コード例 #1
0
    def reshape_data_xgboost(self, full_series, dynamic_features,
                             static_features, outcome, grouping):

        flat_df, timesteps = flatten(full_series, dynamic_features, grouping,
                                     static_features, outcome)
        temporal_features = set(flat_df.columns) - set(static_features)
        temporal_features = set(temporal_features) - set([outcome, grouping])

        X_working = flat_df
        y_working = X_working[outcome].astype(int)
        training_groups = X_working[grouping]
        X_working_static = X_working[static_features]
        X_working_static.loc[grouping] = training_groups
        X_working = X_working[temporal_features]
        X_working = scale(X_working, temporal_features)

        slopes_df = generate_slopes(X_working, temporal_features,
                                    static_features, grouping, training_groups)
        aggregate_df = generate_aggregates(X_working, temporal_features,
                                           grouping, training_groups)

        slopes_static_aggregate_df = pd.concat([slopes_df, X_working_static],
                                               axis=1,
                                               join='inner')
        slopes_static_aggregate_df = pd.concat(
            [slopes_static_aggregate_df, aggregate_df], axis=1, join='inner')

        slopes_static_aggregate_df = slopes_static_aggregate_df.loc[:,
                                                                    ~slopes_static_aggregate_df
                                                                    .columns.
                                                                    duplicated(
                                                                    )]
        slopes_static_aggregate_groups = slopes_static_aggregate_df[grouping]
        slopes_static_aggregate_df.drop(columns=[grouping],
                                        inplace=True,
                                        axis=1)

        return slopes_static_aggregate_df, y_working
コード例 #2
0
def main():
    configs = json.load(open('Configuration.json', 'r'))

    grouping = configs['data']['grouping']
    static_features = configs['data']['static_columns']

    outcomes = (configs['data']['classification_outcome'])
    timeseries_path = configs['paths']['data_path']

    for outcome in outcomes:
        decision_maker = DecisionMaker()

        time_series_nosmote = pd.read_csv(timeseries_path +
                                          "NonSMOTEDTimeSeries/" + outcome +
                                          "FlatTimeSeries1Day.csv")
        time_series = pd.read_csv(timeseries_path + "SMOTEDTimeSeries/" +
                                  outcome + "FlatTimeSeries1Day.csv")

        X_cols = (time_series.columns).tolist()
        X_cols.remove(outcome)
        X_train, X_valid, y_train, y_valid = train_test_split(
            time_series[X_cols],
            time_series[outcome],
            test_size=0.33,
            stratify=time_series[outcome],
            random_state=42)

        test_ids = set([x.partition('.')[0] for x in X_valid[grouping]])

        #print(test_ids.intersection(time_series_nosmote[grouping] ))
        X_test1 = time_series_nosmote.loc[time_series_nosmote[grouping].isin(
            test_ids)]
        distrs_percents = [
            get_distribution_percentages(
                (time_series_nosmote[outcome]).astype(int))
        ]
        print(distrs_percents)

        #####feature selector
        temporal_features = set(X_train.columns) - set(static_features)

        feature_selector = XGBoostClassifier(X_train[temporal_features],
                                             y_train, outcome, grouping)
        fs_y, fs_ths, fs_id, fs_fi = feature_selector.run_xgb("temporal")

        feature_selector.predict(X_valid[temporal_features], y_valid)

        decision_maker.add_classifier(outcome + "Tmp", fs_y, fs_ths, fs_id,
                                      fs_fi)

        featuredf = pd.DataFrame()

        temporal_features.remove(grouping)
        featuredf['features'] = list(temporal_features)
        featuredf['imp'] = fs_fi
        featuredf = featuredf[featuredf['imp'] > 0]

        ########################################
        #baseline and static
        baseline_features = featuredf['features']

        baseline_features = set(
            [x.partition('_')[0] for x in list(baseline_features)])

        baseline_features = [x + "_0" for x in list(baseline_features)]

        baseline_features.insert(0, grouping)
        baseline_static_features = baseline_features + static_features

        slopes_df = generate_slopes(X_train, static_features, grouping)
        slopes_static_baseline_df = pd.concat(
            [slopes_df, X_train[baseline_static_features]],
            axis=1,
            join='inner')

        slopes_static_baseline_df = slopes_static_baseline_df.loc[:,
                                                                  ~slopes_static_baseline_df
                                                                  .columns.
                                                                  duplicated()]

        slopes_df_test = generate_slopes(X_valid, static_features, grouping)
        slopes_static_baseline_test_df = pd.concat(
            [slopes_df_test, X_valid[baseline_static_features]],
            axis=1,
            join='inner')
        slopes_static_baseline_test_df = slopes_static_baseline_test_df.loc[:,
                                                                            ~slopes_static_baseline_test_df
                                                                            .
                                                                            columns
                                                                            .
                                                                            duplicated(
                                                                            )]

        slopes_static_baseline_classifier = XGBoostClassifier(
            slopes_static_baseline_df, y_train, outcome, grouping)

        bs_y, bs_ths, bs_id, bs_fi = slopes_static_baseline_classifier.run_xgb(
            "baseline_static_slope")
        slopes_static_baseline_classifier.predict(
            slopes_static_baseline_test_df, y_valid)
        tf.keras.backend.clear_session()

        decision_maker.add_classifier(outcome + "bss", bs_y, bs_ths, bs_id,
                                      bs_fi)

        ####LSTM autoencoder
        smoted_stacked_series = pd.read_csv(timeseries_path +
                                            "SMOTEDTimeSeries/" + outcome +
                                            "StackedTimeSeries1Day.csv")
コード例 #3
0
def main():
    configs = json.load(open('Configuration.json', 'r'))
    epochs = configs['training']['epochs']
    grouping = configs['data']['grouping']
    dynamic_features = configs['data']['dynamic_columns']
    static_features = configs['data']['static_columns']

    outcomes = configs['data']['classification_outcome']
    lookback = configs['data']['batch_size']
    timeseries_path = configs['paths']['data_path']
    saved_models_path = configs['paths']['saved_models_path']

    ##read, impute and scale dataset
    non_smotedtime_series = pd.read_csv(timeseries_path +
                                        "TimeSeriesAggregatedUpto0.csv")
    non_smotedtime_series[dynamic_features] = impute(non_smotedtime_series,
                                                     dynamic_features)
    normalized_timeseries = scale(non_smotedtime_series, dynamic_features)
    normalized_timeseries.insert(0, grouping, non_smotedtime_series[grouping])

    ##start working per outcome
    for outcome in outcomes:
        decision_maker = DecisionMaker()
        fold_ind, train_ind, test_ind = get_train_test_split(
            non_smotedtime_series[outcome].astype(int),
            non_smotedtime_series[grouping])

        ##Load LSTM models if they exist, otherwise train new models and save them
        filename = saved_models_path + configs['model'][
            'name'] + outcome + '.h5'

        X_train, X_train_y0, X_valid_y0, X_valid, y_valid, X_test, y_test, timesteps, \
        n_features = \
            process_data(normalized_timeseries, non_smotedtime_series, outcome, grouping, lookback,
                         train_ind, test_ind)

        if os.path.isfile(filename):
            autoencoder = LSTMAutoEncoder(configs['model']['name'] + outcome,
                                          outcome,
                                          timesteps,
                                          n_features,
                                          saved_model=filename)
            autoencoder.summary()

        else:
            autoencoder = LSTMAutoEncoder(configs['model']['name'] + outcome,
                                          outcome, timesteps, n_features)
            autoencoder.summary()

            autoencoder.fit(X_train_y0, X_train_y0, epochs, lookback,
                            X_valid_y0, X_valid_y0, 2)
            autoencoder.plot_history()
            ###save model
            filename = saved_models_path + configs['model'][
                'name'] + outcome + '.h5'
            autoencoder.save_model(filename)

        ####Predicting using the fitted model (loaded or trained)

        train_x_predictions = autoencoder.predict(X_train)
        mse_train = np.mean(np.power(
            lstm_flatten(X_train) - lstm_flatten(train_x_predictions), 2),
                            axis=1)

        test_x_predictions = autoencoder.predict(X_test)

        mse_test = np.mean(np.power(
            lstm_flatten(X_test) - lstm_flatten(test_x_predictions), 2),
                           axis=1)

        test_error_df = pd.DataFrame({
            'Reconstruction_error': mse_test,
            'True_class': y_test.tolist()
        })


        pred_y, best_threshold, precision_rt, recall_rt = \
              autoencoder.predict_binary(test_error_df.True_class, test_error_df.Reconstruction_error)

        autoencoder.output_performance(test_error_df.True_class,
                                       test_error_df.Reconstruction_error,
                                       pred_y)
        autoencoder.plot_reconstruction_error(test_error_df, best_threshold)
        autoencoder.plot_roc(test_error_df)
        autoencoder.plot_pr(precision_rt, recall_rt)

        #Feature Selector
        training_loc = train_ind[0]  #+train_ind[1]
        training_ids = non_smotedtime_series.iloc[training_loc]
        training_ids = training_ids[grouping]

        testing_ids = non_smotedtime_series.iloc[test_ind[1]]
        testing_ids = testing_ids[grouping]

        flat_df, timesteps = flatten(non_smotedtime_series, dynamic_features,
                                     grouping, static_features, outcome)
        temporal_features = set(flat_df.columns) - set(static_features)
        temporal_features = set(temporal_features) - set([outcome, grouping])

        X_train = flat_df.loc[flat_df[grouping].isin(training_ids)]
        y_train = X_train[outcome].astype(int)
        training_groups = X_train[grouping]
        X_train_static = X_train[static_features]
        X_train_static.loc[grouping] = training_groups
        X_train = X_train[temporal_features]
        X_train = scale(X_train, temporal_features)
        X_train['mse'] = mse_train

        #X_train, y_train = smote(X_train, y_train)
        X_test = flat_df.loc[flat_df[grouping].isin(testing_ids)]
        y_test = X_test[outcome].astype(int)
        testing_groups = X_test[grouping]
        X_test_static = X_test[static_features]
        X_test_static.loc[grouping] = testing_groups
        X_test = X_test[temporal_features]
        X_test = scale(X_test, temporal_features)
        X_test['mse'] = mse_test

        feature_selector = XGBoostClassifier(X_train, y_train, outcome,
                                             grouping)  #
        feature_selector.fit("temporal", training_groups)

        y_pred_binary, best_threshold, precision_rt, recall_rt = feature_selector.predict(
            X_test, y_test)
        feature_selector.plot_pr(precision_rt, recall_rt, "XGBoost Temporal")

        featuredf = pd.DataFrame()

        temporal_features = set(temporal_features) - set([outcome])
        featuredf['features'] = list(temporal_features)
        #featuredf['imp'] = fs_fi
        #featuredf = featuredf[featuredf['imp'] > 0]
        ########
        baseline_features = featuredf['features']

        baseline_features = set(
            [x.partition('_')[0] for x in list(baseline_features)])

        baseline_features = [x + "_0" for x in list(baseline_features)]

        baseline_features.insert(0, grouping)
        baseline_static_features = baseline_features + static_features

        slopes_df = generate_slopes(X_train, temporal_features,
                                    static_features, grouping, training_groups)

        aggregate_df = generate_aggregates(X_train, temporal_features,
                                           grouping, training_groups)

        slopes_static_baseline_train_df = pd.concat(
            [slopes_df, X_train_static], axis=1, join='inner')

        slopes_static_baseline_train_df = slopes_static_baseline_train_df.loc[:,
                                                                              ~slopes_static_baseline_train_df
                                                                              .
                                                                              columns
                                                                              .
                                                                              duplicated(
                                                                              )]
        slopes_static_baseline_train_groups = slopes_static_baseline_train_df[
            grouping]
        slopes_static_baseline_train_df.drop(columns=[grouping],
                                             inplace=True,
                                             axis=1)
        slopes_static_baseline_train_df['mse'] = mse_train

        slopes_df_test = generate_slopes(X_test, temporal_features,
                                         static_features, grouping,
                                         testing_groups)

        slopes_static_baseline_test_df = pd.concat(
            [slopes_df_test, X_test_static], axis=1, join='inner')
        slopes_static_baseline_test_df = slopes_static_baseline_test_df.loc[:,
                                                                            ~slopes_static_baseline_test_df
                                                                            .
                                                                            columns
                                                                            .
                                                                            duplicated(
                                                                            )]
        slopes_static_baseline_test_groups = slopes_static_baseline_test_df[
            grouping]
        slopes_static_baseline_test_df.drop(columns=[grouping],
                                            inplace=True,
                                            axis=1)
        slopes_static_baseline_test_df['mse'] = mse_test

        slopes_static_baseline_classifier = XGBoostClassifier(
            slopes_static_baseline_train_df, y_train, outcome, grouping)

        #bs_y, bs_ths, bs_id, bs_fi = slopes_static_baseline_classifier.fit("baseline_static_slope",
        #                                                                      slopes_static_baseline_train_groups)
        slopes_static_baseline_classifier.fit(
            "baseline_static_slope", slopes_static_baseline_train_groups)
        y_pred_binary, best_threshold, precision_rt, recall_rt = \
            slopes_static_baseline_classifier.predict( slopes_static_baseline_test_df, y_test)
        slopes_static_baseline_classifier.plot_pr(precision_rt, recall_rt,
                                                  "XGBoost Static")
コード例 #4
0
def main () :
    configs = json.load(open('Configuration.json', 'r'))
    epochs = configs['training']['epochs']
    grouping = configs['data']['grouping']
    dynamic_features = configs['data']['dynamic_columns']
    static_features = configs['data']['static_columns']

    outcomes = configs['data']['classification_outcome']
    lookback = configs['data']['batch_size']
    timeseries_path = configs['paths']['data_path']
    autoencoder_models_path = configs['paths']['autoencoder_models_path']
    xgboost_models_path = configs['paths']['xgboost_models_path']
    test_data_path = configs['paths']['test_data_path']

    ##read, impute and scale dataset
    non_smotedtime_series = pd.read_csv(timeseries_path + "TimeSeriesAggregatedUpto0.csv")
    non_smotedtime_series[dynamic_features] = impute(non_smotedtime_series, dynamic_features)
    normalized_timeseries = scale(non_smotedtime_series, dynamic_features)
    normalized_timeseries.insert(0, grouping, non_smotedtime_series[grouping])

    #intialise classification report which will house results of all outcomes
    classification_report = ClassificationReport()

    ##start working per outcome
    for outcome in outcomes :
        fold_ind, train_ind, test_ind = get_train_test_split(non_smotedtime_series[outcome].astype(int),
                                                             non_smotedtime_series[grouping])

        ##Load LSTM models if they exist, otherwise train new models and save them
        autoencoder_filename = autoencoder_models_path + configs['model']['name'] + outcome + '.h5'
        X_train, X_train_y0, X_valid_y0, X_valid, y_valid, X_test, y_test, timesteps, \
        n_features = \
            process_data(normalized_timeseries, non_smotedtime_series, outcome, grouping, lookback,
                         train_ind, test_ind)
        if ("3D" not in outcome) and ("5D" not in outcome) :
            if os.path.isfile(autoencoder_filename):
                print(" Autoencoder trained model exists for oucome", outcome,"file:" , autoencoder_filename)
                autoencoder = LSTMAutoEncoder(configs['model']['name'] + outcome, outcome,
                                              timesteps, n_features,saved_model = autoencoder_filename)
                autoencoder.summary()


            else :
                print("Autencoder trained model does not exist for outcome", outcome, "file:", autoencoder_filename)
                autoencoder = LSTMAutoEncoder(configs['model']['name'] + outcome, outcome, timesteps, n_features)
                autoencoder.summary()

                autoencoder.fit(X_train_y0, X_train_y0, epochs, lookback, X_valid_y0, X_valid_y0, 2)
                autoencoder.plot_history()

            train_x_predictions = autoencoder.predict(X_train)
            mse_train = np.mean(np.power(lstm_flatten(X_train) - lstm_flatten(train_x_predictions), 2), axis=1)

            test_x_predictions = autoencoder.predict(X_test)

            mse_test = np.mean(np.power(lstm_flatten(X_test) - lstm_flatten(test_x_predictions), 2), axis=1)

            test_error_df = pd.DataFrame({'Reconstruction_error' : mse_test,
                                          'True_class' : y_test.tolist()})

            pred_y, best_threshold, precision_rt, recall_rt = \
                  autoencoder.predict_binary(test_error_df.True_class, test_error_df.Reconstruction_error)

            autoencoder.output_performance(test_error_df.True_class, pred_y)
            autoencoder.plot_reconstruction_error(test_error_df, best_threshold)
            autoencoder.plot_roc(test_error_df)
            autoencoder.plot_pr(precision_rt, recall_rt)

            #Feature Selector
            training_loc = train_ind[0]#+train_ind[1]
            training_ids = non_smotedtime_series.iloc[training_loc]
            training_ids = training_ids[grouping]

            testing_ids = non_smotedtime_series.iloc[test_ind[1]]
            testing_ids = testing_ids[grouping]

            flat_df ,timesteps= flatten (non_smotedtime_series, dynamic_features, grouping, static_features, outcome)
            temporal_features = set(flat_df.columns) - set(static_features)
            temporal_features = set(temporal_features) - set([outcome,grouping])

            X_train = flat_df.loc[flat_df[grouping].isin(training_ids)]
            y_train = X_train[outcome].astype(int)
            training_groups  = X_train[grouping]
            X_train_static = X_train[static_features]
            X_train_static.loc[grouping] = training_groups
            X_train = X_train[temporal_features]
            X_train = scale(X_train, temporal_features)
            X_train['mse'] = mse_train
            X_test = flat_df.loc[flat_df[grouping].isin(testing_ids)]
            y_test = X_test[outcome].astype(int)
            testing_groups = X_test[grouping]
            X_test_static = X_test[static_features]
            X_test_static.loc[grouping] = testing_groups
            X_test= X_test[temporal_features]
            X_test = scale(X_test, temporal_features)
            X_test['mse'] = mse_test

            ########
            slopes_df = generate_slopes ( X_train, temporal_features, static_features,grouping, training_groups)
            aggregate_df = generate_aggregates ( X_train, temporal_features, grouping, training_groups )

            slopes_static_aggregate_train_df = pd.concat([slopes_df, X_train_static], axis=1,join='inner')
            slopes_static_aggregate_train_df = pd.concat([slopes_static_aggregate_train_df, aggregate_df], axis=1,join='inner')
            slopes_static_aggregate_train_df = slopes_static_aggregate_train_df.loc[:, ~slopes_static_aggregate_train_df.columns.duplicated()]
            slopes_static_aggregate_train_groups  = slopes_static_aggregate_train_df[grouping]
            slopes_static_aggregate_train_df.drop(columns = [grouping], inplace=True, axis=1)
            slopes_static_aggregate_train_df['mse'] = mse_train

            slopes_df_test = generate_slopes ( X_test, temporal_features, static_features, grouping, testing_groups)
            aggregate_df_test = generate_aggregates ( X_test, temporal_features, grouping, testing_groups )
            slopes_static_aggregate_test_df = pd.concat([slopes_df_test, X_test_static], axis=1,join='inner')
            slopes_static_aggregate_test_df = pd.concat([slopes_static_aggregate_test_df, aggregate_df_test], axis=1,join='inner')
            slopes_static_aggregate_test_df = slopes_static_aggregate_test_df.loc[:, ~slopes_static_aggregate_test_df.columns.duplicated()]
            slopes_static_aggregate_test_df.drop(columns = [grouping], inplace=True,axis=1)
            slopes_static_aggregate_test_df['mse'] = mse_test


            xgboost_filename = xgboost_models_path + configs['model']['name'] + outcome + '.bin'
            if os.path.isfile(xgboost_filename) :
                slopes_static_baseline_classifier = XGBoostClassifier(slopes_static_aggregate_train_df,
                                                                  y_train, outcome, grouping,
                                                                      saved_model = xgboost_filename)

            else:
                slopes_static_baseline_classifier = XGBoostClassifier(slopes_static_aggregate_train_df,
                                                                  y_train, outcome, grouping)
                slopes_static_baseline_classifier.fit("aggregate_slopes_static_slope", slopes_static_aggregate_train_groups)
                #slopes_static_baseline_classifier.save_model(xgboost_filename)

            y_pred_binary, best_threshold, precision_rt, recall_rt, yhat = \
                slopes_static_baseline_classifier.predict( slopes_static_aggregate_test_df, y_test)

            slopes_static_baseline_classifier.output_performance(y_test, y_pred_binary)
            slopes_static_baseline_classifier.plot_pr(precision_rt, recall_rt, "XGBoost Static")
            to_write_for_plotting = slopes_static_aggregate_test_df
            to_write_for_plotting['outcome'] = y_test
            to_write_for_plotting.to_csv(test_data_path+outcome+".csv", index=False)

            #add to classification report

            classification_report.add_model_result(outcome,y_test, y_pred_binary, best_threshold,
                                                   precision_rt, recall_rt, yhat)


            #delete variables
            del slopes_static_aggregate_train_df
            del slopes_static_aggregate_test_df
            del X_train
            del X_train_y0
            del X_valid_y0
            del X_valid
            del y_valid
            del X_test
            del y_test
            del timesteps
            del train_x_predictions
            del test_x_predictions
            del test_error_df
    #risk_score_visualiser = Visualiser(normalized_timeseries, non_smotedtime_series,
     #                                  dynamic_features, static_features
      #                                 )
    #After fitting model to all outcomes, plot and get summary statistics
    classification_report.plot_distributions_vs_aucs()
    classification_report.plot_pr_auc()
    classification_report.plot_auc()