Пример #1
0
    def predictor(self, file):
        instance1 = data_ingestion.data_getter()
        data = instance1.data_load(file)

        instance2 = preprocessingfile.LA_preprocess()

        set0 = instance2.initialize_columns(data)
        set1 = instance2.imputer(set0)
        final_data = set1[['Loan_ID']]
        set2 = instance2.drop_Loan_id(set1)
        encoded_data = instance2.encode_cat_f(set2)
        new_data = instance2.drop_columns(encoded_data)

        ss_LA = pickle.load(open('pickle_files/veena_LA_stan_scaler.pkl',
                                 'rb'))
        model_LA = pickle.load(open('pickle_files/LAOpt8model.sav', 'rb'))

        ss_result = ss_LA.fit_transform(new_data)
        result = model_LA.predict(ss_result)

        new_data['output'] = result
        new_data['output'] = np.where(new_data['output'] == 0, "Rejected",
                                      "Accepted")
        final_data['Output'] = new_data['output']
        return final_data
Пример #2
0
    def predictor(self, file):
        instance1 = data_ingestion.data_getter()
        data = instance1.data_load(file)

        instance2 = preprocessingfile.LA_preprocess()

        set0 = instance2.initialize_columns(data)
        final_data = set0[['ID']]

        set1 = instance2.drop_columns(set0)
        new_data = instance2.encoder(set1)
        # ============  pandas-prof. report ============================
        new_data.to_csv(r'graph_input_files\graph_data.csv', index_label=False)
        # ==============================================================
        ss_LA = pickle.load(open('pickle_files/LA_Std_scaler.pkl', 'rb'))
        model_LA = pickle.load(open('pickle_files/DTModel-1.pkl', 'rb'))

        ss_result = ss_LA.transform(new_data)
        result = model_LA.predict(ss_result)

        new_data['output'] = result
        new_data['output'] = np.where(new_data['output'] == 0, "Rejected",
                                      "Accepted")
        final_data['Output'] = new_data['output']
        return final_data
    def predictor(self,file):
        '''
              Description: This method takes the csv file from MA_bulk_predict routes in app.py
                                       and calls pre defined preprocessing classes from prediction folder to give output file.
              Output: Each method returns an output dataframe along with 4 chart names which are created and
                                       stored in static folder of main directory.
              On Failure: Raise Exception.
     '''
        try:
            instance1 = data_ingestion.data_getter()
            data = instance1.data_load(file)
            instance2 = preprocessingfile.MA_preprocess()
            visuals = data_visualization.Data_Visualisation()
            visuals.delete_old_graphs('MA')

            new_data = instance2.initialize_columns(data)
            data_num = new_data.select_dtypes(include='number')  # to get list of all numerical features in data
            num_col_ls = list(data_num.columns)
            # ============  pandas-prof. report ============================
            new_data.to_csv(r'graph_input_files\graph_data.csv', index_label=False)
            # ==============================================================
            final_data = new_data[['CONCAT']]
            new_data=instance2.drop_columns(new_data)
            model_MA= instance1.decompress_pickle('pickle_files/Mortgage_RE.pbz2')

            result = model_MA.predict(new_data)
            final_data['output'] = result
            return final_data
        except Exception as e:
            raise e
Пример #4
0
    def predictor(self, file):
        instance1 = data_ingestion.data_getter()
        data = instance1.data_load(file)
        df = data[['type', 'nameOrig', 'nameDest']].copy(deep=True)
        instance2 = preprocessingfile.preprocess()

        set1 = instance2.new_feature(data)
        set2 = instance2.drop_columns(set1)

        model = joblib.load('pickle_files/pickle_fraud.pkl')
        ss = joblib.load('pickle_files/scaler_fraud.pkl')
        en = joblib.load('pickle_files/encode_fraud.pkl')
        set2[[
            'amount', 'oldbalanceOrg', 'newbalanceOrig', 'oldbalanceDest',
            'newbalanceDest', 'errorBalanceOrig', 'errorBalanceDest'
        ]] = ss.transform(set2[[
            'amount', 'oldbalanceOrg', 'newbalanceOrig', 'oldbalanceDest',
            'newbalanceDest', 'errorBalanceOrig', 'errorBalanceDest'
        ]])
        set2.type = en.transform(set2.type)
        result = model.predict(set2)

        df['output'] = result

        df['output'] = np.where(df['output'] == 0, "Trusted", "Fraud")
        return df
Пример #5
0
    def predictor(self, file):
        instance1 = data_ingestion.data_getter()
        data = instance1.data_load(file)

        instance2 = preprocessingfile.LR_preprocess()

        set0 = instance2.initialize_columns(data)
        set1 = instance2.drop_col(set0)
        set2 = instance2.feature_engg(set1)
        set3 = instance2.outlier_removal(set2)
        set4 = instance2.imputer(set3)
        #set4 = instance2.drop_col(set3)

        #final_data = data['RowID']
        #final_data = pd.DataFrame()
        # ============  pandas-prof. report ============================
        set4.to_csv(r'graph_input_files\graph_data.csv', index_label=False)
        # ==============================================================
        lr_model = pickle.load(open('pickle_files/loan_risk.pkl', 'rb'))

        result = lr_model.predict(set4)

        set4['output'] = result
        set4['output'] = np.where(set4['output'] == 0, "Risky", "Safe")
        #final_data['Output']=set4['output']
        final_data = {
            'RowID': [i for i in set0['RowID']],
            'Output': [i for i in set4['output']]
        }
        return pd.DataFrame(final_data)
Пример #6
0
    def predictor(self, file):
        file_object = open(r'.\Prediction_logs\data_getter_logs.txt', 'a+')

        instance1 = data_ingestion.data_getter(file_object)
        data = instance1.data_load(file)

        file_object = open(r'.\Prediction_logs\preprocessing_logs.txt', 'a+')
        instance2 = preprocessingfile.preprocess(file_object)
        set1 = instance2.set_columns(data)

        remove = instance2.remove_columns(set1)

        type1 = instance2.set_type(remove)

        impute = instance2.imputation(type1)

        feature = instance2.feature_remove(impute)

        scaled = instance2.scaling(feature)

        encoder = instance2.encoding(scaled)

        model = joblib.load('model.pkl')

        result = model.predict(encoder)

        encoder['output'] = result

        return encoder
Пример #7
0
    def retrain_model(self):
        file_object = open(r'.\Retraining_logs\retraining_logs.txt', 'a+')
        instance1 = data_ingestion.data_getter(file_object)
        data = instance1.data_load(self.file)

        instance2 = retraining_preprocessing.preprocess(file_object)

        set1 = instance2.set_columns(data)

        target = instance2.target(set1)

        remove = instance2.remove_columns(target)

        type1 = instance2.set_type(remove)

        impute = instance2.imputation(type1)

        feature = instance2.feature_remove(impute)

        scaled = instance2.scaling(feature)

        encoder = instance2.encoding(scaled)

        result = model_fitter.model_fit(encoder, file_object)

        result.training()
Пример #8
0
    def predictor(self, file):
        instance1 = data_ingestion.data_getter()
        data = instance1.data_load(file)

        instance2 = preprocessingfile.LE_preprocess()
        data_final = instance2.initialize_columns(data)

        # ============  pandas-prof. report ============================
        new_data.to_csv(r'graph_input_files\graph_data.csv', index_label=False)
        # ==============================================================
        le_model = pickle.load(open('pickle_files/LE-DecTreeModel.pkl', 'rb'))

        result = le_model.predict(data_final)

        data_final['output'] = result
        data_final['output'] = np.where(data_final['output'] == 0,
                                        "Not Eligible", "Eligible")

        data_final['RowID'] = pd.Series(
            [i for i in range(len(data_final['output']))])

        final_data = {
            'RowID': [i for i in data_final['RowID']],
            'Output': [i for i in data_final['output']]
        }
        return pd.DataFrame(final_data)
Пример #9
0
    def predictor(self, file):
        '''
                  Description: This method takes the csv file from LE_bulk_predict routes in app.py
                                           and calls pre defined preprocessing classes from prediction folder to give output file.
                  Output: Each method returns an output dataframe along with 4 chart names which are created and
                                           stored in static folder of main directory.
                  On Failure: Raise Exception.
         '''
        try:
            instance1 = data_ingestion.data_getter()
            data = instance1.data_load(file)
            instance2 = preprocessingfile.LE_preprocess()
            visuals = data_visualization.Data_Visualisation()
            visuals.delete_old_graphs('LE')

            data_final = instance2.initialize_columns(data)
            data_num = data_final.select_dtypes(
                include='number'
            )  # to get list of all numerical features in data
            num_col_ls = list(data_num.columns)
            # ============  pandas-prof. report ============================
            data_final.to_csv(r'graph_input_files\graph_data.csv',
                              index_label=False)
            # ==============================================================
            le_model = instance1.decompress_pickle(
                'pickle_files/LE-DecTreeModel.pbz2')

            result = le_model.predict(data_final)
            data_final['output'] = result
            chart3_name, imp_feature = visuals.feature_importance(
                data_final["output"], data_final.drop('output', axis=1),
                'LE')  # feature importance
            if imp_feature in num_col_ls:
                chart4_name = visuals.numeric_summary(
                    data_final, imp_feature,
                    'LR')  # most important feature graph
            else:
                chart4_name = visuals.categorical_summary(
                    data_final, imp_feature, 'LR')

            data_final['output'] = np.where(data_final['output'] == 0,
                                            "Not Eligible", "Eligible")

            data_final['RowID'] = pd.Series(
                [i for i in range(len(data_final['output']))])

            final_data = {
                'RowID': [i for i in data_final['RowID']],
                'Output': [i for i in data_final['output']]
            }
            chart1_name = visuals.count_plot("Output", final_data,
                                             'LE')  # count plot
            chart2_name = visuals.heat_map("output", data_final,
                                           'LE')  # heat map
            return pd.DataFrame(
                final_data), chart1_name, chart2_name, chart3_name, chart4_name
        except Exception as e:
            raise e
Пример #10
0
    def predictor(self, file):
        '''
                  Description: This method takes the csv file from LA_bulk_predict routes in app.py
                                           and calls pre defined preprocessing classes from prediction folder to give output file.
                  Output: Each method returns an output dataframe along with 4 chart names which are created and
                                           stored in static folder of main directory.
                  On Failure: Raise Exception.
         '''
        try:
            instance1 = data_ingestion.data_getter()
            data = instance1.data_load(file)
            instance2 = preprocessingfile.LA_preprocess()
            visuals = data_visualization.Data_Visualisation()
            visuals.delete_old_graphs('LA')

            set0 = instance2.initialize_columns(data)
            data_num = set0.select_dtypes(
                include='number'
            )  # to get list of all numerical features in data
            num_col_ls = list(data_num.columns)
            final_data = set0[['ID']]
            set1 = instance2.drop_columns(set0)
            new_data = instance2.encoder(set1)
            # ============  pandas-prof. report ============================
            new_data.to_csv(r'graph_input_files\graph_data.csv',
                            index_label=False)
            # ==============================================================
            ss_LA = instance1.decompress_pickle(
                'pickle_files/LA_Std_scaler.pbz2')
            model_LA = instance1.decompress_pickle(
                'pickle_files/DTModel-1.pbz2')

            ss_result = ss_LA.transform(new_data)
            result = model_LA.predict(ss_result)

            new_data['output'] = result
            chart3_name, imp_feature = visuals.feature_importance(
                new_data["output"], new_data.drop('output', axis=1),
                'LA')  # feature importance
            if imp_feature in num_col_ls:
                chart4_name = visuals.numeric_summary(
                    new_data, imp_feature,
                    'LA')  # most important feature graph
            else:
                chart4_name = visuals.categorical_summary(
                    new_data, imp_feature, 'LA')

            new_data['output'] = np.where(new_data['output'] == 0, "Rejected",
                                          "Accepted")
            final_data['Output'] = new_data['output']

            chart1_name = visuals.count_plot("Output", final_data,
                                             'LA')  # count plot
            chart2_name = visuals.heat_map("output", new_data,
                                           'LA')  # heat map
            return final_data, chart1_name, chart2_name, chart3_name, chart4_name
        except Exception as e:
            raise e
Пример #11
0
    def predictor(self, file):
        '''
              Description: This method takes the csv file from LR_bulk_predict routes in app.py
                                       and calls pre defined preprocessing classes from prediction folder to give output file.
              Output: Each method returns an output dataframe along with 4 chart names which are created and
                                       stored in static folder of main directory.
              On Failure: Raise Exception.
     '''
        try:
            instance1 = data_ingestion.data_getter()
            data = instance1.data_load(file)
            instance2 = preprocessingfile.LR_preprocess()
            visuals = data_visualization.Data_Visualisation()
            visuals.delete_old_graphs('LR')

            set0 = instance2.initialize_columns(data)
            data_num = set0.select_dtypes(
                include='number'
            )  # to get list of all numerical features in data
            num_col_ls = list(data_num.columns)

            set1 = instance2.drop_col(set0)
            set2 = instance2.feature_engg(set1)
            set3 = instance2.outlier_removal(set2)
            set4 = instance2.imputer(set3)
            # ============  pandas-prof. report ============================
            set4.to_csv(r'graph_input_files\graph_data.csv', index_label=False)
            # ==============================================================
            lr_model = instance1.decompress_pickle(
                'pickle_files/loan_risk.pbz2')

            result = lr_model.predict(set4)
            set4['output'] = result
            chart3_name, imp_feature = visuals.feature_importance(
                set4["output"], set4.drop('output', axis=1),
                'LR')  # feature importance
            if imp_feature in num_col_ls:
                chart4_name = visuals.numeric_summary(
                    set4, imp_feature, 'LR')  # most important feature graph
            else:
                chart4_name = visuals.categorical_summary(
                    set4, imp_feature, 'LR')

            set4['output'] = np.where(set4['output'] == 0, "Risky", "Safe")
            final_data = {
                'RowID': [i for i in set0['RowID']],
                'Output': [i for i in set4['output']]
            }

            chart1_name = visuals.count_plot("Output", final_data,
                                             'LR')  # count plot
            chart2_name = visuals.heat_map("output", set4, 'LR')  # heat map
            print(chart3_name)
            print(chart4_name)
            return pd.DataFrame(
                final_data), chart1_name, chart2_name, chart3_name, chart4_name
        except Exception as e:
            raise e
Пример #12
0
 def retrainer(self, file):
     instance1 = data_ingestion.data_getter()
     data = instance1.data_load(file)
     instance2 = preprocessingfile.retrain()
     set0 = instance2.initialize_columns(data)
     set1 = instance2.drop_columns(set0)
     new_data = instance2.obj_to_cat(set2)
     x, y = instance2.smote(new_data)
     instance2.model(x, y)
    def predictor(self,file):
        '''
                  Description: This method takes the csv file from PLR_bulk_predict routes in app.py
                                           and calls pre defined preprocessing classes from prediction folder to give output file.
                  Output: Each method returns an output dataframe along with 4 chart names which are created and
                                           stored in static folder of main directory.
                  On Failure: Raise Exception.
         '''
        try:
            instance1 = data_ingestion.data_getter()
            visuals = data_visualization.Data_Visualisation()
            visuals.delete_old_graphs('LPR')

            data = instance1.data_load(file[0])
            data1 = data[['SK_ID_CURR']]
            bureau=instance1.data_load(file[1])
            previos_application=instance1.data_load(file[5])
            pos_cash=instance1.data_load(file[4])
            insta_payments=instance1.data_load(file[3])
            credit_card=instance1.data_load(file[2])

            instance2 = preprocessingfile.LPR_preprocess()
            data=instance2.error_flag_column(data)
            data=instance2.new_columns(data)
            application_bureau=instance2.joining_berau_application(bureau,data)
            application_bureau=instance2.feature_engineering(bureau,application_bureau)
            application_bureau_prev=instance2.joining_previousapplication_to_applicationbereau(previos_application,application_bureau)
            application_bureau_prev=instance2.Joining_POS_CASH_balance_to_application_bureau_prev_data(pos_cash,application_bureau_prev)
            application_bureau_prev=instance2.joining_InstallmentsPaymentsdata_to_application_bureau_prev_data(insta_payments,application_bureau_prev)
            application_bureau_prev=instance2.Joining_Creditcardbalancedata_to_application_bureau_prev(application_bureau_prev,credit_card)
            application_bureau_prev=instance2.featurization(application_bureau_prev)
            application_bureau_prev=instance2.feature_selection(application_bureau_prev)

            data_num = application_bureau_prev.select_dtypes(include='number')  # to get list of all numerical features in data
            num_col_ls = list(data_num.columns)

            model = instance1.decompress_pickle(r'pickle_files\DecTreePLR.pbz2')
            output = model.predict(application_bureau_prev)
            application_bureau_prev['result'] = output
            chart3_name, imp_feature = visuals.feature_importance(application_bureau_prev["result"],
                                                                  application_bureau_prev.drop('result', axis=1),'LPR')  # feature importance
            if imp_feature in num_col_ls:
                chart4_name = visuals.numeric_summary(application_bureau_prev, imp_feature, 'LPR')  # most important feature graph
            else:
                chart4_name = visuals.categorical_summary(application_bureau_prev, imp_feature, 'LPR') # most imp feature categorical graph
            data1['result']=application_bureau_prev["result"]
            data1['result'] = np.where(data1['result'] == 0, "Loan Repayed", "Not Repayed")
            chart1_name = visuals.count_plot("result", data1,'LPR')  # count plot
            chart2_name = visuals.heat_map("result", application_bureau_prev, 'LPR')  # heat map
            return data1,chart1_name,chart2_name,chart3_name,chart4_name
        except Exception as e:
            raise e
    def predictor(self, file):
        '''
                  Description: This method takes the csv file from MA_bulk_predict routes in app.py
                                           and calls pre defined preprocessing classes from prediction folder to give output file.
                  Output: Each method returns an output dataframe along with 4 chart names which are created and
                                           stored in static folder of main directory.
                  On Failure: Raise Exception.
         '''
        try:
            instance1 = data_ingestion.data_getter()
            data = instance1.data_load(file)

            instance2 = preprocessingfile.MS_preprocess()
            visuals = data_visualization.Data_Visualisation()
            visuals.delete_old_graphs('MS')

            set0 = instance2.rename(data)
            data_num = set0.select_dtypes(include='number')  # to get list of all numerical features in data
            num_col_ls = list(data_num.columns)

            set1 = instance2.drop_columns(set0)
            set2 = instance2.days_passed(set1)
            set3 = instance2.job(set2)
            set4 = instance2.education_cat(set3)
            set5 = instance2.contacted_month(set4)
            data_final = instance2.contacts_before_campaign(set5)

            data_final = pd.get_dummies(data_final,drop_first=True)
            data_final = instance2.columns_match(data_final)
            data_final = data_final[['last_call_duration','age','days_passed_recent','contacts_during_campaign','housing_loan_yes','personal_loan_yes','day_of_week_tue','last_contacted_month_may-aug','day_of_week_thu','marital_married','day_of_week_wed','day_of_week_mon','marital_single','contacts_during_campaign']]
            MS__model = instance1.decompress_pickle('pickle_files/MS_randomforest_model4.pbz2')
            result = MS__model.predict(data_final)
            data_final["output"]=result
            chart3_name, imp_feature = visuals.feature_importance(data_final["output"], data_final.drop('output', axis=1),'MS')  # feature importance
            if imp_feature in num_col_ls:
                chart4_name = visuals.numeric_summary(data_final, imp_feature, 'MS')  # most important feature graph
            else:
                chart4_name = visuals.categorical_summary(data_final, imp_feature, 'MS')

            final_result=pd.DataFrame(result, columns=['Output'])
            final_result["SrNo."]=np.arange(len(final_result["Output"]))
            pop_col=final_result.pop('Output')
            final_result["output"]=pop_col
            final_result['output'] = np.where(final_result['output'] == 0, "Not Subscribed", "Term Deposit")

            chart1_name = visuals.count_plot("output", final_result, 'MS')  # count plot
            chart2_name = visuals.heat_map("output", data_final, 'MS')  # heat map
            return final_result, chart1_name, chart2_name, chart3_name, chart4_name
        except Exception as e:
            raise e
Пример #15
0
    def predictor(self, file):
        instance1 = data_ingestion.data_getter()
        data = instance1.data_load(file)

        instance2 = preprocessingfile.Fraud_preprocess()

        set0 = instance2.initialize_columns(data)
        set1 = instance2.drop_columns(set0)
        df = set1[['customer']]
        set2 = set1.drop(['customer'], axis=1)
        new_data = instance2.obj_to_cat(set2)

        model_Fraud = joblib.load('pickle_files/Fraud_new_model.pkl')
        result = model_Fraud.predict(new_data)

        df['output'] = result
        df['output'] = np.where(df['output'] == 0, "Trusted", "Fraud")
        return df
Пример #16
0
    def predictor(self, file):
        instance1 = data_ingestion.data_getter()
        data = instance1.data_load(file)

        instance2 = preprocessingfile.MA_preprocess()

        new_data = instance2.initialize_columns(data)
        # ============  pandas-prof. report ============================
        new_data.to_csv(r'graph_input_files\graph_data.csv', index_label=False)
        # ==============================================================
        final_data = new_data[['CONCAT']]
        new_data = instance2.drop_columns(new_data)
        model_MA = pickle.load(open('pickle_files/Mortgage_RE.pkl', 'rb'))

        result = model_MA.predict(new_data)

        final_data['output'] = result
        return final_data
Пример #17
0
def preprocessing_new():
    print("started!!!")

    all_batch_files = glob.glob("Training_Raw_files_validated/Good_Raw/*.csv")
    if len(all_batch_files) > 0:
        data_get = data_getter()
        pre = Preprocessing()
        data = data_get.data_load(all_batch_files[0])
        prep = Get_independet_dependent_data()

        X, y = prep.get_independent_dependent_data(data)

        print("independent columns:-", X.columns)

        expTextToNum = ExpTextToNum(data)
        drop_cols = ['Loan ID', 'Customer ID', 'Bankruptcies', 'Tax Liens']
        ##testing the combo
        credit_score_normalizer = CreditScoreNormalizer(data)

        currentScore_normalizer = CurrentLoanAmountNormalizer(X)
        monthly_deliquent = MonthDeliquent(X)
        home_own_spell = HomeOwnSpell(X)
        purpose_Spell = PurposeSpell(X)
        ###For categorical values we need one hot encoding also
        ohe = OneHotEncoder(handle_unknown='ignore', sparse=False)

        # KNN Imputation Need to be done here
        #knn_imputer = KNN_Imputer(X)

        # num_col = pre.get_numerical_col(data)


        ct2 = make_column_transformer((expTextToNum, ['Years in current job']), \
                                      (credit_score_normalizer, ['Credit Score']), \
                                      (currentScore_normalizer, ['Current Loan Amount']), \
                                      (monthly_deliquent,['Months since last delinquent']),\
                                      (home_own_spell,['Home Ownership']),\
                                      (purpose_Spell,['Purpose']),\

                                      (ohe, ['Term', 'Home Ownership', 'Purpose']),\

                                      ('drop', drop_cols), \
                                      remainder='passthrough')

        pipchk = make_pipeline(ct2, ArrayToDf(X))
        dataframe = pipchk.fit_transform(X)

        # Normalized_data = ct2.fit_transform(X)
        dataframe.to_csv("dataframe.csv")
        data1 = pd.read_csv('dataframe.csv', index_col=[0])

        print("Dataframe")
        print(dataframe)

        problem_utils = utills()

        trans_dataframe = pre.fill_missing_value_KNN_imputer(data1)
        trans_dataframe.to_csv(problem_utils.saveResult_result())
        print(
            "///////////////////KNN Data Transform Data frame///////////////////"
        )
        #print(trans_dataframe)

        # print(dataframe.columns)
        # print(Normalized_data)
        #shutil.move("Training_Raw_files_validated/Good_Raw/credit.csv", "Preprocessed_Data")
        print("done")

        return Response("preprocessing successfull!!")
Пример #18
0
def warns(*args, **kwargs):
    pass


warnings.warn = warns

ALLOWED_EXTENSIONS = set(['csv', 'xlsx', 'data'])


def allowed_file(filename):
    return '.' in filename and filename.rsplit(
        '.', 1)[1].lower() in ALLOWED_EXTENSIONS


# instance for data ingestion class (to load compressed pickle file)..
pickle_load = data_ingestion.data_getter()

# load models stored in compressed pickle files ..
ss_LA = pickle_load.decompress_pickle('pickle_files/LA_Std_scaler.pbz2')
model_LA = pickle_load.decompress_pickle('pickle_files/DTModel-1.pbz2')
model_Fraud = joblib.load('pickle_files/Fraud_new_model.pkl')
model_LR = pickle_load.decompress_pickle('pickle_files/loan_risk.pbz2')
model_LE = pickle_load.decompress_pickle('pickle_files/LE-DecTreeModel.pbz2')
model_MA = pickle_load.decompress_pickle('pickle_files/Mortgage_RE.pbz2')
model_MS = pickle_load.decompress_pickle(
    'pickle_files/MS_randomforest_model4.pbz2')

# instances of prediction.py file for bulk upload of different models ..
LA_instance = LA_predict()
Fraud_instance = Fraud_predict()
LR_instance = LR_predict()