Esempio n. 1
0
    # ###################################################################################################################
    print("Adding forecaster.zip to system path")
    sys.path.insert(0, "forecaster.zip")
    # ###################################################################################################################

    from support_func import get_current_date, get_sample_customer_list, date_check, \
        get_sample_customer_list_new_addition, get_previous_sundays
    from properties import MODEL_BUILDING
    import properties as p
    from run_weekly import run_weekly
    from run_monthly import run_monthly

    ####################################################################################################################

    # Getting Current Date Time for AppName
    appName = "_".join([MODEL_BUILDING, get_current_date()])
    ####################################################################################################################

    # conf = SparkConf()

    spark = SparkSession \
        .builder \
        .appName(appName) \
        .enableHiveSupport() \
        .getOrCreate()

    sc = spark.sparkContext
    sqlContext = spark

    print("Setting LOG LEVEL as ERROR")
    sc.setLogLevel("ERROR")
Esempio n. 2
0
    start_time = time.time()
    if 'backlog' in kwargs.keys() and kwargs.get('backlog'):
        print("run_weekly() - Backlog -- Running MA")
        build_prediction_weekly(sc=sc, sqlContext=sqlContext, _model_bld_date_string=_model_bld_date_string,
                                backlog_run=True, _bottlers=_bottlers)
    else:
        build_prediction_weekly(sc=sc, sqlContext=sqlContext, _model_bld_date_string=_model_bld_date_string,
                                _bottlers=_bottlers)
    print("Time taken for running WEEKLY MODELS:\t\t--- %s seconds ---" % (time.time() - start_time))


if __name__ == "__main__":
    ####################################################################################################################

    # Getting Current Date Time for AppName
    appName = "_".join([MODEL_BUILDING, "WEEKLY", get_current_date()])
    ####################################################################################################################

    # conf = SparkConf()

    spark = SparkSession \
        .builder \
        .appName(appName) \
        .enableHiveSupport() \
        .getOrCreate()

    sc = spark.sparkContext
    sqlContext = spark

    print ("Setting LOG LEVEL as ERROR")
    sc.setLogLevel("ERROR")
Esempio n. 3
0
    # from run_distributed_arima import _run_dist_arima
    # from run_distributed_prophet import _run_dist_prophet
    # from run_moving_average import _run_moving_average_weekly
    # from support_func import assign_category, get_current_date
    # from transform_data.spark_dataframe_func import final_select_dataset
    # from properties import MODEL_BUILDING, weekly_pdt_cat_123_location, weekly_pdt_cat_7_location
    # from pyspark.sql.functions import *
    # from transform_data.data_transform import string_to_gregorian
    # from support_func import get_current_date, get_sample_customer_list
    # import properties as p

    ###################################################################################################################

    # Getting Current Date Time for AppName
    appName_Weekly = "_".join([MODEL_BUILDING, "W", get_current_date()])
    ####################################################################################################################

    # conf = SparkConf().setAppName(appName)
    #
    # sc = SparkContext(conf=conf)
    # sqlContext = HiveContext(sparkContext=sc)

    spark = SparkSession \
        .builder \
        .config("spark.sql.warehouse.dir",
                "wasb://[email protected]/user/sshuser/spark-warehouse") \
        .appName(appName_Weekly) \
        .enableHiveSupport() \
        .getOrCreate()
Esempio n. 4
0
"""
author: rajarshi
"""
from pyspark import SparkContext, SparkConf
from pyspark.sql import HiveContext
from pyspark.sql.functions import *
from support_func import _generate_invoice, get_order_dates_between
from transform.support_func import _get_visit_list_from_invoice
from support_func import get_current_date
from properties import *

appName = "_".join(["CSO_INVOICE_GEN_", get_current_date()])
conf = SparkConf() \
    .setAppName(appName) \
    .set("spark.io.compression.codec", "snappy")

sc = SparkContext(conf=conf)
sqlContext = HiveContext(sparkContext=sc)

print "Setting LOG LEVEL as ERROR"
sc.setLogLevel("ERROR")

print "Add cso.zip to system path"
import sys
sys.path.insert(0, "cso.zip")

# Obtain visit list from invoice data. TODO : This will be replaced.
_visit_list = _get_visit_list_from_invoice(sqlContext=sqlContext,
                                           start_date=START_DATE_INVOICE,
                                           end_date=END_DATE_INVOICE)
Esempio n. 5
0
    from support_func import get_current_date, get_sample_customer_list, date_check, \
        get_sample_customer_list_new_addition, get_previous_sundays
    from properties import MODEL_BUILDING
    import properties as p
    from run_weekly import run_weekly
    from run_monthly import run_monthly

    ####################################################################################################################
    mdl_bld_date_string = ["".join(sys.argv[1])]
    _model_bld_date_string_stg = mdl_bld_date_string[0]

    # Getting Current Date Time for AppName
    appName = "_".join([
        MODEL_BUILDING, "SIMULATION_RUN::Dated::", _model_bld_date_string_stg,
        get_current_date()
    ])
    ####################################################################################################################

    # conf = SparkConf()

    spark = SparkSession \
        .builder \
        .appName(appName) \
        .enableHiveSupport() \
        .getOrCreate()

    sc = spark.sparkContext
    sqlContext = spark

    print("Setting LOG LEVEL as ERROR")
    print("Adding forecaster.zip to system path")
    sys.path.insert(0, "forecaster.zip")
    ####################################################################################################################

    from support_func import get_current_date, get_sample_customer_list, date_check
    from properties import MODEL_BUILDING
    import properties as p
    # from run_weekly import run_weekly
    # from run_monthly import run_monthly
    from run_weekly_baseline import run_weekly_baseline
    from run_monthly_baseline import run_monthly_baseline

    ####################################################################################################################

    # Getting Current Date Time for AppName
    appName = "_".join([MODEL_BUILDING, get_current_date()])
    ####################################################################################################################

    # conf = SparkConf()

    spark = SparkSession \
        .builder \
        .appName(appName) \
        .enableHiveSupport() \
        .getOrCreate()

    sc = spark.sparkContext
    sqlContext = spark

    print ("Setting LOG LEVEL as ERROR")
    sc.setLogLevel("ERROR")
    ###################################################################################################################

    # from run_distributed_prophet_monthly import _run_dist_prophet_monthly
    # from run_moving_average import _run_moving_average_monthly
    # from support_func import assign_category, get_current_date, _get_last_day_of_previous_month
    # from properties import MODEL_BUILDING, monthly_pdt_cat_456_location, monthly_pdt_cat_8910_location
    # from pyspark.sql.functions import *
    # from transform_data.data_transform import string_to_gregorian
    # from support_func import get_current_date, get_sample_customer_list
    # import properties as p

    ###################################################################################################################

    # Getting Current Date Time for AppName
    appName_Monthly = "_".join([MODEL_BUILDING, "M", get_current_date()])
    ####################################################################################################################

    # conf = SparkConf().setAppName(appName)
    #
    # sc = SparkContext(conf=conf)
    # sqlContext = HiveContext(sparkContext=sc)

    spark = SparkSession \
        .builder \
        .config("spark.sql.warehouse.dir",
                "wasb://[email protected]/user/sshuser/spark-warehouse") \
        .appName(appName_Monthly) \
        .enableHiveSupport() \
        .getOrCreate()