Example #1
0
 def post(self):
     json_data = request.get_json(force=True)
     app_context.init()
     app_context.application_context = json_data
     log_info(
         "Resource Layout_Detector_WF  Layout_Detector service started",
         app_context.application_context)
     task_id = str("LD-" + str(time.time()).replace('.', '')[0:13])
     task_starttime = eval(str(time.time()).replace('.', '')[0:13])
     #json_data = request.get_json(force = True)
     try:
         error_validator = ValidationResponse(DOWNLOAD_FOLDER)
         if error_validator.format_error(json_data) is True:
             response_gen = Response(json_data, DOWNLOAD_FOLDER)
             response = response_gen.workflow_response(
                 task_id, task_starttime)
             log_info(
                 "Resource Layout_Detector_WF Layout_Detector api response completed",
                 app_context.application_context)
             return jsonify(response)
     except FormatError as e:
         log_error(
             "Resource Layout_Detector_WF Input json format is not correct or dict_key is missing",
             app_context.application_context, e)
         return Status.ERR_request_input_format.value
Example #2
0
def processRequest(data):
    file_ops = FileOperation()
    producer_tok = Producer(config.bootstrap_server)
    DOWNLOAD_FOLDER = file_ops.file_download(config.download_folder)
    task_id = str("ANNO-" + str(time.time()).replace('.', ''))
    task_starttime = str(time.time()).replace('.', '')
    input_params, workflow_id, jobid, tool_name, step_order = file_ops.json_input_format(
        data)

    log_info("processing -- received message " + str(jobid), data)

    try:
        response_gen = Response(data, DOWNLOAD_FOLDER)
        result_response = response_gen.workflow_response(
            task_id, task_starttime)

        if result_response != None:
            if "errorID" not in result_response.keys():
                push_output(producer_tok, config.output_topic, result_response,
                            jobid, task_id, data)
                log_info(
                    "processing completed successfully, published at %s" %
                    (config.output_topic), data)
            else:
                log_info("processing failed, informed WFM", data)

    except Exception as e:
        log_exception("exception encountered ", LOG_WITHOUT_CONTEXT, e)
Example #3
0
def vision_ocr_request_worker():
    file_ops            = FileOperation()
    DOWNLOAD_FOLDER     = file_ops.create_file_download_dir(config.download_folder)
    producer_tok        = Producer(config.bootstrap_server)
    log_info("vision_ocr_request_worker : starting thread ", LOG_WITHOUT_CONTEXT)

    while True:
        data            = processQueue.get(block=True)
        #################
        task_id         = str("vision_ocr" + str(time.time()).replace('.', ''))
        ###################
        task_starttime  = str(time.time()).replace('.', '')
        input_files, workflow_id, jobid, tool_name, step_order = file_ops.json_input_format(data)
        
        log_info("vision_ocr_request_worker processing -- received message "+str(jobid), data)

        try:
            response_gen    = Response(data, DOWNLOAD_FOLDER)

            file_value_response = response_gen.workflow_response(task_id, task_starttime, False)
            if file_value_response != None:
                if "errorID" not in file_value_response.keys():
                    push_output(producer_tok, config.output_topic, file_value_response, jobid, task_id,data)
                    log_info("vision_ocr_request_worker : response send to topic %s"%(config.output_topic), LOG_WITHOUT_CONTEXT)
                else:
                    log_info("vision_ocr_request_worker : error send to error handler", data)

            log_info('vision_ocr_request_worker - request in internal queue {}'.format(Queue.qsize()), data)

            processQueue.task_done()
        except Exception as e:
            log_exception("vision_ocr_request_worker ",  LOG_WITHOUT_CONTEXT, e)

        controlQueue.put(1)
Example #4
0
def word_detector_request_worker():
    file_ops = FileOperation()
    DOWNLOAD_FOLDER = file_ops.create_file_download_dir(config.download_folder)
    producer_tok = Producer(config.bootstrap_server)
    log_info("word_detector_request_worker : starting thread ",
             LOG_WITHOUT_CONTEXT)

    while True:
        data = Queue.get(block=True)
        #################
        task_id = str("word_detector" + str(time.time()).replace('.', ''))
        ###################
        task_starttime = str(time.time()).replace('.', '')
        input_files, workflow_id, jobid, tool_name, step_order = file_ops.json_input_format(
            data)

        log_info(
            "word_detector_request_worker processing -- received message " +
            str(jobid), data)

        try:
            response_gen = Response(data, DOWNLOAD_FOLDER)

            file_value_response = response_gen.workflow_response(
                task_id, task_starttime, False)
            if file_value_response != None:
                push_output(producer_tok, config.output_topic,
                            file_value_response, jobid, task_id, data)
                log_info(
                    "word_detector_request_worker : response send to topic %s"
                    % (config.output_topic), LOG_WITHOUT_CONTEXT)
            else:
                erro_obj = {
                    'code': 400,
                    'jobID': jobid,
                    'message': "Word detector failed"
                }
                producer_tok.push_data_to_queue(
                    config.KAFKA_ANUVAAD_ETL_WF_ERROR_TOPIC, erro_obj)

                log_info(
                    "word_detector_request_worker : error send to error handler",
                    data)

            log_info(
                'word_detector_request_worker - request in internal queue {}'.
                format(Queue.qsize()), data)

            Queue.task_done()
        except Exception as e:
            log_exception("word_detector_request_worker ", LOG_WITHOUT_CONTEXT,
                          e)
Example #5
0
    def post(self):
        json_data = request.get_json(force=True)
        app_context.init()
        app_context.application_context = json_data
        log_info("Resource Vision_OCR service started", app_context.application_context)

        try:
            error_validator = ValidationResponse(DOWNLOAD_FOLDER)
            if error_validator.format_error(json_data) is True:
                response_gen = Response(json_data, DOWNLOAD_FOLDER)
                response = response_gen.nonwf_response()
                log_info("Resource Vision_OCR api response completed", app_context.application_context)
                return jsonify(response)
        except FormatError as e:
            log_error("Resource Vision_OCR Input json format is not correct or dict_key is missing", app_context.application_context, e)
            return Status.ERR_request_input_format.value
Example #6
0
 def post(self):
     log_info("Resource BlockMerger", "Block-merger service started", None)
     json_data = request.get_json(force=True)
     try:
         error_validator = ValidationResponse(DOWNLOAD_FOLDER)
         if error_validator.format_error(json_data) is True:
             response_gen = Response(json_data, DOWNLOAD_FOLDER)
             response = response_gen.nonwf_response()
             log_info("Resource BlockMerger",
                      "Block-merger api response completed", None)
             return jsonify(response)
     except FormatError as e:
         log_error(
             "Resource BlockMerger",
             "Input json format is not correct or dict_key is missing",
             None, e)
         return Status.ERR_request_input_format.value
Example #7
0
def process_merger_kf():
    file_ops = FileOperation()
    DOWNLOAD_FOLDER = file_ops.create_file_download_dir(config.download_folder)
    task_id = str("BM-" + str(time.time()).replace('.', ''))
    task_starttime = str(time.time()).replace('.', '')
    # instatiation of consumer for respective topic
    try:
        consumer_class = Consumer(config.input_topic, config.bootstrap_server)
        consumer = consumer_class.consumer_instantiate()
        log_info("process_merger_kf", "trying to receive value from consumer",
                 None)
        thread_instance = 0
        for msg in consumer:
            try:
                data = msg.value
                task_id = str("BM-" + str(time.time()).replace('.', ''))
                task_starttime = str(time.time()).replace('.', '')
                input_files, workflow_id, jobid, tool_name, step_order = file_ops.json_input_format(
                    data)
                log_info("process_merger_kf", "kafka request arrived ", jobid)
                response_gen = Response(data, DOWNLOAD_FOLDER)
                t1 = threading.Thread(
                    target=response_gen.multi_thred_block_merger,
                    args=(task_id, task_starttime, jobid),
                    name='BM-thread-' + str(thread_instance))
                t1.start()
                thread_instance += 1
                log_info("multithread", "block-merger running on multithread",
                         None)
                '''
                file_value_response = response_gen.workflow_response(task_id, task_starttime)
                if "errorID" not in file_value_response.keys():
                    producer = Producer()
                    producer.push_data_to_queue(config.output_topic, file_value_response, jobid, task_id)
                else:
                    log_info("process_merger_kf", "error send to error handler", jobid)'''
            except Exception as e:
                log_exception("process_pdf_kf",
                              "exception while consuming the records", jobid,
                              e)

    except KafkaConsumerError as e:
        response_custom = CustomResponse(Status.ERR_STATUS.value, None, None)
        response_custom.status_code['message'] = str(e)
        file_ops.error_handler(response_custom.status_code,
                               "KAFKA_CONSUMER_ERROR", True)
        log_exception("process_pdf_kf", "Consumer didn't instantiate", None, e)
    except KafkaProducerError as e:
        response_custom = e.code
        response_custom['message'] = e.message
        file_ops.error_handler(response_custom, "KAFKA_PRODUCER_ERROR", True)
        log_exception("process_pdf_kf",
                      "response send to topic %s" % (config.output_topic),
                      response_custom['jobID'], e)
Example #8
0
 def post(self):
     log_info("Resource BlockMergerWF", "Block-merger service started",
              None)
     task_id = str("BM-" + str(time.time()).replace('.', ''))
     task_starttime = str(time.time()).replace('.', '')
     json_data = request.get_json(force=True)
     try:
         error_validator = ValidationResponse(DOWNLOAD_FOLDER)
         if error_validator.format_error(json_data) is True:
             response_gen = Response(json_data, DOWNLOAD_FOLDER)
             response = response_gen.workflow_response(
                 task_id, task_starttime)
             log_info("Resource BlockMergerWF",
                      "Block-merger api response completed", None)
             return jsonify(response)
     except FormatError as e:
         log_error(
             "Resource BlockMergerWF",
             "Input json format is not correct or dict_key is missing",
             None, e)
         return Status.ERR_request_input_format.value
Example #9
0
def block_merger_request_worker():
    file_ops = FileOperation()
    DOWNLOAD_FOLDER = file_ops.create_file_download_dir(config.download_folder)
    producer_tok = Producer(config.bootstrap_server)

    while True:
        data = blockMergerQueue.get(block=True)
        task_id = str("BM-" + str(time.time()).replace('.', ''))
        task_starttime = str(time.time()).replace('.', '')
        input_files, workflow_id, jobid, tool_name, step_order = file_ops.json_input_format(
            data)

        log_info(
            "block_merger_request_worker processing -- received message " +
            str(jobid), data)

        response_gen = Response(data, DOWNLOAD_FOLDER)

        file_value_response = response_gen.workflow_response(
            task_id, task_starttime, False)
        if file_value_response != None:
            if "errorID" not in file_value_response.keys():
                push_output(producer_tok, config.output_topic,
                            file_value_response, jobid, task_id)
                log_info(
                    "process_block_merger_kf : response send to topic %s" %
                    (config.output_topic), None)
            else:
                log_info(
                    "process_block_merger_kf : error send to error handler",
                    jobid)

        log_info(
            'block_merger_request_worker - request in internal queue {}'.
            format(blockMergerQueue.qsize()), jobid)

        blockMergerQueue.task_done()