Exemplo n.º 1
0
def connect(host="localhost", port=5672, usr="******", pas="******"):
    logger = get_logger(name=curr_module_name())
    try:
        logger.info("Connecting to rabbit: %s:%s@%s", usr, pas, host)
        credentials = PlainCredentials(usr, pas)
        parameters = ConnectionParameters(host, port, "/", credentials)
        conn = BlockingConnection(parameters)
    except Exception as e:
        logger.error("Exception connecting to rabbit: %s:%s@%s", usr, pas, host, exc_info=True)
        conn = None
    return conn
Exemplo n.º 2
0
def stop_ampq_channels():
    worker_stopping = 1
    logger = get_logger(name=curr_module_name())
    for ch in channels:
        logger.info('\tStopping worker (%d)...', worker_stopping)

        if ch.is_open:
            try:
                ch.close()
            except Exception as e:
                logger.error('Exception closing channel', exc_info=True)
        logger.info('\tStopped worker (%d).', worker_stopping)
        worker_stopping += 1
Exemplo n.º 3
0
def connect_with_retries(host="localhost", port=5672, usr="******", pas="******"):
    delay = 0.5
    retries = 0
    while retries < MAX_RECONNECT_RETRIES:
        try:
            conn = connect(host=host, port=port, usr=usr, pas=pas)
            ch = conn.channel()
            if conn and conn.is_open:
                return ch
        except Exception as e:
            logger = get_logger(name=curr_module_name())
            logger.error("############ error Connecting to rabbit: %s:%s@%s", usr, pas, host, exc_info=True)
            time.sleep(delay)
            delay *= 2
            retries += 1
    raise Exception("Couldn't connect to rabbit-{}:{}@{}. {} retries".format(usr, pas, host, MAX_RECONNECT_RETRIES))
Exemplo n.º 4
0
def start_responses_recvr(host='127.0.0.1', port=5672, usr='******', pas='******', stats=None,
                          msgs_dir=None, trace_err_only=False, verbose=False):
    logger = get_logger(name=curr_module_name())

    def count_message_processed():
        global pending_tasks
        pending_tasks = -1 if pending_tasks == -1 else pending_tasks - 1

        if pending_tasks == 0:
            logger.info("Processed all messages... exiting.")
            stats.sumaryPrettyPrint()
            stats.closeCsvFile()
            sys.exit()
        else:  # pending_tasks < 0 -> infinite. > 0 is the nr msgs pending
            logger.debug('-- still msgs_pending: %d', get_pending_nr())

    def handle_response(response_msg):
        msg = json.loads(response_msg.body.decode('utf-8'))
        logger.debug(">>>> response received: %s from queue %s correlation_id: %d \
                      pending_msgs: %d is_error: %s",
                     threading.current_thread().name, TASK_RESPONSES_POOL,
                     msg['correlation_id'], get_pending_nr(), str(is_error(msg)))
        if verbose:
            logger.info('---------------------------------------- MSG:')
            for key, value in msg.items():
                logger.info('\t%s:-> %s', key, str(value))
            logger.info('---------------------------------------------')

        stats.trackMsg(msg)

        if msgs_dir and (not trace_err_only or is_error(msg)):
            trace_msg(msgs_dir, msg)

        response_msg.ack()

        count_message_processed()

    curr_th_name = threading.current_thread().name

    logger.info(">> Starting receiver %s connecting to rabbitmq: %s:%s@%s",
                curr_th_name, usr, pas, host)

    subscriber = init_subscriber(host=host, port=port, usr=usr, pas=pas)
    subscriber.subscribe(handle_response)
Exemplo n.º 5
0
def connect_and_declare(host="localhost", port=5672, usr="******", pas="******", destinations=None):
    """ connects to RabbitMQ and does queue/exchange declarations
        destinations: name(s) of destinations, can be str or list
    """
    if not destinations:
        destinations = [TASK_REQUESTS_POOL, TASK_RESPONSES_POOL]
    elif isinstance(destinations, basestring):
        destinations = [destinations]

    ch = connect_with_retries(host=host, port=port, usr=usr, pas=pas)

    for destination in destinations:
        try:
            declare_and_bind(ch, destination, routing_key="#")
        except pika.exceptions.ChannelClosed as e:
            logger = get_logger(name=curr_module_name())
            logger.warning("Destination with name=%s already exists, error. Skipping", destination, exc_info=True)
            ch = connect_with_retries(host=host, port=port, usr=usr, pas=pas)

    return ch
Exemplo n.º 6
0
channels = []  # stores all executor thread channels.
stop = False  # False until the executor is asked to stop
MB_10 = 10485760

DEFAULT_DESTINATION = DestinationNames.get_for(TASK_REQUESTS_POOL)


if __name__ == '__main__':
    parser = argparse.ArgumentParser(description=globals()['__doc__'], add_help=True)
    parser.add_argument('--host', default='127.0.0.1', dest='host')
    parser.add_argument('--port', default=5672, dest='port', type=int)
    parser.add_argument('--user', default='guest', dest='usr')
    parser.add_argument('--pass', default='guest', dest='pas')
    parser.add_argument('--workers', default=1, dest='workers', type=int)
    parser.add_argument('--tasks', default=-1, dest='tasks_nr', type=int)
    parser.add_argument('--max-retries', default=0, dest='max_retries', type=int)
    parser.add_argument('--verbose', action='store_true', dest='verbose')
    parser.add_argument('--queue', default=DEFAULT_DESTINATION, dest='queue')

    register_signals_handling()

    if len(sys.argv) == 1:
        parser.print_help()
        sys.exit(1)

    args = parser.parse_args()
    start_executors(args.workers, args.host, args.port, args.usr, args.pas,  queue=args.queue,
                    tasks_nr=args.tasks_nr, max_retries=args.max_retries, verbose=args.verbose)
    get_logger(name=curr_module_name()).info('Executor exiting now.')
Exemplo n.º 7
0
def log_exc(txt):
    logger = get_logger(name=curr_module_name())
    logger.error(txt)
Exemplo n.º 8
0
    args = parser.parse_args()

    set_msgs_to_process(args.tasks)

    csvAutoSave = args.stats_csv_filename is not None

    if args.msgs_dir:
        init_dir(args.msgs_dir)
    if args.stats_csv_filename:
        init_dir(os.path.dirname(args.stats_csv_filename))

    global stats
    stats = TaskStatistics(csvAuto=csvAutoSave, csvFileName=args.stats_csv_filename)

    if args.stats_interval > 0:  # print stats every stats_interval seconds
        logger = get_logger(name=curr_module_name())
        logger.info('>>>>> args.stats_interval: %d', args.stats_interval)

        def print_stats_every(interval):
            while True:
                time.sleep(interval)
                stats.sumaryPrettyPrint()

        # daemon=True is not supported by python 2.7
        stats_th = threading.Thread(target=print_stats_every,
                                    args=(args.stats_interval,),
                                    name='stats_th')
        stats_th.daemon = True
        stats_th.start()

    for x in range(0, args.workers):
Exemplo n.º 9
0
def start_executor(host='127.0.0.1', port=5672, usr='******', pas='******', queue=DEFAULT_DESTINATION,
                   tasks_nr=1, max_retries=0, verbose=False, custom_callback=None,
                   ok_returncodes=(0, )):
    curr_th_name = threading.current_thread().name
    logger = get_logger(name=curr_module_name())
    logger.info(">> Starting executor %s connecting to rabbitmq: %s:%s@%s for executing %d tasks.",
                curr_th_name, usr, pas, host, tasks_nr)

    ch = connect_and_declare(host=host, port=port, usr=usr, pas=pas, destinations=queue)
    ch.basic_qos(prefetch_count=1)  # consume msgs one at a time

    channels.append(ch)

    def is_ok_returncode(code):
        return code in ok_returncodes

    def create_response_for(msg):
        resp = {}
        resp.update(msg)
        resp['executor_name'] = get_executor_name()
        resp['retries'] = msg.get('retries', 0)
        resp['max_retries'] = msg.get('max_retries', 0)
        return resp

    def should_retry(response_msg):
        is_error = response_msg['returncode'] != 0
        returncode = response_msg['returncode']
        is_retriable = returncode not in response_msg['non_retriable']
        current_retries = response_msg.get('retries', 0)
        msg_max_retries = response_msg.get('max_retries', max_retries)
        retries_pending = current_retries < msg_max_retries
        return is_error and is_retriable and retries_pending

    def send_response(response_msg):
        if should_retry(response_msg):
            logger.debug('---- retrying msg correlation_id: %d current_retries: %d of %d',
                         response_msg['correlation_id'], response_msg['retries'],
                         response_msg['max_retries'])
            tgt_exch = TASK_REQUESTS_POOL
            response_msg['retries'] += 1
        else:
            tgt_exch = TASK_RESPONSES_POOL

        response_str = json.dumps(response_msg)
        props = pika.BasicProperties(
                             delivery_mode=2,  # make message persistent
                          )
        ch.basic_publish(exchange=tgt_exch, routing_key='', body=response_str, properties=props)

    def tasks_nr_generator(tasks_nr):
        tasks_nr_gen = tasks_nr
        while True:
            tasks_nr_gen = tasks_nr_gen - 1 if tasks_nr_gen > 0 else tasks_nr_gen
            yield tasks_nr_gen

    def trace_msg(msg, context_info=''):
        logger.info(u'------------- MSG: %s', context_info)
        for key, value in msg.items():
            try:
                logger.info(u'\t%s:-> %s', key, value)
            except Exception as e:
                logger.error('error in trace message::::', exc_info=True)
                sys.exit(777)
        logger.info('---------------------------------------------')

    def execute_command(command):
        p = subprocess.Popen(command, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
        o, e = p.communicate()

        return p.returncode, o.decode('utf-8'), e.decode('utf-8')

    def handle_message(ch, method, properties, body):
        msg = json.loads(body.decode('utf-8'))
        logger.debug(">>>> msg received: %s from queue %s : correlation_id %d command: %s",
                     curr_th_name, queue, msg['correlation_id'], msg['command'][:50])
        try:
            response_msg = create_response_for(msg)
            response_msg['pre_command_ts'] = currtimemillis()

            returncode, out, err = command_callback(msg['command'])

            response_msg['post_command_ts'] = currtimemillis()
            response_msg['returncode'] = returncode
            response_msg['stdout'] = out
            response_msg['stderr'] = err

        except Exception as exc:
            logger.error('**** Command execution error. Exception for : correlation_id %d ',
                         response_msg['correlation_id'], exc_info=True)
            response_msg['post_command_ts'] = currtimemillis()
            response_msg['returncode'] = -3791
            response_msg['stdout'] = 'Exception trying to execute command.'
            response_msg['stderr'] = repr(exc)
        finally:
            if verbose or not is_ok_returncode(response_msg['returncode']):
                trace_msg(msg, context_info='Request msg: '+str(msg['correlation_id']))
                resp_header = 'Response msg: '+str(response_msg['correlation_id']) + \
                              ' returncode: '+str(response_msg['returncode'])
                trace_msg(response_msg, context_info=resp_header)
            send_response(response_msg)
            ch.basic_ack(method.delivery_tag)

        tasks_nr_new_elem = next(tasks_nr_gen)

        logger.debug("<<<< executed by: executor %s correlation_id: %d pending: %d",
                     curr_th_name, response_msg['correlation_id'], tasks_nr_new_elem)

        if tasks_nr_new_elem == 0:
            logger.info('==== no more tasks to execute. Exiting.')
            stop_and_exit()

    tasks_nr_gen = tasks_nr_generator(tasks_nr)

    command_callback = custom_callback if custom_callback else execute_command

    ch.basic_consume(handle_message, queue=queue, no_ack=False)
    logger.info("<< Ready: executor %s connected to rabbitmq: %s:%s@%s",
                curr_th_name, usr, pas, host)
    ch.start_consuming()
Exemplo n.º 10
0
 def signal_handler(signal, frame):
     logger = get_logger(name=curr_module_name())
     logger.info('Received signal: %s', str(signal))
     stop_and_exit()
Exemplo n.º 11
0
def stop_and_exit():
    stop_ampq_channels()
    logger = get_logger(name=curr_module_name())
    logger.info('Stopped all AMQP channels.')
    global stop
    stop = True