コード例 #1
0
ファイル: tasks.py プロジェクト: patochectp/kirin
def purge_rt_update(self, config):
    func_name = "purge_rt_update"
    connector = config["connector"]

    logger = logging.LoggerAdapter(logging.getLogger(__name__),
                                   extra={"connector": connector})
    logger.debug("purge realtime update for %s", connector)

    lock_name = make_kirin_lock_name(func_name, connector)
    with get_lock(logger, lock_name,
                  app.config[str("REDIS_LOCK_TIMEOUT_PURGE")]) as locked:
        if not locked:
            logger.warning("%s for %s is already in progress", func_name,
                           connector)
            return

        until = datetime.date.today() - datetime.timedelta(
            days=int(config["nb_days_to_keep"]))
        logger.info("purge realtime update for {} until {}".format(
            connector, until))

        # TODO:  we want to purge on "contributor" later, not "connector".
        RealTimeUpdate.remove_by_connectors_until(connectors=[connector],
                                                  until=until)
        logger.info("%s for %s is finished", func_name, connector)
コード例 #2
0
ファイル: tasks.py プロジェクト: woshilapin/kirin
def _purge(process_name: str, contributor: Contributor,
           delete_rows_func: Callable[[Contributor], Dict]) -> Dict:
    logger = logging.LoggerAdapter(logging.getLogger(process_name),
                                   extra={"contributor": contributor.id})
    logger.debug("preparing '{}' for contributor '{}'".format(
        process_name, contributor.id))

    lock_name = make_kirin_lock_name(process_name, contributor.id)
    with get_lock(logger, lock_name,
                  app.config["REDIS_LOCK_TIMEOUT_PURGE"]) as locked:
        if not locked:
            logger.warning(
                "operation '{}' for contributor '{}' is already in progress".
                format(process_name, contributor.id))
            return {}

        start_datetime = datetime.datetime.utcnow()
        logger.info("executing '{}' for contributor '{}'".format(
            process_name, contributor.id))

        purge_count = delete_rows_func(contributor)
        duration = (datetime.datetime.utcnow() -
                    start_datetime).total_seconds()
        logger.warning(
            "operation '{}' for contributor '{}' is finished: purged {} row(s) in {} s"
            .format(process_name, contributor.id, purge_count, duration))
        new_relic.record_custom_parameter(
            "{}:{}_count".format(contributor.id, process_name), purge_count)
        new_relic.record_custom_parameter(
            "{}:{}_duration".format(contributor.id, process_name), duration)
        return purge_count
コード例 #3
0
def gtfs_poller(self, config):
    func_name = "gtfs_poller"
    contributor = (
        model.Contributor.query_existing()
        .filter_by(id=config.get("contributor"), connector_type=ConnectorType.gtfs_rt.value)
        .first()
    )

    logger = logging.LoggerAdapter(logging.getLogger(__name__), extra={"contributor": contributor.id})

    lock_name = make_kirin_lock_name(func_name, contributor.id)
    with get_lock(logger, lock_name, app.config[str("REDIS_LOCK_TIMEOUT_POLLER")]) as locked:
        if not locked or not config.get("feed_url"):
            new_relic.ignore_transaction()
            return

        retrieval_interval = config.get("retrieval_interval", 10)
        if _is_last_call_too_recent(func_name, contributor.id, retrieval_interval):
            # do nothing if the last call is too recent
            new_relic.ignore_transaction()
            return

        logger.debug("polling of %s", config.get("feed_url"))

        # We do a HEAD request at the very beginning of polling and we compare it with the previous one to check if
        # the gtfs-rt is changed.
        # If the HEAD request or Redis get/set fail, we just ignore this part and do the polling anyway
        if not _is_newer(config):
            new_relic.ignore_transaction()
            manage_db_no_new(connector_type=ConnectorType.gtfs_rt.value, contributor_id=contributor.id)
            return

        try:
            response = _retrieve_gtfsrt(config)
            response.raise_for_status()
        except Exception as e:
            manage_db_error(
                data="",
                connector_type=ConnectorType.gtfs_rt.value,
                contributor_id=contributor.id,
                error="Http Error",
                is_reprocess_same_data_allowed=True,
            )
            logger.debug(six.text_type(e))
            return

        wrap_build(KirinModelBuilder(contributor), response.content)
        logger.info("%s for %s is finished", func_name, contributor.id)
コード例 #4
0
def gtfs_poller(self, config):
    func_name = 'gtfs_poller'
    logger = logging.LoggerAdapter(logging.getLogger(__name__), extra={'contributor': config['contributor']})
    logger.debug('polling of %s', config['feed_url'])

    contributor = config['contributor']
    lock_name = make_kirin_lock_name(func_name, contributor)
    with get_lock(logger, lock_name, app.config['REDIS_LOCK_TIMEOUT_POLLER']) as locked:
        if not locked:
            new_relic.ignore_transaction()
            return

        # We do a HEAD request at the very beginning of polling and we compare it with the previous one to check if
        # the gtfs-rt is changed.
        # If the HEAD request or Redis get/set fail, we just ignore this part and do the polling anyway
        if not _is_newer(config):
            new_relic.ignore_transaction()
            manage_db_no_new(connector='gtfs-rt', contributor=contributor)
            return

        try:
            response = requests.get(config['feed_url'], timeout=config.get('timeout', 1))
            response.raise_for_status()

        except Exception as e:
            manage_db_error(data='', connector='gtfs-rt', contributor=contributor,
                            status='KO', error='Http Error')
            logger.debug(str(e))
            return

        nav = navitia_wrapper.Navitia(url=config['navitia_url'],
                                      token=config['token'],
                                      timeout=5,
                                      cache=redis,
                                      query_timeout=app.config.get('NAVITIA_QUERY_CACHE_TIMEOUT', 600),
                                      pubdate_timeout=app.config.get('NAVITIA_PUBDATE_CACHE_TIMEOUT', 600))\
            .instance(config['coverage'])

        proto = gtfs_realtime_pb2.FeedMessage()
        try:
            proto.ParseFromString(response.content)
        except DecodeError:
            manage_db_error(proto, 'gtfs-rt', contributor=contributor, status='KO', error='Decode Error')
            logger.debug('invalid protobuf')
        else:
            model_maker.handle(proto, nav, contributor)
            logger.info('%s for %s is finished', func_name, contributor)
コード例 #5
0
ファイル: __init__.py プロジェクト: woshilapin/kirin
 def attempt_worker_launch(self):
     with get_lock(self.logger, self.lock_name,
                   WORKER_REDIS_TIMEOUT_LOCK.total_seconds()) as locked:
         if locked:
             should_wait = True
             try:
                 should_wait = self.process()
             except Exception as e:
                 self.logger.warning("{0} worker died: {1}".format(
                     self.connector_type.value, e))
             finally:
                 self.try_db_commit()
                 if should_wait:
                     time.sleep(
                         QUEUE_WORKER_CONF_RELOAD_INTERVAL.total_seconds())
         else:
             self.logger.debug("Lock {%s} not acquired (already taken)",
                               self.lock_name)
             time.sleep(QUEUE_WORKER_CONF_RELOAD_INTERVAL.total_seconds())
コード例 #6
0
ファイル: tasks.py プロジェクト: patochectp/kirin
def purge_trip_update(self, config):
    func_name = "purge_trip_update"
    contributor = config["contributor"]
    logger = logging.LoggerAdapter(logging.getLogger(__name__),
                                   extra={"contributor": contributor})
    logger.debug("purge trip update for %s", contributor)

    lock_name = make_kirin_lock_name(func_name, contributor)
    with get_lock(logger, lock_name,
                  app.config[str("REDIS_LOCK_TIMEOUT_PURGE")]) as locked:
        if not locked:
            logger.warning("%s for %s is already in progress", func_name,
                           contributor)
            return
        until = datetime.date.today() - datetime.timedelta(
            days=int(config["nb_days_to_keep"]))
        logger.info("purge trip update for {} until {}".format(
            contributor, until))

        TripUpdate.remove_by_contributors_and_period(
            contributors=[contributor], start_date=None, end_date=until)
        logger.info("%s for %s is finished", func_name, contributor)
コード例 #7
0
ファイル: __init__.py プロジェクト: woshilapin/kirin
    def poll(self, model_builder):
        contributor_id = self.config.get("contributor")

        logger = logging.LoggerAdapter(logging.getLogger(__name__), extra={"contributor": contributor_id})

        logger.debug("start polling")
        lock_name = make_kirin_lock_name(self.name, contributor_id)
        logger.debug("try getting lock")
        with get_lock(logger, lock_name, app.config["REDIS_LOCK_TIMEOUT_POLLER"]) as locked:
            logger.debug("locked acquired")
            if not locked or not self.config.get("feed_url"):
                logger.debug("lock is not aquired, ignoring the transaction")
                new_relic.ignore_transaction()
                return

            if self._is_last_call_too_recent():
                # do nothing if the last call is too recent
                logger.debug("last call was too recent, ignoring the transaction")
                new_relic.ignore_transaction()
                return

            logger.debug("polling of %s", self.config.get("feed_url"))
            self._poll_feed(model_builder)
            logger.debug("%s for %s is finished", self.name, contributor_id)