def purge_rt_update(self, config): func_name = "purge_rt_update" connector = config["connector"] logger = logging.LoggerAdapter(logging.getLogger(__name__), extra={"connector": connector}) logger.debug("purge realtime update for %s", connector) lock_name = make_kirin_lock_name(func_name, connector) with get_lock(logger, lock_name, app.config[str("REDIS_LOCK_TIMEOUT_PURGE")]) as locked: if not locked: logger.warning("%s for %s is already in progress", func_name, connector) return until = datetime.date.today() - datetime.timedelta( days=int(config["nb_days_to_keep"])) logger.info("purge realtime update for {} until {}".format( connector, until)) # TODO: we want to purge on "contributor" later, not "connector". RealTimeUpdate.remove_by_connectors_until(connectors=[connector], until=until) logger.info("%s for %s is finished", func_name, connector)
def _purge(process_name: str, contributor: Contributor, delete_rows_func: Callable[[Contributor], Dict]) -> Dict: logger = logging.LoggerAdapter(logging.getLogger(process_name), extra={"contributor": contributor.id}) logger.debug("preparing '{}' for contributor '{}'".format( process_name, contributor.id)) lock_name = make_kirin_lock_name(process_name, contributor.id) with get_lock(logger, lock_name, app.config["REDIS_LOCK_TIMEOUT_PURGE"]) as locked: if not locked: logger.warning( "operation '{}' for contributor '{}' is already in progress". format(process_name, contributor.id)) return {} start_datetime = datetime.datetime.utcnow() logger.info("executing '{}' for contributor '{}'".format( process_name, contributor.id)) purge_count = delete_rows_func(contributor) duration = (datetime.datetime.utcnow() - start_datetime).total_seconds() logger.warning( "operation '{}' for contributor '{}' is finished: purged {} row(s) in {} s" .format(process_name, contributor.id, purge_count, duration)) new_relic.record_custom_parameter( "{}:{}_count".format(contributor.id, process_name), purge_count) new_relic.record_custom_parameter( "{}:{}_duration".format(contributor.id, process_name), duration) return purge_count
def gtfs_poller(self, config): func_name = "gtfs_poller" contributor = ( model.Contributor.query_existing() .filter_by(id=config.get("contributor"), connector_type=ConnectorType.gtfs_rt.value) .first() ) logger = logging.LoggerAdapter(logging.getLogger(__name__), extra={"contributor": contributor.id}) lock_name = make_kirin_lock_name(func_name, contributor.id) with get_lock(logger, lock_name, app.config[str("REDIS_LOCK_TIMEOUT_POLLER")]) as locked: if not locked or not config.get("feed_url"): new_relic.ignore_transaction() return retrieval_interval = config.get("retrieval_interval", 10) if _is_last_call_too_recent(func_name, contributor.id, retrieval_interval): # do nothing if the last call is too recent new_relic.ignore_transaction() return logger.debug("polling of %s", config.get("feed_url")) # We do a HEAD request at the very beginning of polling and we compare it with the previous one to check if # the gtfs-rt is changed. # If the HEAD request or Redis get/set fail, we just ignore this part and do the polling anyway if not _is_newer(config): new_relic.ignore_transaction() manage_db_no_new(connector_type=ConnectorType.gtfs_rt.value, contributor_id=contributor.id) return try: response = _retrieve_gtfsrt(config) response.raise_for_status() except Exception as e: manage_db_error( data="", connector_type=ConnectorType.gtfs_rt.value, contributor_id=contributor.id, error="Http Error", is_reprocess_same_data_allowed=True, ) logger.debug(six.text_type(e)) return wrap_build(KirinModelBuilder(contributor), response.content) logger.info("%s for %s is finished", func_name, contributor.id)
def gtfs_poller(self, config): func_name = 'gtfs_poller' logger = logging.LoggerAdapter(logging.getLogger(__name__), extra={'contributor': config['contributor']}) logger.debug('polling of %s', config['feed_url']) contributor = config['contributor'] lock_name = make_kirin_lock_name(func_name, contributor) with get_lock(logger, lock_name, app.config['REDIS_LOCK_TIMEOUT_POLLER']) as locked: if not locked: new_relic.ignore_transaction() return # We do a HEAD request at the very beginning of polling and we compare it with the previous one to check if # the gtfs-rt is changed. # If the HEAD request or Redis get/set fail, we just ignore this part and do the polling anyway if not _is_newer(config): new_relic.ignore_transaction() manage_db_no_new(connector='gtfs-rt', contributor=contributor) return try: response = requests.get(config['feed_url'], timeout=config.get('timeout', 1)) response.raise_for_status() except Exception as e: manage_db_error(data='', connector='gtfs-rt', contributor=contributor, status='KO', error='Http Error') logger.debug(str(e)) return nav = navitia_wrapper.Navitia(url=config['navitia_url'], token=config['token'], timeout=5, cache=redis, query_timeout=app.config.get('NAVITIA_QUERY_CACHE_TIMEOUT', 600), pubdate_timeout=app.config.get('NAVITIA_PUBDATE_CACHE_TIMEOUT', 600))\ .instance(config['coverage']) proto = gtfs_realtime_pb2.FeedMessage() try: proto.ParseFromString(response.content) except DecodeError: manage_db_error(proto, 'gtfs-rt', contributor=contributor, status='KO', error='Decode Error') logger.debug('invalid protobuf') else: model_maker.handle(proto, nav, contributor) logger.info('%s for %s is finished', func_name, contributor)
def attempt_worker_launch(self): with get_lock(self.logger, self.lock_name, WORKER_REDIS_TIMEOUT_LOCK.total_seconds()) as locked: if locked: should_wait = True try: should_wait = self.process() except Exception as e: self.logger.warning("{0} worker died: {1}".format( self.connector_type.value, e)) finally: self.try_db_commit() if should_wait: time.sleep( QUEUE_WORKER_CONF_RELOAD_INTERVAL.total_seconds()) else: self.logger.debug("Lock {%s} not acquired (already taken)", self.lock_name) time.sleep(QUEUE_WORKER_CONF_RELOAD_INTERVAL.total_seconds())
def purge_trip_update(self, config): func_name = "purge_trip_update" contributor = config["contributor"] logger = logging.LoggerAdapter(logging.getLogger(__name__), extra={"contributor": contributor}) logger.debug("purge trip update for %s", contributor) lock_name = make_kirin_lock_name(func_name, contributor) with get_lock(logger, lock_name, app.config[str("REDIS_LOCK_TIMEOUT_PURGE")]) as locked: if not locked: logger.warning("%s for %s is already in progress", func_name, contributor) return until = datetime.date.today() - datetime.timedelta( days=int(config["nb_days_to_keep"])) logger.info("purge trip update for {} until {}".format( contributor, until)) TripUpdate.remove_by_contributors_and_period( contributors=[contributor], start_date=None, end_date=until) logger.info("%s for %s is finished", func_name, contributor)
def poll(self, model_builder): contributor_id = self.config.get("contributor") logger = logging.LoggerAdapter(logging.getLogger(__name__), extra={"contributor": contributor_id}) logger.debug("start polling") lock_name = make_kirin_lock_name(self.name, contributor_id) logger.debug("try getting lock") with get_lock(logger, lock_name, app.config["REDIS_LOCK_TIMEOUT_POLLER"]) as locked: logger.debug("locked acquired") if not locked or not self.config.get("feed_url"): logger.debug("lock is not aquired, ignoring the transaction") new_relic.ignore_transaction() return if self._is_last_call_too_recent(): # do nothing if the last call is too recent logger.debug("last call was too recent, ignoring the transaction") new_relic.ignore_transaction() return logger.debug("polling of %s", self.config.get("feed_url")) self._poll_feed(model_builder) logger.debug("%s for %s is finished", self.name, contributor_id)