def execute(self): """ Associate amps and write unassoc_<datetime>_dat.xml. """ install_path, data_path = get_config_paths() amp_handler = AmplitudeHandler(install_path, data_path) event = amp_handler.getEvent(self._eventid) if event is None: # # This shouldn't ever happen, but the code is here just # in case it does # datadir = os.path.join(data_path, self._eventid, 'current') if not os.path.isdir(datadir): raise NotADirectoryError('%s is not a valid directory.' % datadir) eventxml = os.path.join(datadir, 'event.xml') if not os.path.isfile(eventxml): raise FileNotFoundError('%s does not exist.' % eventxml) origin = read_event_file(eventxml) event = { 'id': self._eventid, 'netid': origin['netid'], 'network': origin['network'], 'time': origin['time'].strftime(constants.TIMEFMT), 'lat': origin['lat'], 'lon': origin['lon'], 'depth': origin['depth'], 'mag': origin['mag'], 'locstring': origin['locstring'] } amp_handler.insertEvent(event) amp_handler.associateOne(self._eventid, pretty_print=True)
def execute(self): """ Assemble ShakeMap input data and write and ShakeMapInputContainer named shake_data.hdf in the event's 'current' directory. Raises: NotADirectoryError: When the event data directory does not exist. FileNotFoundError: When the the event's event.xml file does not exist. RuntimeError: When there are problems parsing the configuration. ValidateError: When there are configuration items missing or mis- configured. """ install_path, data_path = get_config_paths() datadir = os.path.join(data_path, self._eventid, 'current') if not os.path.isdir(datadir): raise NotADirectoryError('%s is not a valid directory.' % datadir) eventxml = os.path.join(datadir, 'event.xml') self.logger.debug('Looking for event.xml file...') if not os.path.isfile(eventxml): raise FileNotFoundError('%s does not exist.' % eventxml) # Prompt for a comment string if none is provided on the command line if self.comment is None: if sys.stdout is not None and sys.stdout.isatty(): self.comment = input( 'Please enter a comment for this version.\n' 'comment: ') else: self.comment = '' # find any source.txt or moment.xml files momentfile = os.path.join(datadir, 'moment.xml') sourcefile = os.path.join(datadir, 'source.txt') if not os.path.isfile(sourcefile): sourcefile = None if not os.path.isfile(momentfile): momentfile = None # # Clear away results from previous runs # products_path = os.path.join(datadir, 'products') if os.path.isdir(products_path): shutil.rmtree(products_path, ignore_errors=True) pdl_path = os.path.join(datadir, 'pdl') if os.path.isdir(pdl_path): shutil.rmtree(pdl_path, ignore_errors=True) # Look for any .transferred file and delete it save_file = os.path.join(datadir, SAVE_FILE) if os.path.isfile(save_file): os.remove(save_file) # # Look for global configs in install_path/config # spec_file = get_configspec() validator = get_custom_validator() self.logger.debug('Looking for configuration files...') modules = ConfigObj( os.path.join(install_path, 'config', 'modules.conf'), configspec=spec_file) gmpe_sets = ConfigObj( os.path.join(install_path, 'config', 'gmpe_sets.conf'), configspec=spec_file) global_config = ConfigObj( os.path.join(install_path, 'config', 'model.conf'), configspec=spec_file) # # this is the event specific model.conf (may not be present) # prefer model.conf to model_select.conf # event_config_file = os.path.join(datadir, 'model.conf') event_config_zc_file = os.path.join(datadir, 'model_select.conf') if os.path.isfile(event_config_file): event_config = ConfigObj(event_config_file, configspec=spec_file) elif os.path.isfile(event_config_zc_file): event_config = ConfigObj(event_config_zc_file, configspec=spec_file) else: event_config = ConfigObj() # # start merging event_config # global_config.merge(event_config) global_config.merge(modules) global_config.merge(gmpe_sets) results = global_config.validate(validator) if not isinstance(results, bool) or not results: config_error(global_config, results) check_config(global_config, self.logger) global_data_path = os.path.join(os.path.expanduser('~'), 'shakemap_data') # # If there is a prediction_location->file file, then we need # to expand any macros; this could have the event ID, so we # can't just use the file_type handler in the configspec # if 'file' in global_config['interp']['prediction_location']: loc_file = global_config['interp']['prediction_location']['file'] if loc_file and loc_file != 'None': # 'None' is a string here loc_file = path_macro_sub(loc_file, ip=install_path, dp=data_path, gp=global_data_path, ei=self._eventid) if not os.path.isfile(loc_file): raise FileNotFoundError("prediction file '%s' is not " "a valid file" % loc_file) global_config['interp']['prediction_location']['file'] = \ loc_file config = global_config.dict() self.logger.debug('Looking for data files...') datafiles = glob.glob(os.path.join(datadir, '*_dat.xml')) if os.path.isfile(os.path.join(datadir, 'stationlist.xml')): datafiles.append(os.path.join(datadir, 'stationlist.xml')) datafiles += glob.glob(os.path.join(datadir, '*_dat.json')) if os.path.isfile(os.path.join(datadir, 'stationlist.json')): datafiles.append(os.path.join(datadir, 'stationlist.json')) self.logger.debug('Looking for rupture files...') # look for geojson versions of rupture files rupturefile = os.path.join(datadir, 'rupture.json') if not os.path.isfile(rupturefile): # failing any of those, look for text file versions rupturefiles = glob.glob(os.path.join(datadir, '*_fault.txt')) rupturefile = None if len(rupturefiles): rupturefile = rupturefiles[0] # # Sort out the version history. Get the most recent backup file and # extract the existing history. Then add a new line for this run. # timestamp = datetime.datetime.utcnow().strftime('%FT%TZ') originator = config['system']['source_network'] backup_dirs = sorted( glob.glob(os.path.join(datadir, '..', 'backup*')), reverse=True) if len(backup_dirs): # # Backup files exist so find the latest one and extract its # history, then add a new line that increments the version # bu_file = os.path.join(backup_dirs[0], 'shake_data.hdf') bu_ic = ShakeMapInputContainer.load(bu_file) history = bu_ic.getVersionHistory() bu_ic.close() version = int( backup_dirs[0].replace( os.path.join(datadir, '..', 'backup'), '')) version += 1 new_line = [timestamp, originator, version, self.comment] history['history'].append(new_line) elif os.path.isfile(os.path.join(datadir, 'shake_data.hdf')): # # No backups are available, but there is an existing shake_data # file. Extract its history and update the timestamp and # source network (but leave the version alone). # If there is no history, just start a new one with version 1 # bu_file = os.path.join(datadir, 'shake_data.hdf') bu_ic = ShakeMapInputContainer.load(bu_file) history = bu_ic.getVersionHistory() bu_ic.close() if 'history' in history: new_line = [timestamp, originator, history['history'][-1][2], self.comment] history['history'][-1] = new_line else: history = {'history': []} new_line = [timestamp, originator, 1, self.comment] history['history'].append(new_line) else: # # No backup and no existing file. Make this version 1 # history = {'history': []} new_line = [timestamp, originator, 1, self.comment] history['history'].append(new_line) hdf_file = os.path.join(datadir, 'shake_data.hdf') self.logger.debug('Creating input container...') shake_data = ShakeMapInputContainer.createFromInput( hdf_file, config, eventxml, history, rupturefile=rupturefile, sourcefile=sourcefile, momentfile=momentfile, datafiles=datafiles) self.logger.debug('Created HDF5 input container in %s' % shake_data.getFileName()) ah = AmplitudeHandler(install_path, data_path) event = ah.getEvent(self._eventid) if event is None: origin = shake_data.getRuptureObject().getOrigin() event = {'id': self._eventid, 'netid': origin.netid, 'network': origin.network, 'time': origin.time.strftime(constants.TIMEFMT), 'lat': origin.lat, 'lon': origin.lon, 'depth': origin.depth, 'mag': origin.mag, 'locstring': origin.locstring} ah.insertEvent(event) shake_data.close()
def test_amps(): try: install_path, data_path = get_config_paths() # dbfile location homedir = os.path.dirname(os.path.abspath(__file__)) dbfile = os.path.join(homedir, '..', '..', 'data', 'install', 'data', 'amps.db') if os.path.isfile(dbfile): os.remove(dbfile) handler = AmplitudeHandler(install_path, data_path) # test inserting events into the database event = { 'id': 'ci37889959', 'netid': 'ci', 'network': '', 'time': datetime(2018, 3, 7, 18, 5, 0).strftime(constants.TIMEFMT), 'lat': 35.487, 'lon': -120.027, 'depth': 8.0, 'locstring': 'Somewhere in California', 'mag': 3.7 } handler.insertEvent(event) info = handler.getStats() assert info['events'] == 1 # Try getting and updating an event event_out = handler.getEvent('ci37889959') current_time = time.time() del event_out['network'] event_out['repeats'] = [1, 2, 3] event_out['lastrun'] = current_time handler.insertEvent(event_out, update=True) event_out = handler.getEvent('ci37889959') assert event_out['network'] == '' assert set(event_out['repeats']) == set([1, 2, 3]) assert event_out['lastrun'] == current_time homedir = os.path.dirname(os.path.abspath(__file__)) xmlfile = os.path.join(homedir, '..', '..', 'data', 'ampdata', 'USR_100416_20180307_180450.xml') handler.insertAmps(xmlfile) info = handler.getStats() assert info['events'] == 1 assert info['stations'] == 1 assert info['station_min'] == datetime(2018, 3, 7, 18, 4, 49, tzinfo=timezone.utc) assert info['station_max'] == datetime(2018, 3, 7, 18, 4, 49, tzinfo=timezone.utc) assert info['channels'] == 3 assert info['pgms'] == 15 eqtime = timestr_to_timestamp(event['time']) eqlat = event['lat'] eqlon = event['lon'] df = handler.associate(eqtime, eqlat, eqlon) vsum = 0 for row in df: if row[2] == 'pga': vsum += row[3] np.testing.assert_almost_equal(vsum, 0.010621814475025483) # get repeats repeats = handler.getRepeats() assert repeats[0][0] == 'ci37889959' assert set(repeats[0][2]) == set([1, 2, 3]) # delete event handler.deleteEvent('ci37889959') info = handler.getStats() assert info['events'] == 0 assert handler.getEvent('ci37889959') is None del handler os.remove(dbfile) # test global associator event2 = { 'id': 'ci0000001', 'netid': 'ci', 'network': '', 'time': datetime(2000, 1, 1, 0, 0, 1).strftime(constants.TIMEFMT), 'lat': 37.487, 'lon': -122.027, 'depth': 8.0, 'locstring': 'Somewhere in California', 'mag': 4.7 } handler = AmplitudeHandler(install_path, data_path) handler.insertEvent(event) handler.insertEvent(event2) handler.insertAmps(xmlfile) associated = handler.associateAll(pretty_print=True) assert len(associated) == 1 # Do an association bad event and an old event count = handler.associateOne('kk1234567') assert count == -1 count = handler.associateOne('ci0000001') assert count == 0 del handler os.remove(dbfile) shutil.rmtree(os.path.join(data_path, event['id'])) # test event associator; a lot of this is just for coverage of # various edge cases. xmlfile2 = os.path.join(homedir, '..', '..', 'data', 'ampdata', 'USR_100416_20180307_180450_2.xml') xmlfile3 = os.path.join(homedir, '..', '..', 'data', 'ampdata', 'USR_100416_20180307_180450_3.xml') xmlfile4 = os.path.join(homedir, '..', '..', 'data', 'ampdata', 'USR_100416_20180307_180450_4.xml') xmlfile5 = os.path.join(homedir, '..', '..', 'data', 'ampdata', 'USR_100416_20180307_180450_5.xml') xmlfile6 = os.path.join(homedir, '..', '..', 'data', 'ampdata', 'USR_100416_20180307_180450_6.xml') handler = AmplitudeHandler(install_path, data_path) handler.insertEvent(event) handler.insertEvent(event2) handler.insertAmps(xmlfile) handler.insertAmps(xmlfile2) handler.insertAmps(xmlfile3) handler.insertAmps(xmlfile4) handler.insertAmps(xmlfile5) handler.insertAmps(xmlfile6) associated = handler.associateOne(event['id'], pretty_print=False) assert associated == 30 del handler os.remove(dbfile) shutil.rmtree(os.path.join(data_path, event['id'])) # test clean methods handler = AmplitudeHandler(install_path, data_path) handler.insertEvent(event) handler.insertAmps(xmlfile) # Add another event with the alternate time encoding xmlfile = os.path.join(homedir, '..', '..', 'data', 'ampdata', 'TA109C_BH..2018_095_193003x.xml') handler.insertAmps(xmlfile) info = handler.getStats() assert info['stations'] == 2 handler.cleanEvents(threshold=1) handler.cleanAmps(threshold=1) info = handler.getStats() assert info['events'] == 0 assert info['stations'] == 0 assert info['channels'] == 0 assert info['pgms'] == 0 except Exception: assert 1 == 2 finally: if os.path.isfile(dbfile): os.remove(dbfile)
class Queue(object): def __init__(self, pargs): current_time = int(time.time()) self.MEMORY_UPDATE_TIME = current_time self.ASSOCIATE_UPDATE_TIME = current_time self.DB_MAINTENANCE_TIME = current_time self.children = {} self.attached = pargs.attached self.install_path, self.data_path = get_config_paths() self.config = get_config(self.install_path) # # Get shake.conf for the autorun modules # config_file = os.path.join(self.install_path, 'config', 'shake.conf') spec_file = get_configspec('shake') shake_config = ConfigObj(config_file, configspec=spec_file) results = shake_config.validate(Validator()) if not isinstance(results, bool) or not results: config_error(shake_config, results) self.shake_cmds = shlex.split(shake_config['autorun_modules']) # # Turn this process into a daemon # self.logpath = os.path.join(self.install_path, 'logs') if not os.path.isdir(self.logpath): os.makedirs(self.logpath) pidfile = os.path.join(self.logpath, 'queue.pid') self.filelock = lockfile.FileLock(pidfile) if self.filelock.is_locked(): if pargs.break_lock: self.filelock.break_lock() else: logger = self.getLogger() logger.error("pid lock file '%s' exists, can't start " "sm_queue; exiting..." % (pidfile)) sys.exit(-1) def queueMainLoop(self): context = daemon.DaemonContext(working_directory=self.data_path, pidfile=self.filelock) with self.getContext(context): self.logger = self.getLogger() # # Create the database for running and queued events. # self.eventQueue = EventQueue(self.install_path) # # Create the socket # qsocket = socket.socket(socket.AF_INET, socket.SOCK_STREAM) qsocket.bind(('', self.config['port'])) # Set a timeout so that we can occasionally look for other # things to do qsocket.settimeout(30) qsocket.listen(5) # # Get a connection to the event database # self.ampHandler = AmplitudeHandler(self.install_path, self.data_path) self.logger.info('sm_queue initiated') # # At startup we want to see what was running when this process # shut down and try to restart them. # running = self.eventQueue.getRunningEvents() for eventid, command in running: self.logger.info("Startup: Running event %s" % (eventid)) event = self.ampHandler.getEvent(eventid) # Update the XML because the DB may have newer information self.writeEventXml(event) p = subprocess.Popen(command) self.children[eventid] = {'popen': p, 'start_time': time.time()} while True: # # Do routine stuff # self.doPeriodicTasks() # # Now wait for a connection # try: (clientsocket, address) = qsocket.accept() except socket.timeout: # # Normal timeout; do routine tasks and then go # back to waiting for a connection # continue # # Got a connection # hostname, _, _ = socket.gethostbyaddr(address[0]) # hostname = socket.getfqdn(hostname) self.logger.info('Got connection from %s at port %s' % (hostname, address[1])) if hostname not in self.config['servers']: self.logger.warning('Connection from %s refused: not in ' 'valid servers list' % hostname) clientsocket.close() continue # # The accept() should guarantee that there's something # to read, but something could go wrong... # try: clientsocket.settimeout(5) data = clientsocket.recv(MAX_SIZE) except socket.timeout: self.logger.warning('Did not get data from connection, ' 'continuing') clientsocket.close() continue else: clientsocket.close() # # Decode the data and do something # try: cmd = json.loads(data.decode('utf8')) except json.decoder.JSONDecodeError: self.logger.warning("Couldn't decode data from %s: " "ignoring" % hostname) continue if not isinstance(cmd, dict) or 'type' not in cmd or \ 'data' not in cmd or 'id' not in cmd['data']: self.logger.warning('Bad data from %s: ignoring' % hostname) continue if cmd['type'] == 'origin': self.logger.info('Received "origin" for event %s' % cmd['data']['id']) if 'action' in cmd['data']: action = cmd['data']['action'] else: action = 'Origin received' self.processOrigin(cmd['data'], action) elif cmd['type'] == 'cancel': self.logger.info('Received "cancel" for event %s' % cmd['data']['id']) self.processCancel(cmd['data']) else: self.logger.info('Received "%s" for event %s' % cmd['type'], cmd['data']['id']) self.processOther(cmd['data'], cmd['type']) def doPeriodicTasks(self): """ Check for finished children and start any needed timed repeats. Returns: nothing: Nothing. """ # # Do routine stuff: # first check for repeats to queue # current_time = int(time.time()) repeats = self.ampHandler.getRepeats() for eventid, otime, rep_list in repeats: while rep_list is not None and rep_list[0] < current_time: event = self.ampHandler.getEvent(eventid) if eventid in self.children: # # Event is already running; pop this repeat and move on # rep_list.pop(0) if len(rep_list) == 0: rep_list = None event['repeats'] = rep_list self.ampHandler.insertEvent(event, update=True) continue rep_list.pop(0) if len(rep_list) == 0: rep_list = None event['repeats'] = rep_list self.ampHandler.insertEvent(event, update=True) if event['lastrun'] == 0: # This is a delayed first run self.logger.info('Queueing event %s after network delay' % eventid) self.dispatchEvent(event, 'Event added') else: self.logger.info('Queueing repeat of event %s' % eventid) self.dispatchEvent(event, 'Scheduled repeat') break # # Run the associator and dispatch events with new data # if self.config['associate_interval'] >= 0 and \ self. ASSOCIATE_UPDATE_TIME + \ self.config['associate_interval'] < current_time: self.ASSOCIATE_UPDATE_TIME = current_time self.associateAll() # # Reap any dead children, and then try to run queued events # _ = self.reapChildren() self.runQueuedEvents() # # Print memory usage once per hour to see how much we're leaking... # if self.MEMORY_UPDATE_TIME + 3600 < current_time: self.MEMORY_UPDATE_TIME = current_time process = psutil.Process(os.getpid()) mem = getattr(process.memory_full_info(), 'uss', 0) / 1048576.0 self.logger.info('Currently using %.1f MB' % mem) # # Do the occasional DB cleanup once per day; keep amps for 30 # days and events for 1 year # if self.DB_MAINTENANCE_TIME + 86400 < current_time: self.DB_MAINTENANCE_TIME = current_time # # First do the assocication to make sure we don't drop any # amps that might associate # if self.config['associate_interval'] >= 0: self.ASSOCIATE_UPDATE_TIME = current_time self.associateAll() # # Now clean out the amps and events # self.ampHandler.cleanAmps(threshold=30) self.ampHandler.cleanEvents(threshold=365) return def getLogger(self): """Set up a logger for this process. Returns: logging.logger: An instance of a logger. """ if not os.path.isdir(self.logpath): os.makedirs(self.logpath) logger = logging.getLogger('queue_logger') logger.setLevel(logging.INFO) if self.attached: handler = logging.StreamHandler() else: logfile = os.path.join(self.logpath, 'queue.log') handler = TimedRotatingFileHandler(logfile, when='midnight', backupCount=30) formatter = logging.Formatter( fmt='%(asctime)s - %(levelname)s - %(message)s', datefmt='%Y-%m-%d %H:%M:%S') handler.setFormatter(formatter) logger.addHandler(handler) logger.propagate = False return logger def associateAll(self): """Do the associateAll method of the the AmplitudeHandler and then process all of the events with updated data. Returns: nothing: Nothing. """ event_list = self.ampHandler.associateAll(pretty_print=True) for eventid in event_list: event = self.ampHandler.getEvent(eventid) self.processOrigin(event, 'Data association') return def writeEventXml(self, event): """ Create the event directory if it doesn't exist and write/re-write the event.xml file Args: event (dict): The event data structure. Returns: nothing: Nothing. """ ttemp = event['time'] try: dt = datetime.strptime(ttemp, constants.TIMEFMT) except ValueError: try: dt = datetime.strptime(ttemp, constants.ALT_TIMEFMT) except ValueError: self.logger.error("Can't parse input time %s" % ttemp) return event['time'] = dt event_dir = os.path.join(self.data_path, event['id'], 'current') if not os.path.isdir(event_dir): os.makedirs(event_dir) event_xml = os.path.join(event_dir, 'event.xml') self.logger.info('Writing event %s to event.xml' % (event['id'])) val = write_event_file(event, event_xml) if val: self.logger.error('Error writing event.xml: %s' % val) event['time'] = ttemp return def moveEventDirectory(self, oldid, newid): """Change the name of an existing event directory to a new ID. """ try: shutil.move(os.path.join(self.data_path, oldid), os.path.join(self.data_path, newid)) except shutil.Error as e: self.logger("Error trying to move data directory %s to %s: %s" % (oldid, newid, str(e))) return def processOrigin(self, event, action): """ Determine if an event should be processed (or reprocessed) and dispatch it for processing. Args: event (dict): The event data structure. action (str): The "type" of the trigger that caused this function to be called. Returns: nothing: Nothing. """ current_time = int(time.time()) force_run = False dispatch = True # # See if we already have this event, make a decision # existing = self.ampHandler.getEvent(event['id']) if existing is None and 'alt_eventids' in event: # # We haven't processed this ID, but the ID may have changed # for eid in event['alt_eventids'].split(','): if eid == event['id']: continue alt_exists = self.ampHandler.getEvent(eid) if alt_exists is None: continue # # We processed this event under a different ID # If the event is currently running with the old ID, kill it # if eid in self.children: self.children[eid]['popen'].kill() self.children[eid]['popen'].wait() del self.children[eid] self.eventQueue.deleteRunningEvent(eid) # Delete the old event from the database self.ampHandler.deleteEvent(eid) # Move the old event directory to the new ID self.moveEventDirectory(eid, event['id']) # Now treat the the new event ID as a new event. existing = None # But force it to run (because we want to update the event ID), # bypassing date and magnitude checks that come later force_run = True break if existing is None: # # This is a new event (or an event with a changed ID) # update = False # Do we want to run this event? if not force_run and self.magnitudeTooSmall(event): self.logger.info('Event %s (mag=%f) too small, skipping' % (event['id'], event['mag'])) return if not force_run and self.eventTooOldOrInFuture(event): self.logger.info('Event %s too old or too far in the future, ' 'skipping' % event['id']) return # # Looks like we'll be running this event, get the repeats # (if any) and toss the ones that have already passed # replist = None try: dt = datetime.strptime(event['time'], constants.TIMEFMT) except ValueError: try: dt = datetime.strptime(event['time'], constants.ALT_TIMEFMT) except ValueError: self.logger.error("Can't parse input time %s" % event['time']) return event_timestamp = int(dt.replace(tzinfo=timezone.utc).timestamp()) for mag in sorted(self.config['repeats'].keys(), reverse=True): if event['mag'] > mag: replist = [x + event_timestamp for x in self.config['repeats'][mag] if event_timestamp + x > current_time] break # # The first time we run an event, we need to check its # network ID against those in the delay list. If present, # we add the required delay as the first repeat, but # don't dispatch the event. If the delay time has already # passed, just treat this as a normal event # if event['netid'] in self.config['network_delays']: delay = self.config['network_delays'][event['netid']] if event_timestamp + delay > current_time: self.logger.info('Delaying processing event %s due to ' 'network delay configuration.' % (event['id'])) replist.insert(0, event_timestamp + delay) dispatch = False event['repeats'] = replist if len(replist) > 0 else None event['lastrun'] = 0 else: # # We've run this event before # update = True # # We want to update the event info in the database but # save the lastrun and repeats settings # event['lastrun'] = existing['lastrun'] event['repeats'] = copy.copy(existing['repeats']) # # Insert or update the event info in the database, then # possibly queue the event to be run. # self.ampHandler.insertEvent(event, update=update) if dispatch is True: self.dispatchEvent(event, action) return def processOther(self, data, action): """A trigger has been issued for an event. Treat this as an origin update. If the event in question is not in our database, ignore the message. Args: data (dict): The event information dictionary. action (str): The "type" of the trigger that caused this function to be called. Returns: nothing: Nothing. """ eventid = data['id'] existing = self.ampHandler.getEvent(eventid) if existing: self.processOrigin(existing, action) else: if 'alt_eventids' in data: for eid in data['alt_eventids'].split(','): if eid == eventid: continue existing = self.ampHandler.getEvent(eid) if existing: self.processOrigin(existing, action) return self.logger.info('Trigger of action "%s" is for unprocessed ' 'event %s: ignoring' % (action, data['id'])) return def processCancel(self, data): """We've received a cancellation of an event: run 'shake cancel'. Args: data (dict): The dictionary must have an event ID under the 'id' key. Returns: nothing: Nothing. """ eventid = data['id'] existing = self.ampHandler.getEvent(eventid) if existing: self.dispatchEvent(data, 'cancel') return if 'alt_eventids' in data: for eid in data['alt_eventids'].split(','): if eid == eventid: continue existing = self.ampHandler.getEvent(eid) if existing: self.dispatchEvent(existing, 'cancel') return self.logger.info('cancel is for unprocessed event %s: ignoring' % eventid) return def magnitudeTooSmall(self, event): """ Return False if the magnitude is greater than the threshold magnitude of the first metro box within which it falls, or the global minmag if it does not fall within a box; return true otherwise. Args: event (dict): The event dictionary; must contain at least "mag", "lon", and "lat" keys. Returns: bool: True if the event is too small to process; False otherwise. """ mag = event['mag'] lon = event['lon'] lat = event['lat'] pt = Point((lon, lat)) for boxname in sorted(self.config['boxes']): boxdict = self.config['boxes'][boxname] if pt.within(boxdict['poly']): if mag >= boxdict['mag']: return False else: return True # # Not in any boxes # if mag >= self.config['minmag']: return False return True def eventTooOldOrInFuture(self, event): """ Return True if the event is too old or too far in the future to process; return False otherwise. Args: event (dict): The event data structure. Returns: bool: True if the event is older than old_event_age or is more than future_event_age in the future; returns False otherwise. """ current_time = time.time() try: event_time = datetime.strptime(event['time'], constants.TIMEFMT).\ replace(tzinfo=timezone.utc).timestamp() except ValueError: event_time = datetime.strptime(event['time'], constants.ALT_TIMEFMT).\ replace(tzinfo=timezone.utc).timestamp() if self.config['old_event_age'] >= 0 and \ event_time + self.config['old_event_age'] < current_time: return True if self.config['future_event_age'] >= 0 and \ event_time - self.config['future_event_age'] > current_time: return True return False def dispatchEvent(self, event, action): """ Queue a run for the specified event. Args: event (dict): The data structure of the event to process. action (str): 'cancel', 'test', or some other string. 'cancel' starts the cancel process, 'test' queues the process 'echo eventid'. Any other string queues the shake process to be run at the next opportunity. See the configuration file 'queue.conf' for the exact commands that will be run. Returns: nothing: Nothing. """ eventid = event['id'] if action == 'cancel': # # Cancellations aren't queued, they're run immediately # self.logger.info('Canceling event %s' % eventid) if eventid in self.children: self.logger.info('Event %s is running; killing...' % eventid) self.children[eventid]['popen'].kill() self.children[eventid]['popen'].wait() del self.children[eventid] self.eventQueue.deleteRunningEvent(eventid) cmd = self.config['cancel_command'].replace( 'shake', self.config['shake_path']) cmd = cmd.replace('<EVID>', eventid) cmd = cmd.split() p = subprocess.Popen(cmd) self.children[eventid] = {'popen': p, 'start_time': time.time()} self.eventQueue.insertRunningEvent(eventid, cmd) return self.logger.info('Queueing event %s due to action "%s"' % (eventid, action)) # # Add the action as the assemble/augment comment, or replace the # comment if it is already there. # for ix, shcmd in enumerate(self.shake_cmds): if shcmd not in ['assemble', 'augment']: continue if len(self.shake_cmds) == ix + 1: # This shouldn't happen self.shake_cmds.append('-c') self.shake_cmds.append('"%s"' % action) elif self.shake_cmds[ix + 1] == '-c': self.shake_cmds[ix + 2] = '"%s"' % action else: self.shake_cmds.insert(ix + 1, '-c') self.shake_cmds.insert(ix + 2, '"%s"' % action) break if action == 'test': cmd = self.config['shake_command'].replace('shake', 'echo') else: cmd = self.config['shake_command'].replace( 'shake', self.config['shake_path']) cmd = cmd.replace('<EVID>', eventid) cmd = cmd.split() + self.shake_cmds self.eventQueue.queueEvent(eventid, cmd, event['mag']) return def runQueuedEvents(self): """If there is space, run events from the queue """ if len(self.children) >= self.config['max_subprocesses']: self.logger.info('Processing queue is full; waiting for open ' 'slots.') return current_time = int(time.time()) mtw = self.config['max_trigger_wait'] queued = self.eventQueue.getQueuedEvents() for eventid, command in queued: event = self.ampHandler.getEvent(eventid) if eventid in self.children: # # Event is currently running, don't run it but make sure # there's a repeat pretty soon # if event['repeats']: if event['repeats'][0] > current_time + mtw: event['repeats'].insert(0, current_time + mtw) else: event['repeats'] = [current_time + mtw] self.ampHandler.insertEvent(event, update=True) self.logger.info('Event %s is currently running, shelving ' 'this update' % event['id']) self.eventQueue.dequeueEvent(eventid) continue if event['repeats']: delta_t = current_time - event['repeats'][0] if delta_t > -mtw: # We're due for a rerun anyway, so just leave the # event queued self.logger.info('Event %s will repeat soon, shelving ' 'this update' % eventid) self.eventQueue.dequeueEvent(eventid) continue if current_time - event['lastrun'] < mtw: # # We ran this event very recently, but don't have a repeat # scheduled in the near future, so let's skip this one # but make sure something happens relatively soon # if event['repeats']: event['repeats'].insert(0, current_time + mtw) else: event['repeats'] = [current_time + mtw] self.ampHandler.insertEvent(event, update=True) self.logger.info('Event %s ran recently, shelving this ' 'update' % event['id']) self.eventQueue.dequeueEvent(eventid) continue self.logger.info("Running event %s" % (eventid)) # Update the XML because the DB may have newer information self.writeEventXml(event) p = subprocess.Popen(command) self.children[eventid] = {'popen': p, 'start_time': time.time()} self.eventQueue.dequeueEvent(eventid) self.eventQueue.insertRunningEvent(eventid, command) if len(self.children) >= self.config['max_subprocesses']: self.logger.info('Processing queue is full; waiting for open ' 'slots.') break return def reapChildren(self): """ Look through the list of child processes, reap the ones that have finished, and kill any that are taking too long. Returns: nothing: Nothing. Completed or killed child processes are removed from the list of children. """ to_delete = [] current_time = time.time() for eventid, info in self.children.items(): returncode = info['popen'].poll() if returncode is not None: self.logger.info('Reaped child for event %s (return code %d)' % (eventid, returncode)) event = self.ampHandler.getEvent(eventid) if event: event['lastrun'] = current_time self.ampHandler.insertEvent(event, update=True) to_delete.append(eventid) self.eventQueue.deleteRunningEvent(eventid) continue # # Kill children who take too long # if info['start_time'] + self.config['max_process_time'] < \ current_time: self.logger.warning('Event %s taking too long, killing' % eventid) info['popen'].kill() info['popen'].wait() self.logger.warning('Reaped child for killed event %s' % eventid) to_delete.append(eventid) self.eventQueue.deleteRunningEvent(eventid) for eventid in to_delete: del self.children[eventid] return to_delete def getContext(self, context): """Returns a context based on the value of the 'attached' argument. If attached is True, then the function returns an instance of the Dummycontext; if it is False the function returns the 'context' argument. Args: context (Context manager): A valid context manager. Returns: Context manager: If attached is True, the function returns an instance of the Dummycontext; if False, returns the 'context' argument. """ if self.attached: return Dummycontext() else: return context