class Controller(object): def __init__(self, config): self._config = config self.uve = SnmpUve(self._config) self._logger = self.uve.logger() self.sleep_time() self._keep_running = True def stop(self): self._keep_running = False def sleep_time(self, newtime=None): if newtime: self._sleep_time = newtime else: self._sleep_time = self._config.frequency() return self._sleep_time def get_net_devices(self): return list(self._config.devices()) def do_work(self, i): self._logger.debug('@do_work(%d):started...' % i) cdir = mkdtemp() input_file = os.path.join(cdir, 'in.data') output_file = os.path.join(cdir, 'out.data') with open(input_file, 'wb') as f: pickle.dump(dict(out=output_file, netdev=self.get_net_devices(), instance=i), f) f.flush() proc = subprocess.Popen('contrail-snmp-scanner --input %s' % ( input_file), shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) o,e = proc.communicate() self._logger.debug('@do_work(%d): scan done with %d\nstdout:' \ '\n%s\nstderr:\n%s\n' % (i, proc.returncode, o, e)) with open(output_file, 'rb') as f: d = pickle.load(f) os.unlink(input_file) os.unlink(output_file) os.rmdir(cdir) for dev, data in d.items(): self.uve.send(data['snmp']) self.uve.send_flow_uve({'name': dev, 'flow_export_source_ip': data['flow_export_source_ip']}) self._logger.debug('@do_work(%d):Done!' % i) def run(self): i = 0 while self._keep_running: self.do_work(i) gevent.sleep(self._sleep_time) i += 1
class Controller(object): def __init__(self, config): self._config = config self.uve = SnmpUve(self._config) self._hostname = socket.gethostname() self._logger = self.uve.logger() self.sleep_time() self._keep_running = True self.last = set() self._sem = None self._config.set_cb(self.notify) self._mnt = MaxNinTtime(3, self._sleep_time) self._state = 'full_scan' # replace it w/ fsm self._if_data = None # replace it w/ fsm self._cleanup = None self._members = None self._partitions = None self._prouters = None def _make_if_cdata(self, data): if_cdata = {} t = time.time() for dev in data: if 'snmp' in data[dev]: if 'ifMib' in data[dev]['snmp']: if 'ifTable' in data[dev]['snmp']['ifMib']: if_cdata[dev] = dict(map(lambda x: ( x['ifIndex'], (x['ifOperStatus'], t)), filter(lambda x: 'ifOperStatus' in x\ and 'ifDescr' in x, data[dev][ 'snmp']['ifMib']['ifTable']))) elif 'ifOperStatus' in data[dev]['snmp']: if_cdata[dev] = dict((k, (v, t)) for k, v in data[dev]['snmp']['ifOperStatus'].items()) return if_cdata def _set_status(self, _dict, dev, intf, val): if dev not in _dict: _dict[dev] = {} _dict[dev][intf] = val def _check_and_update_ttl(self, up2down): t = time.time() expry = 3 * self._fast_scan_freq for dev in self._if_data: for intf in self._if_data[dev]: if self._if_data[dev][intf][0] == 1: if t - self._if_data[dev][intf][1] > expry: self._set_status(up2down, dev, intf, 7) #no resp self._if_data[dev][intf] = (7, t) def _get_if_changes(self, if_cdata): down2up, up2down, others = {}, {}, {} for dev in if_cdata: if dev in self._if_data: for intf in if_cdata[dev]: if intf in self._if_data[dev]: if if_cdata[dev][intf][0] != self._if_data[dev][ intf][0]: if self._if_data[dev][intf][0] == 1: self._set_status(up2down, dev, intf, if_cdata[dev][intf][0]) elif if_cdata[dev][intf][0] == 1: self._set_status(down2up, dev, intf, if_cdata[dev][intf][0]) else: self._set_status(others, dev, intf, if_cdata[dev][intf][0]) self._if_data[dev][intf] = if_cdata[dev][intf] else: self._if_data[dev] = if_cdata[dev] for intf in self._if_data[dev]: if self._if_data[dev][intf][0] == 1: self._set_status(down2up, dev, intf, if_cdata[dev][intf][0]) else: self._set_status(others, dev, intf, if_cdata[dev][intf][0]) return down2up, up2down, others def _chk_if_change(self, data): if_cdata = self._make_if_cdata(data) down2up, up2down, others = self._get_if_changes(if_cdata) self._check_and_update_ttl(up2down) self._logger.debug('@chk_if_change: down2up(%s), up2down(%s), ' \ 'others(%s)' % (', '.join(down2up.keys()), ', '.join(up2down.keys()), ', '.join(others.keys()))) return down2up, up2down, others def _extra_call_params(self): if self._state != 'full_scan': return dict(restrict='ifOperStatus') return {} def _analyze(self, data): ret = True time = self._fast_scan_freq if self._state != 'full_scan': down2up, up2down, others = self._chk_if_change(data) if down2up: self._state = 'full_scan' time = self._mnt.add() elif self._mnt.ready4full_scan(): self._state = 'full_scan' time = 0 elif up2down: self.uve.send_ifstatus_update(self._if_data) ret = False sret = 'chngd: ' + self._state + ', time: ' + str(time) else: self._state = 'fast_scan' self._if_data = self._make_if_cdata(data) self.uve.send_ifstatus_update(self._if_data) sret = 'chngd: %d' % len(self._if_data) self._logger.debug('@do_work(analyze):State %s(%d)->%s!' % ( self._state, len(data), str(sret))) return ret, time def notify(self, svc, msg='', up=True, servers=''): self.uve.conn_state_notify(svc, msg, up, servers) def stop(self): self._keep_running = False def sleep_time(self, newtime=None): if newtime: self._sleep_time = newtime else: self._sleep_time = self._config.frequency() self._fast_scan_freq = self._config.fast_scan_freq() if self._fast_scan_freq > self._sleep_time / 2: self._fast_scan_freq = self._sleep_time / 2 return self._sleep_time def _setup_io(self): cdir = mkdtemp() input_file = os.path.join(cdir, 'in.data') output_file = os.path.join(cdir, 'out.data') return cdir, input_file, output_file def _create_input(self, input_file, output_file, devices, i, restrict=None): if isinstance(devices[0], DeviceDict): devices = DeviceConfig.populate_cfg(devices) with open(input_file, 'wb') as f: data = dict(out=output_file, netdev=devices, instance=i) if restrict: data['restrict'] = restrict pickle.dump(data, f) f.flush() def _run_scanner(self, input_file, output_file, i): proc = subprocess.Popen('contrail-snmp-scanner --input %s' % ( input_file), shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, close_fds=True) self._cleanup = (proc, output_file) o,e = proc.communicate() self._cleanup = None self._logger.debug('@run_scanner(%d): scan done with %d\nstdout:' \ '\n%s\nstderr:\n%s\n' % (i, proc.returncode, o, e)) with open(output_file, 'rb') as f: d = pickle.load(f) self._logger.debug('@run_scanner(%d): loaded %s' % (i, output_file)) return d def _cleanup_io(self, cdir, input_file, output_file): os.unlink(input_file) os.unlink(output_file) os.rmdir(cdir) def _send_uve(self, d): for dev, data in d.items(): if dev: self.uve.send(data['snmp']) self.uve.send_flow_uve({'name': dev, 'flow_export_source_ip': data['flow_export_source_ip']}) self.find_fix_name(data['name'], dev) self._logger.debug('@send_uve:Processed %d!' % (len(d))) def _send_snmp_collector_uve(self, members, partitions, prouters): snmp_collector_info = SnmpCollectorInfo() if self._members != members: self._members = members snmp_collector_info.members = members if self._partitions != partitions: self._partitions = partitions snmp_collector_info.partitions = partitions if self._prouters != prouters: self._prouters = prouters snmp_collector_info.prouters = prouters if snmp_collector_info != SnmpCollectorInfo(): snmp_collector_info.name = self._hostname SnmpCollectorUVE(data=snmp_collector_info).send() # end _send_snmp_collector_uve def _del_uves(self, l): with self._sem: for dev in l: self.uve.delete(dev) def do_work(self, i, devices): self._logger.debug('@do_work(%d):started (%d)...' % (i, len(devices))) sleep_time = self._fast_scan_freq if devices: with self._sem: self._work_set = devices cdir, input_file, output_file = self._setup_io() self._create_input(input_file, output_file, devices, i, **self._extra_call_params()) data = self._run_scanner(input_file, output_file, i) self._cleanup_io(cdir, input_file, output_file) do_send, sleep_time = self._analyze(data) if do_send: self._send_uve(data) gevent.sleep(0) del self._work_set self._logger.debug('@do_work(%d):Processed %d!' % (i, len(devices))) return sleep_time def find_fix_name(self, cfg_name, snmp_name): if snmp_name != cfg_name: self._logger.debug('@find_fix_name: snmp name %s differs from ' \ 'configured name %s, fixed for this run' % ( snmp_name, cfg_name)) for d in self._work_set: if d.name == cfg_name: d.name = snmp_name return def run(self): i = 0 self._sem = Semaphore() self._logger.debug('Starting.. %s' % str( self._config.zookeeper_server())) constnt_schdlr = ConsistentScheduler( self._config._name, zookeeper=self._config.zookeeper_server(), delete_hndlr=self._del_uves, logger=self._logger, cluster_id=self._config.cluster_id()) while self._keep_running: self._logger.debug('@run: ittr(%d)' % i) if constnt_schdlr.schedule(self._config.devices()): members = constnt_schdlr.members() partitions = constnt_schdlr.partitions() prouters = map(lambda x: x.name, constnt_schdlr.work_items()) self._send_snmp_collector_uve(members, partitions, prouters) sleep_time = self.do_work(i, constnt_schdlr.work_items()) self._logger.debug('done work %s' % str(prouters)) i += 1 gevent.sleep(sleep_time) else: gevent.sleep(1) constnt_schdlr.finish()
class Controller(object): def __init__(self, config): self._config = config self._config.random_collectors = self._config.collectors() self._chksum = "" if self._config.collectors(): self._chksum = hashlib.md5("".join(self._config.collectors())).hexdigest() self._config.random_collectors = random.sample(self._config.collectors(), \ len(self._config.collectors())) self.uve = SnmpUve(self._config) self._sandesh = self.uve.sandesh_instance() self._hostname = socket.gethostname() self._logger = self.uve.logger() self.sleep_time() self.last = set() self._sem = Semaphore() self._config.set_cb(self.notify) self._mnt = MaxNinTtime(3, self._sleep_time) self._state = 'full_scan' # replace it w/ fsm self._if_data = None # replace it w/ fsm self._cleanup = None self._members = None self._partitions = None self._prouters = {} self._config_handler = SnmpConfigHandler(self._sandesh, self._config.rabbitmq_params(), self._config.cassandra_params()) self._consistent_scheduler = ConsistentScheduler(self._config._name, zookeeper=self._config.zookeeper_server(), delete_hndlr=self._del_uves, logger=self._logger, cluster_id=self._config.cluster_id()) def _make_if_cdata(self, data): if_cdata = {} t = time.time() for dev in data: if 'snmp' in data[dev]: if 'ifMib' in data[dev]['snmp']: if 'ifTable' in data[dev]['snmp']['ifMib']: if_cdata[dev] = dict(map(lambda x: ( x['ifIndex'], (x['ifOperStatus'], t)), filter(lambda x: 'ifOperStatus' in x\ and 'ifDescr' in x, data[dev][ 'snmp']['ifMib']['ifTable']))) elif 'ifOperStatus' in data[dev]['snmp']: if_cdata[dev] = dict((k, (v, t)) for k, v in data[dev]['snmp']['ifOperStatus'].items()) return if_cdata def _delete_if_data(self, dev): if dev in self._if_data: del self._if_data[dev] def _set_status(self, _dict, dev, intf, val): if dev not in _dict: _dict[dev] = {} _dict[dev][intf] = val def _check_and_update_ttl(self, up2down): t = time.time() expry = 3 * self._fast_scan_freq for dev in self._if_data: for intf in self._if_data[dev]: if self._if_data[dev][intf][0] == 1: if t - self._if_data[dev][intf][1] > expry: self._set_status(up2down, dev, intf, 7) #no resp self._if_data[dev][intf] = (7, t) def _get_if_changes(self, if_cdata): down2up, up2down, others = {}, {}, {} for dev in if_cdata: if dev in self._if_data: for intf in if_cdata[dev]: if intf in self._if_data[dev]: if if_cdata[dev][intf][0] != self._if_data[dev][ intf][0]: if self._if_data[dev][intf][0] == 1: self._set_status(up2down, dev, intf, if_cdata[dev][intf][0]) elif if_cdata[dev][intf][0] == 1: self._set_status(down2up, dev, intf, if_cdata[dev][intf][0]) else: self._set_status(others, dev, intf, if_cdata[dev][intf][0]) self._if_data[dev][intf] = if_cdata[dev][intf] else: self._if_data[dev] = if_cdata[dev] for intf in self._if_data[dev]: if self._if_data[dev][intf][0] == 1: self._set_status(down2up, dev, intf, if_cdata[dev][intf][0]) else: self._set_status(others, dev, intf, if_cdata[dev][intf][0]) return down2up, up2down, others def _chk_if_change(self, data): if_cdata = self._make_if_cdata(data) down2up, up2down, others = self._get_if_changes(if_cdata) self._check_and_update_ttl(up2down) self._logger.debug('@chk_if_change: down2up(%s), up2down(%s), ' \ 'others(%s)' % (', '.join(down2up.keys()), ', '.join(up2down.keys()), ', '.join(others.keys()))) return down2up, up2down, others def _extra_call_params(self): if self._state != 'full_scan': return dict(restrict='ifOperStatus') return {} def _analyze(self, data): ret = True time = self._fast_scan_freq if self._state != 'full_scan': down2up, up2down, others = self._chk_if_change(data) if down2up: self._state = 'full_scan' time = self._mnt.add() elif self._mnt.ready4full_scan(): self._state = 'full_scan' time = 0 elif up2down: self.uve.send_ifstatus_update(self._if_data) ret = False sret = 'chngd: ' + self._state + ', time: ' + str(time) else: self._state = 'fast_scan' self._if_data = self._make_if_cdata(data) self.uve.send_ifstatus_update(self._if_data) sret = 'chngd: %d' % len(self._if_data) self._logger.debug('@do_work(analyze):State %s(%d)->%s!' % ( self._state, len(data), str(sret))) return ret, time def notify(self, svc, msg='', up=True, servers=''): self.uve.conn_state_notify(svc, msg, up, servers) def sleep_time(self, newtime=None): if newtime: self._sleep_time = newtime else: self._sleep_time = self._config.frequency() self._fast_scan_freq = self._config.fast_scan_freq() if self._fast_scan_freq > self._sleep_time / 2: self._fast_scan_freq = self._sleep_time / 2 return self._sleep_time def _setup_io(self): cdir = mkdtemp() input_file = os.path.join(cdir, 'in.data') output_file = os.path.join(cdir, 'out.data') return cdir, input_file, output_file def _create_input(self, input_file, output_file, devices, i, restrict=None): if isinstance(devices[0], DeviceDict): devices = DeviceConfig.populate_cfg(devices) with open(input_file, 'wb') as f: data = dict(out=output_file, netdev=devices, instance=i) if restrict: data['restrict'] = restrict pickle.dump(data, f) f.flush() def _run_scanner(self, input_file, output_file, i): proc = subprocess.Popen('contrail-snmp-scanner --input %s' % ( input_file), shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, close_fds=True) self._cleanup = (proc, output_file) o,e = proc.communicate() self._cleanup = None self._logger.debug('@run_scanner(%d): scan done with %d\nstdout:' \ '\n%s\nstderr:\n%s\n' % (i, proc.returncode, o, e)) with open(output_file, 'rb') as f: d = pickle.load(f) self._logger.debug('@run_scanner(%d): loaded %s' % (i, output_file)) return d def _cleanup_io(self, cdir, input_file, output_file): os.unlink(input_file) os.unlink(output_file) os.rmdir(cdir) def _send_uve(self, d): for dev, data in d.items(): if dev: self.uve.send(data['snmp']) self.uve.send_flow_uve({'name': dev, 'flow_export_source_ip': data['flow_export_source_ip']}) self.find_fix_name(data['name'], dev) self._logger.debug('@send_uve:Processed %d!' % (len(d))) def _send_snmp_collector_uve(self, members, partitions, prouters): snmp_collector_info = SnmpCollectorInfo() if self._members != members: self._members = members snmp_collector_info.members = members if self._partitions != partitions: self._partitions = partitions snmp_collector_info.partitions = partitions new_prouters = {p.name: p for p in prouters} if self._prouters.keys() != new_prouters.keys(): deleted_prouters = [v for p, v in self._prouters.iteritems() \ if p not in new_prouters] self._del_uves(deleted_prouters) self._prouters = new_prouters snmp_collector_info.prouters = self._prouters.keys() if snmp_collector_info != SnmpCollectorInfo(): snmp_collector_info.name = self._hostname SnmpCollectorUVE(data=snmp_collector_info).send() # end _send_snmp_collector_uve def _del_uves(self, l): with self._sem: for dev in l: self._delete_if_data(dev.name) self.uve.delete(dev) def do_work(self, i, devices): self._logger.debug('@do_work(%d):started (%d)...' % (i, len(devices))) sleep_time = self._fast_scan_freq if devices: with self._sem: self._work_set = devices cdir, input_file, output_file = self._setup_io() self._create_input(input_file, output_file, devices, i, **self._extra_call_params()) data = self._run_scanner(input_file, output_file, i) self._cleanup_io(cdir, input_file, output_file) do_send, sleep_time = self._analyze(data) if do_send: self._send_uve(data) gevent.sleep(0) del self._work_set self._logger.debug('@do_work(%d):Processed %d!' % (i, len(devices))) return sleep_time def find_fix_name(self, cfg_name, snmp_name): if snmp_name != cfg_name: self._logger.debug('@find_fix_name: snmp name %s differs from ' \ 'configured name %s, fixed for this run' % ( snmp_name, cfg_name)) for d in self._work_set: if d.name == cfg_name: d.name = snmp_name return def sighup_handler(self): if self._config._args.conf_file: config = ConfigParser.SafeConfigParser() config.read(self._config._args.conf_file) if 'DEFAULTS' in config.sections(): try: collectors = config.get('DEFAULTS', 'collectors') if type(collectors) is str: collectors = collectors.split() new_chksum = hashlib.md5("".join(collectors)).hexdigest() if new_chksum != self._chksum: self._chksum = new_chksum self._config.random_collectors = \ random.sample(collectors, len(collectors)) # Reconnect to achieve load-balance irrespective of list self.uve.sandesh_reconfig_collectors( self._config.random_collectors) except ConfigParser.NoOptionError as e: pass # end sighup_handler def _snmp_walker(self): i = 0 while True: self._logger.debug('@run: ittr(%d)' % i) devices = map(lambda e: DeviceDict(e[0].split(':')[-1], e[1].obj), self._config_handler.get_physical_routers()) if self._consistent_scheduler.schedule(devices): members = self._consistent_scheduler.members() partitions = self._consistent_scheduler.partitions() work_items = self._consistent_scheduler.work_items() self._send_snmp_collector_uve(members, partitions, work_items) sleep_time = self.do_work(i, work_items) self._logger.debug('done work %s' % str(self._prouters.keys())) i += 1 gevent.sleep(sleep_time) else: gevent.sleep(1) # end _snmp_walker def run(self): """ @sighup SIGHUP handler to indicate configuration changes """ gevent.signal(signal.SIGHUP, self.sighup_handler) self.gevs = [ gevent.spawn(self._config_handler.start), gevent.spawn(self._snmp_walker) ] try: gevent.joinall(self.gevs) except KeyboardInterrupt: self._logger.error('Exiting on ^C') except gevent.GreenletExit: self._logger.error('Exiting on gevent-kill') finally: self._logger.error('stopping everything!') self.stop() # end run def stop(self): self.uve.killall() l = len(self.gevs) for i in range(0, l): self._logger.error('killing %d of %d' % (i+1, l)) self.gevs[0].kill() self._logger.error('joining %d of %d' % (i+1, l)) self.gevs[0].join() self._logger.error('stopped %d of %d' % (i+1, l)) self.gevs.pop(0) self._consistent_scheduler.finish()
class Controller(object): def __init__(self, config): self._config = config self._me = socket.gethostname() + ':' + str(os.getpid()) self.uve = SnmpUve(self._config) self._logger = self.uve.logger() self.sleep_time() self._keep_running = True self.last = set() self._sem = None self._config.set_cb(self.notify) self._mnt = MaxNinTtime(3, self._sleep_time) self._state = 'full_scan' # replace it w/ fsm self._if_data = None # replace it w/ fsm self._cleanup = None def _make_if_cdata(self, data): if_cdata = {} t = time.time() for dev in data: if 'snmp' in data[dev]: if 'ifMib' in data[dev]['snmp']: if 'ifTable' in data[dev]['snmp']['ifMib']: if_cdata[dev] = dict(map(lambda x: ( x['ifIndex'], (x['ifOperStatus'], t)), filter(lambda x: 'ifOperStatus' in x\ and 'ifDescr' in x, data[dev][ 'snmp']['ifMib']['ifTable']))) elif 'ifOperStatus' in data[dev]['snmp']: if_cdata[dev] = dict((k, (v, t)) for k, v in data[dev]['snmp']['ifOperStatus'].items()) return if_cdata def _set_status(self, _dict, dev, intf, val): if dev not in _dict: _dict[dev] = {} _dict[dev][intf] = val def _check_and_update_ttl(self, up2down): t = time.time() expry = 3 * self._fast_scan_freq for dev in self._if_data: for intf in self._if_data[dev]: if self._if_data[dev][intf][0] == 1: if t - self._if_data[dev][intf][1] > expry: self._set_status(up2down, dev, intf, 7) #no resp def _get_if_changes(self, if_cdata): down2up, up2down, others = {}, {}, {} t = time.time() for dev in if_cdata: if dev in self._if_data: for intf in if_cdata[dev]: if intf in self._if_data[dev]: if if_cdata[dev][intf][0] != self._if_data[dev][ intf][0]: if self._if_data[dev][intf][0] == 1: self._set_status(up2down, dev, intf, if_cdata[dev][intf][0]) elif if_cdata[dev][intf][0] == 1: self._set_status(down2up, dev, intf, if_cdata[dev][intf][0]) else: self._set_status(others, dev, intf, if_cdata[dev][intf][0]) self._if_data[dev][intf] = if_cdata[dev][intf] else: self._if_data[dev] = if_cdata[dev] for intf in self._if_data[dev]: if self._if_data[dev][intf][0] == 1: self._set_status(down2up, dev, intf, if_cdata[dev][intf][0]) else: self._set_status(others, dev, intf, if_cdata[dev][intf][0]) return down2up, up2down, others def _chk_if_change(self, data): if_cdata = self._make_if_cdata(data) down2up, up2down, others = self._get_if_changes(if_cdata) self._check_and_update_ttl(up2down) self._logger.debug('@chk_if_change: down2up(%s), up2down(%s), ' \ 'others(%s)' % (', '.join(down2up.keys()), ', '.join(up2down.keys()), ', '.join(others.keys()))) return down2up, up2down, others def _extra_call_params(self): if self._state != 'full_scan': return dict(restrict='ifOperStatus') return {} def _analyze(self, data): ret = True time = self._fast_scan_freq if self._state != 'full_scan': down2up, up2down, others = self._chk_if_change(data) if down2up: self._state = 'full_scan' time = self._mnt.add() elif self._mnt.ready4full_scan(): self._state = 'full_scan' time = 0 elif up2down: self.uve.send_ifstatus_update(self._if_data) ret = False sret = 'chngd: ' + self._state + ', time: ' + str(time) else: self._state = 'fast_scan' self._if_data = self._make_if_cdata(data) self.uve.send_ifstatus_update(self._if_data) sret = 'chngd: %d' % len(self._if_data) self._logger.debug('@do_work(analyze):State %s(%d)->%s!' % ( self._state, len(data), str(sret))) return ret, time def notify(self, svc, msg='', up=True, servers=''): self.uve.conn_state_notify(svc, msg, up, servers) def stop(self): self._keep_running = False def sleep_time(self, newtime=None): if newtime: self._sleep_time = newtime else: self._sleep_time = self._config.frequency() self._fast_scan_freq = self._config.fast_scan_freq() if self._fast_scan_freq > self._sleep_time / 2: self._fast_scan_freq = self._sleep_time / 2 return self._sleep_time def _setup_io(self): cdir = mkdtemp() input_file = os.path.join(cdir, 'in.data') output_file = os.path.join(cdir, 'out.data') return cdir, input_file, output_file def _create_input(self, input_file, output_file, devices, i, restrict=None): with open(input_file, 'wb') as f: data = dict(out=output_file, netdev=devices, instance=i) if restrict: data['restrict'] = restrict pickle.dump(data, f) f.flush() def _run_scanner(self, input_file, output_file, i): proc = subprocess.Popen('contrail-snmp-scanner --input %s' % ( input_file), shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE, close_fds=True) self._cleanup = (proc, output_file) o,e = proc.communicate() self._cleanup = None self._logger.debug('@run_scanner(%d): scan done with %d\nstdout:' \ '\n%s\nstderr:\n%s\n' % (i, proc.returncode, o, e)) with open(output_file, 'rb') as f: d = pickle.load(f) self._logger.debug('@run_scanner(%d): loaded %s' % (i, output_file)) return d def _cleanup_io(self, cdir, input_file, output_file): os.unlink(input_file) os.unlink(output_file) os.rmdir(cdir) def _send_uve(self, d): for dev, data in d.items(): if dev: self.uve.send(data['snmp']) self.uve.send_flow_uve({'name': dev, 'flow_export_source_ip': data['flow_export_source_ip']}) self.find_fix_name(data['name'], dev) self._logger.debug('@send_uve:Processed %d!' % (len(d))) def _del_uves(self, l): with self._sem: for dev in l: self.uve.delete(dev) def do_work(self, i, devices): self._logger.debug('@do_work(%d):started (%d)...' % (i, len(devices))) sleep_time = self._fast_scan_freq if devices: with self._sem: self._work_set = devices cdir, input_file, output_file = self._setup_io() self._create_input(input_file, output_file, devices, i, **self._extra_call_params()) data = self._run_scanner(input_file, output_file, i) self._cleanup_io(cdir, input_file, output_file) do_send, sleep_time = self._analyze(data) if do_send: self._send_uve(data) gevent.sleep(0) del self._work_set self._logger.debug('@do_work(%d):Processed %d!' % (i, len(devices))) return sleep_time def find_fix_name(self, cfg_name, snmp_name): if snmp_name != cfg_name: self._logger.debug('@find_fix_name: snmp name %s differs from ' \ 'configured name %s, fixed for this run' % ( snmp_name, cfg_name)) for d in self._work_set: if d.name == cfg_name: d.name = snmp_name return def run(self): i = 0 self._sem = Semaphore() constnt_schdlr = ConsistentScheduler( self._config._name, zookeeper=self._config.zookeeper_server(), delete_hndlr=self._del_uves, logger=self._logger) while self._keep_running: self._logger.debug('@run: ittr(%d)' % i) if constnt_schdlr.schedule(self._config.devices()): sleep_time = self.do_work(i, constnt_schdlr.work_items()) self._logger.debug('done work %s' % str( map(lambda x: x.name, constnt_schdlr.work_items()))) i += 1 gevent.sleep(sleep_time) else: gevent.sleep(1) constnt_schdlr.finish()
class Controller(object): def __init__(self, config): self._config = config self._me = socket.gethostname() + ':' + str(os.getpid()) self.uve = SnmpUve(self._config) self._logger = self.uve.logger() self.sleep_time() self._keep_running = True self.last = set() self.bucketsize = 47 self._config.set_cb(self.notify) def notify(self, svc, msg='', up=True, servers=''): self.uve.conn_state_notify(svc, msg, up, servers) def stop(self): self._keep_running = False def sleep_time(self, newtime=None): if newtime: self._sleep_time = newtime else: self._sleep_time = self._config.frequency() return self._sleep_time def device2partition(self, key): return struct.unpack( 'H', hashlib.md5(key).digest()[-2:])[0] % self.bucketsize def get_devices_inuse(self): d = [] for i in self._dscvrd_workers: if i['name'] != self._me: d += i['devices'].split(', ') return d def get_net_devices(self, pc): devs = set() inuse = self.get_devices_inuse() for dev in self._config.devices(): if dev.name not in inuse: part = self.device2partition(dev.name) if pc.own_partition(part): devs.add(dev) return list(devs) def do_work(self, i, devices): self._logger.debug('@do_work(%d):started...' % i) if devices: cdir = mkdtemp() input_file = os.path.join(cdir, 'in.data') output_file = os.path.join(cdir, 'out.data') with open(input_file, 'wb') as f: pickle.dump(dict(out=output_file, netdev=devices, instance=i), f) f.flush() proc = subprocess.Popen('contrail-snmp-scanner --input %s' % (input_file), shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) o, e = proc.communicate() self._logger.debug('@do_work(%d): scan done with %d\nstdout:' \ '\n%s\nstderr:\n%s\n' % (i, proc.returncode, o, e)) with open(output_file, 'rb') as f: d = pickle.load(f) os.unlink(input_file) os.unlink(output_file) os.rmdir(cdir) for dev, data in d.items(): self.uve.send(data['snmp']) self.uve.send_flow_uve({ 'name': dev, 'flow_export_source_ip': data['flow_export_source_ip'] }) current = set(d.keys()) for dev in (self.last - current): self.uve.delete(dev) self.last = current self._logger.debug('@do_work(%d):Done!' % i) def workers(self): try: a = self._config._disc.subscribe(self._config._name, 0) gevent.sleep(0) self._dscvrd_workers = a.read() self._logger.debug( '@workers(discovery):%s' % (' '.join(map(lambda x: x['name'], self._dscvrd_workers)))) return set(map(lambda x: x['name'], self._dscvrd_workers)) except Exception as e: import traceback traceback.print_exc() self._logger.exception('@workers(discovery):%s\n%s' % (str(e), str(dir(e)))) self._dscvrd_workers = [] return set([]) def notify_hndlr(self, p): self._logger.debug('@notify_hndlr: New partition %s' % str(p)) def run(self): i = 0 self._config._disc.publish( self._config._name, dict(name=self._me, devices=', '.join(self.last), time='0')) w = self.workers() gevent.sleep(1) pc = PartitionClient(self._config._name, self._me, list(w), self.bucketsize, self.notify_hndlr, self._config.zookeeper_server()) while self._keep_running: w_ = self.workers() if w == w_: i += 1 self.do_work(i, self.get_net_devices(pc)) t = self._sleep_time else: try: pc.update_cluster_list(list(w_)) self._logger.debug('@run(libpartition): updated %s' % (str(w_))) w = w_ t = 1 except Exception as e: import traceback traceback.print_exc() self._logger.exception('@run(libpartition):%s' % str(e)) self._config._disc.publish( self._config._name, dict(name=self._me, devices=', '.join(self.last), time='0')) gevent.sleep(t)
class Controller(object): def __init__(self, config): self._config = config self._me = socket.gethostname() + ':' + str(os.getpid()) self.uve = SnmpUve(self._config) self._logger = self.uve.logger() self.sleep_time() self._keep_running = True self.last = set() self._sem = None self._config.set_cb(self.notify) def notify(self, svc, msg='', up=True, servers=''): self.uve.conn_state_notify(svc, msg, up, servers) def stop(self): self._keep_running = False def sleep_time(self, newtime=None): if newtime: self._sleep_time = newtime else: self._sleep_time = self._config.frequency() return self._sleep_time def _setup_io(self): cdir = mkdtemp() input_file = os.path.join(cdir, 'in.data') output_file = os.path.join(cdir, 'out.data') return cdir, input_file, output_file def _create_input(self, input_file, output_file, devices, i): with open(input_file, 'wb') as f: pickle.dump(dict(out=output_file, netdev=devices, instance=i), f) f.flush() def _run_scanner(self, input_file, output_file, i): proc = subprocess.Popen('contrail-snmp-scanner --input %s' % ( input_file), shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) o,e = proc.communicate() self._logger.debug('@run_scanner(%d): scan done with %d\nstdout:' \ '\n%s\nstderr:\n%s\n' % (i, proc.returncode, o, e)) with open(output_file, 'rb') as f: d = pickle.load(f) self._logger.debug('@run_scanner(%d): loaded %s' % (i, output_file)) return d def _cleanup_io(self, cdir, input_file, output_file): os.unlink(input_file) os.unlink(output_file) os.rmdir(cdir) def _send_uve(self, d): for dev, data in d.items(): self.uve.send(data['snmp']) self.uve.send_flow_uve({'name': dev, 'flow_export_source_ip': data['flow_export_source_ip']}) self.find_fix_name(data['name'], dev) self._logger.debug('@send_uve:Processed %d!' % (len(d))) def _del_uves(self, l): with self._sem: for dev in l: self.uve.delete(dev) def do_work(self, i, devices): self._logger.debug('@do_work(%d):started (%d)...' % (i, len(devices))) if devices: with self._sem: self._work_set = devices cdir, input_file, output_file = self._setup_io() self._create_input(input_file, output_file, devices, i) self._send_uve(self._run_scanner(input_file, output_file, i)) gevent.sleep(0) self._cleanup_io(cdir, input_file, output_file) del self._work_set self._logger.debug('@do_work(%d):Processed %d!' % (i, len(devices))) def find_fix_name(self, cfg_name, snmp_name): if snmp_name != cfg_name: self._logger.debug('@do_work: snmp name %s differs from ' \ 'configured name %s, fixed for this run' % ( snmp_name, cfg_name)) for d in self._work_set: if d.name == cfg_name: d.name = snmp_name return def run(self): i = 0 self._sem = Semaphore() constnt_schdlr = ConsistentScheduler( self._config._name, zookeeper=self._config.zookeeper_server(), delete_hndlr=self._del_uves, logger=self._logger) while self._keep_running: self._logger.debug('@run: ittr(%d)' % i) if constnt_schdlr.schedule(self._config.devices()): self.do_work(i, constnt_schdlr.work_items()) self._logger.debug('done work %s' % str( map(lambda x: x.name, constnt_schdlr.work_items()))) i += 1 gevent.sleep(self._sleep_time) else: gevent.sleep(1) constnt_schdlr.finish()
class Controller(object): def __init__(self, config): self._config = config self.uve = SnmpUve(self._config) self._logger = self.uve.logger() self.sleep_time() self._keep_running = True def stop(self): self._keep_running = False def sleep_time(self, newtime=None): if newtime: self._sleep_time = newtime else: self._sleep_time = self._config.frequency() return self._sleep_time def get_net_devices(self): return list(self._config.devices()) def do_work(self, i): self._logger.debug('@do_work(%d):started...' % i) cdir = mkdtemp() input_file = os.path.join(cdir, 'in.data') output_file = os.path.join(cdir, 'out.data') with open(input_file, 'wb') as f: pickle.dump( dict(out=output_file, netdev=self.get_net_devices(), instance=i), f) f.flush() proc = subprocess.Popen('contrail-snmp-scanner --input %s' % (input_file), shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) o, e = proc.communicate() self._logger.debug('@do_work(%d): scan done with %d\nstdout:' \ '\n%s\nstderr:\n%s\n' % (i, proc.returncode, o, e)) with open(output_file, 'rb') as f: d = pickle.load(f) os.unlink(input_file) os.unlink(output_file) os.rmdir(cdir) for dev, data in d.items(): self.uve.send(data['snmp']) self.uve.send_flow_uve({ 'name': dev, 'flow_export_source_ip': data['flow_export_source_ip'] }) self._logger.debug('@do_work(%d):Done!' % i) def run(self): i = 0 while self._keep_running: self.do_work(i) gevent.sleep(self._sleep_time) i += 1
class Controller(object): def __init__(self, config): self._config = config self._me = socket.gethostname() + ':' + str(os.getpid()) self.uve = SnmpUve(self._config) self._logger = self.uve.logger() self.sleep_time() self._keep_running = True self.last = set() self._sem = None self._config.set_cb(self.notify) def notify(self, svc, msg='', up=True, servers=''): self.uve.conn_state_notify(svc, msg, up, servers) def stop(self): self._keep_running = False def sleep_time(self, newtime=None): if newtime: self._sleep_time = newtime else: self._sleep_time = self._config.frequency() return self._sleep_time def _setup_io(self): cdir = mkdtemp() input_file = os.path.join(cdir, 'in.data') output_file = os.path.join(cdir, 'out.data') return cdir, input_file, output_file def _create_input(self, input_file, output_file, devices, i): with open(input_file, 'wb') as f: pickle.dump(dict(out=output_file, netdev=devices, instance=i), f) f.flush() def _run_scanner(self, input_file, output_file, i): proc = subprocess.Popen('contrail-snmp-scanner --input %s' % (input_file), shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) o, e = proc.communicate() self._logger.debug('@run_scanner(%d): scan done with %d\nstdout:' \ '\n%s\nstderr:\n%s\n' % (i, proc.returncode, o, e)) with open(output_file, 'rb') as f: d = pickle.load(f) self._logger.debug('@run_scanner(%d): loaded %s' % (i, output_file)) return d def _cleanup_io(self, cdir, input_file, output_file): os.unlink(input_file) os.unlink(output_file) os.rmdir(cdir) def _send_uve(self, d): for dev, data in d.items(): self.uve.send(data['snmp']) self.uve.send_flow_uve({ 'name': dev, 'flow_export_source_ip': data['flow_export_source_ip'] }) self.find_fix_name(data['name'], dev) self._logger.debug('@send_uve:Processed %d!' % (len(d))) def _del_uves(self, l): with self._sem: for dev in l: self.uve.delete(dev) def do_work(self, i, devices): self._logger.debug('@do_work(%d):started (%d)...' % (i, len(devices))) if devices: with self._sem: self._work_set = devices cdir, input_file, output_file = self._setup_io() self._create_input(input_file, output_file, devices, i) self._send_uve(self._run_scanner(input_file, output_file, i)) gevent.sleep(0) self._cleanup_io(cdir, input_file, output_file) del self._work_set self._logger.debug('@do_work(%d):Processed %d!' % (i, len(devices))) def find_fix_name(self, cfg_name, snmp_name): if snmp_name != cfg_name: self._logger.debug('@do_work: snmp name %s differs from ' \ 'configured name %s, fixed for this run' % ( snmp_name, cfg_name)) for d in self._work_set: if d.name == cfg_name: d.name = snmp_name return def run(self): i = 0 self._sem = Semaphore() constnt_schdlr = ConsistentScheduler( self._config._name, zookeeper=self._config.zookeeper_server(), delete_hndlr=self._del_uves, logger=self._logger) while self._keep_running: self._logger.debug('@run: ittr(%d)' % i) if constnt_schdlr.schedule(self._config.devices()): self.do_work(i, constnt_schdlr.work_items()) self._logger.debug( 'done work %s' % str(map(lambda x: x.name, constnt_schdlr.work_items()))) i += 1 gevent.sleep(self._sleep_time) else: gevent.sleep(1) constnt_schdlr.finish()