class FbfProductController(object): """ Wrapper class for an FBFUSE product. """ STATES = ["idle", "preparing", "ready", "starting", "capturing", "stopping", "error"] IDLE, PREPARING, READY, STARTING, CAPTURING, STOPPING, ERROR = STATES def __init__(self, parent, product_id, katpoint_antennas, n_channels, feng_streams, proxy_name, feng_config): """ @brief Construct new instance @param parent The parent FbfMasterController instance @param product_id The name of the product @param katpoint_antennas A list of katpoint.Antenna objects @param n_channels The integer number of frequency channels provided by the CBF. @param feng_streams A string describing the multicast groups containing F-enging data (in the form: spead://239.11.1.150+15:7147) @param proxy_name The name of the proxy associated with this subarray (used as a sensor prefix) #NEED FENG CONFIG @param servers A list of FbfWorkerServer instances allocated to this product controller """ self.log = logging.getLogger("mpikat.fbfuse_product_controller.{}".format(product_id)) self.log.debug("Creating new FbfProductController with args: {}".format( ", ".join([str(i) for i in (parent, product_id, katpoint_antennas, n_channels, feng_streams, proxy_name, feng_config)]))) self._parent = parent self._product_id = product_id self._antennas = ",".join([a.name for a in katpoint_antennas]) self._katpoint_antennas = katpoint_antennas self._antenna_map = {a.name: a for a in self._katpoint_antennas} self._n_channels = n_channels self._streams = ip_range_from_stream(feng_streams) self._proxy_name = proxy_name self._feng_config = feng_config self._servers = [] self._beam_manager = None self._delay_config_server = None self._ca_client = None self._previous_sb_config = None self._managed_sensors = [] self._ibc_mcast_group = None self._cbc_mcast_groups = None self._default_sb_config = { u'coherent-beams-nbeams':400, u'coherent-beams-tscrunch':16, u'coherent-beams-fscrunch':1, u'coherent-beams-antennas':self._antennas, u'coherent-beams-granularity':6, u'incoherent-beam-tscrunch':16, u'incoherent-beam-fscrunch':1, u'incoherent-beam-antennas':self._antennas, u'bandwidth':self._feng_config['bandwidth'], u'centre-frequency':self._feng_config['centre-frequency']} self.setup_sensors() def __del__(self): self.teardown_sensors() def info(self): """ @brief Return a metadata dictionary describing this product controller """ out = { "antennas":self._antennas, "nservers":len(self.servers), "capturing":self.capturing, "streams":self._streams, "nchannels":self._n_channels, "proxy_name":self._proxy_name } return out def add_sensor(self, sensor): """ @brief Add a sensor to the parent object @note This method is used to wrap calls to the add_sensor method on the parent FbfMasterController instance. In order to disambiguate between sensors from describing different products the associated proxy name is used as sensor prefix. For example the "servers" sensor will be seen by clients connected to the FbfMasterController server as "<proxy_name>-servers" (e.g. "FBFUSE_1-servers"). """ prefix = "{}.".format(self._product_id) if sensor.name.startswith(prefix): self._parent.add_sensor(sensor) else: sensor.name = "{}{}".format(prefix,sensor.name) self._parent.add_sensor(sensor) self._managed_sensors.append(sensor) def setup_sensors(self): """ @brief Setup the default KATCP sensors. @note As this call is made only upon an FBFUSE configure call a mass inform is required to let connected clients know that the proxy interface has changed. """ self._state_sensor = LoggingSensor.discrete( "state", description = "Denotes the state of this FBF instance", params = self.STATES, default = self.IDLE, initial_status = Sensor.NOMINAL) self._state_sensor.set_logger(self.log) self.add_sensor(self._state_sensor) self._ca_address_sensor = Sensor.string( "configuration-authority", description = "The address of the server that will be deferred to for configurations", default = "", initial_status = Sensor.UNKNOWN) self.add_sensor(self._ca_address_sensor) self._available_antennas_sensor = Sensor.string( "available-antennas", description = "The antennas that are currently available for beamforming", default = json.dumps({antenna.name:antenna.format_katcp() for antenna in self._katpoint_antennas}), initial_status = Sensor.NOMINAL) self.add_sensor(self._available_antennas_sensor) self._phase_reference_sensor = Sensor.string( "phase-reference", description="A KATPOINT target string denoting the F-engine phasing centre", default="unset,radec,0,0", initial_status=Sensor.UNKNOWN) self.add_sensor(self._phase_reference_sensor) reference_antenna = Antenna("reference,{ref.lat},{ref.lon},{ref.elev}".format( ref=self._katpoint_antennas[0].ref_observer)) self._reference_antenna_sensor = Sensor.string( "reference-antenna", description="A KATPOINT antenna string denoting the reference antenna", default=reference_antenna.format_katcp(), initial_status=Sensor.NOMINAL) self.add_sensor(self._reference_antenna_sensor) self._bandwidth_sensor = Sensor.float( "bandwidth", description = "The bandwidth this product is configured to process", default = self._default_sb_config['bandwidth'], initial_status = Sensor.UNKNOWN) self.add_sensor(self._bandwidth_sensor) self._nchans_sensor = Sensor.integer( "nchannels", description = "The number of channels to be processesed", default = self._n_channels, initial_status = Sensor.UNKNOWN) self.add_sensor(self._nchans_sensor) self._cfreq_sensor = Sensor.float( "centre-frequency", description = "The centre frequency of the band this product configured to process", default = self._default_sb_config['centre-frequency'], initial_status = Sensor.UNKNOWN) self.add_sensor(self._cfreq_sensor) self._cbc_nbeams_sensor = Sensor.integer( "coherent-beam-count", description = "The number of coherent beams that this FBF instance can currently produce", default = self._default_sb_config['coherent-beams-nbeams'], initial_status = Sensor.UNKNOWN) self.add_sensor(self._cbc_nbeams_sensor) self._cbc_nbeams_per_group = Sensor.integer( "coherent-beam-count-per-group", description = "The number of coherent beams packed into a multicast group", default = 1, initial_status = Sensor.UNKNOWN) self.add_sensor(self._cbc_nbeams_per_group) self._cbc_ngroups = Sensor.integer( "coherent-beam-ngroups", description = "The number of multicast groups used for coherent beam transmission", default = 1, initial_status = Sensor.UNKNOWN) self.add_sensor(self._cbc_ngroups) self._cbc_nbeams_per_server_set = Sensor.integer( "coherent-beam-nbeams-per-server-set", description = "The number of beams produced by each server set", default = 1, initial_status = Sensor.UNKNOWN) self.add_sensor(self._cbc_nbeams_per_server_set) self._cbc_tscrunch_sensor = Sensor.integer( "coherent-beam-tscrunch", description = "The number time samples that will be integrated when producing coherent beams", default = self._default_sb_config['coherent-beams-tscrunch'], initial_status = Sensor.UNKNOWN) self.add_sensor(self._cbc_tscrunch_sensor) self._cbc_fscrunch_sensor = Sensor.integer( "coherent-beam-fscrunch", description = "The number frequency channels that will be integrated when producing coherent beams", default = self._default_sb_config['coherent-beams-fscrunch'], initial_status = Sensor.UNKNOWN) self.add_sensor(self._cbc_fscrunch_sensor) self._cbc_antennas_sensor = Sensor.string( "coherent-beam-antennas", description = "The antennas that will be used when producing coherent beams", default = self._default_sb_config['coherent-beams-antennas'], initial_status = Sensor.UNKNOWN) self.add_sensor(self._cbc_antennas_sensor) self._cbc_mcast_groups_sensor = Sensor.string( "coherent-beam-multicast-groups", description = "Multicast groups used by this instance for sending coherent beam data", default = "", initial_status = Sensor.UNKNOWN) self.add_sensor(self._cbc_mcast_groups_sensor) self._cbc_mcast_groups_mapping_sensor = Sensor.string( "coherent-beam-multicast-group-mapping", description = "Mapping of mutlicast group address to the coherent beams in that group", default= "", initial_status = Sensor.UNKNOWN) self.add_sensor(self._cbc_mcast_groups_mapping_sensor) self._ibc_nbeams_sensor = Sensor.integer( "incoherent-beam-count", description = "The number of incoherent beams that this FBF instance can currently produce", default = 1, initial_status = Sensor.UNKNOWN) self.add_sensor(self._ibc_nbeams_sensor) self._ibc_tscrunch_sensor = Sensor.integer( "incoherent-beam-tscrunch", description = "The number time samples that will be integrated when producing incoherent beams", default = self._default_sb_config['incoherent-beam-tscrunch'], initial_status = Sensor.UNKNOWN) self.add_sensor(self._ibc_tscrunch_sensor) self._ibc_fscrunch_sensor = Sensor.integer( "incoherent-beam-fscrunch", description = "The number frequency channels that will be integrated when producing incoherent beams", default = self._default_sb_config['incoherent-beam-fscrunch'], initial_status = Sensor.UNKNOWN) self.add_sensor(self._ibc_fscrunch_sensor) self._ibc_antennas_sensor = Sensor.string( "incoherent-beam-antennas", description = "The antennas that will be used when producing incoherent beams", default = self._default_sb_config['incoherent-beam-antennas'], initial_status = Sensor.UNKNOWN) self.add_sensor(self._ibc_antennas_sensor) self._ibc_mcast_group_sensor = Sensor.string( "incoherent-beam-multicast-group", description = "Multicast group used by this instance for sending incoherent beam data", default = "", initial_status = Sensor.UNKNOWN) self.add_sensor(self._ibc_mcast_group_sensor) self._servers_sensor = Sensor.string( "servers", description = "The worker server instances currently allocated to this product", default = ",".join(["{s.hostname}:{s.port}".format(s=server) for server in self._servers]), initial_status = Sensor.UNKNOWN) self.add_sensor(self._servers_sensor) self._nserver_sets_sensor = Sensor.integer( "nserver-sets", description = "The number of server sets (independent subscriptions to the F-engines)", default = 1, initial_status = Sensor.UNKNOWN) self.add_sensor(self._nserver_sets_sensor) self._nservers_per_set_sensor = Sensor.integer( "nservers-per-set", description = "The number of servers per server set", default = 1, initial_status = Sensor.UNKNOWN) self.add_sensor(self._nservers_per_set_sensor) self._delay_config_server_sensor = Sensor.string( "delay-config-server", description = "The address of the delay configuration server for this product", default = "", initial_status = Sensor.UNKNOWN) self.add_sensor(self._delay_config_server_sensor) def teardown_sensors(self): """ @brief Remove all sensors created by this product from the parent server. @note This method is required for cleanup to stop the FBF sensor pool becoming swamped with unused sensors. """ for sensor in self._managed_sensors: self._parent.remove_sensor(sensor) self._managed_sensors = [] self._parent.mass_inform(Message.inform('interface-changed')) @property def servers(self): return self._servers @property def capturing(self): return self.state == self.CAPTURING @property def idle(self): return self.state == self.IDLE @property def starting(self): return self.state == self.STARTING @property def stopping(self): return self.state == self.STOPPING @property def ready(self): return self.state == self.READY @property def preparing(self): return self.state == self.PREPARING @property def error(self): return self.state == self.ERROR @property def state(self): return self._state_sensor.value() def _verify_antennas(self, antennas): """ @brief Verify that a set of antennas is available to this instance. @param antennas A CSV list of antenna names """ self.log.debug("Verifying antenna set: {}".format(antennas)) antennas_set = set([ant.name for ant in self._katpoint_antennas]) requested_antennas = set(antennas) return requested_antennas.issubset(antennas_set) def set_configuration_authority(self, hostname, port): if self._ca_client: self._ca_client.stop() self._ca_client = KATCPClientResource(dict( name = 'configuration-authority-client', address = (hostname, port), controlled = True)) self._ca_client.start() self._ca_address_sensor.set_value("{}:{}".format(hostname, port)) @coroutine def get_ca_sb_configuration(self, sb_id): self.log.debug("Retrieving schedule block configuration from configuration authority") yield self._ca_client.until_synced() try: response = yield self._ca_client.req.get_schedule_block_configuration(self._proxy_name, sb_id) except Exception as error: self.log.error("Request for SB configuration to CA failed with error: {}".format(str(error))) raise error try: config_dict = json.loads(response.reply.arguments[1]) except Exception as error: self.log.error("Could not parse CA SB configuration with error: {}".format(str(error))) raise error self.log.debug("Configuration authority returned: {}".format(config_dict)) raise Return(config_dict) def reset_sb_configuration(self): self.log.debug("Reseting schedule block configuration") try: self.capture_stop() except Exception as error: self.log.warning("Received error while attempting capture stop: {}".format(str(error))) self._parent._server_pool.deallocate(self._servers) if self._ibc_mcast_group: self._parent._ip_pool.free(self._ibc_mcast_group) if self._cbc_mcast_groups: self._parent._ip_pool.free(self._cbc_mcast_groups) self._cbc_mcast_groups = None self._ibc_mcast_group = None self._servers = [] if self._delay_config_server: self._delay_config_server.stop() self._delay_config_server = None self._beam_manager = None def set_error_state(self, message): self.reset_sb_configuration() self._state_sensor.set_value(self.ERROR) def set_sb_configuration(self, config_dict): """ @brief Set the schedule block configuration for this product @param config_dict A dictionary specifying configuation parameters, e.g. @code { u'coherent-beams-nbeams':100, u'coherent-beams-tscrunch':22, u'coherent-beams-fscrunch':2, u'coherent-beams-antennas':'m007', u'coherent-beams-granularity':6, u'incoherent-beam-tscrunch':16, u'incoherent-beam-fscrunch':1, u'incoherent-beam-antennas':'m008' } @endcode @detail Valid parameters for the configuration dictionary are as follows: coherent-beams-nbeams - The desired number of coherent beams to produce coherent-beams-tscrunch - The number of spectra to integrate in the coherent beamformer coherent-beams-tscrunch - The number of spectra to integrate in the coherent beamformer coherent-beams-fscrunch - The number of channels to integrate in the coherent beamformer coherent-beams-antennas - The specific antennas to use for the coherent beamformer coherent-beams-granularity - The number of beams per output mutlicast group (an integer divisor or multiplier of this number will be used) incoherent-beam-tscrunch - The number of spectra to integrate in the incoherent beamformer incoherent-beam-fscrunch - The number of channels to integrate in the incoherent beamformer incoherent-beam-antennas - The specific antennas to use for the incoherent beamformer centre-frequency - The desired centre frequency in Hz bandwidth - The desired bandwidth in Hz @note FBFUSE reasonably assumes that the user does not know the possible configurations at any given time. As such it tries to satisfy the users request but will not throw an error if the requested configuration is not acheivable, instead opting to provide a reduced configuration. For example the user may request 1000 beams and 6 beams per multicast group but FBFUSE may configure to produce 860 beams and 24 beams per multicast group. If the user can only use 6 beams per multcast group, then in the 24-beam case they must subscribe to the same multicast group 4 times on different nodes. """ if self._previous_sb_config == config_dict: self.log.info("Configuration is unchanged, proceeding with existing configuration") return else: self._previous_sb_config = config_dict self.reset_sb_configuration() self.log.info("Setting schedule block configuration") config = deepcopy(self._default_sb_config) config.update(config_dict) self.log.info("Configuring using: {}".format(config)) requested_cbc_antenna = parse_csv_antennas(config['coherent-beams-antennas']) if not self._verify_antennas(requested_cbc_antenna): raise Exception("Requested coherent beam antennas are not a subset of the available antennas") requested_ibc_antenna = parse_csv_antennas(config['incoherent-beam-antennas']) if not self._verify_antennas(requested_ibc_antenna): raise Exception("Requested incoherent beam antennas are not a subset of the available antennas") # first we need to get one ip address for the incoherent beam self._ibc_mcast_group = self._parent._ip_pool.allocate(1) self._ibc_mcast_group_sensor.set_value(self._ibc_mcast_group.format_katcp()) largest_ip_range = self._parent._ip_pool.largest_free_range() nworkers_available = self._parent._server_pool.navailable() cm = FbfConfigurationManager(len(self._katpoint_antennas), self._feng_config['bandwidth'], self._n_channels, nworkers_available, largest_ip_range) requested_nantennas = len(parse_csv_antennas(config['coherent-beams-antennas'])) mcast_config = cm.get_configuration( config['coherent-beams-tscrunch'], config['coherent-beams-fscrunch'], config['coherent-beams-nbeams'], requested_nantennas, config['bandwidth'], config['coherent-beams-granularity']) self._bandwidth_sensor.set_value(config['bandwidth']) self._cfreq_sensor.set_value(config['centre-frequency']) self._nchans_sensor.set_value(mcast_config['num_chans']) self._cbc_nbeams_sensor.set_value(mcast_config['num_beams']) self._cbc_nbeams_per_group.set_value(mcast_config['num_beams_per_mcast_group']) self._cbc_ngroups.set_value(mcast_config['num_mcast_groups']) self._cbc_nbeams_per_server_set.set_value(mcast_config['num_beams_per_worker_set']) self._cbc_tscrunch_sensor.set_value(config['coherent-beams-tscrunch']) self._cbc_fscrunch_sensor.set_value(config['coherent-beams-fscrunch']) self._cbc_antennas_sensor.set_value(config['coherent-beams-antennas']) self._ibc_tscrunch_sensor.set_value(config['incoherent-beam-tscrunch']) self._ibc_fscrunch_sensor.set_value(config['incoherent-beam-fscrunch']) self._ibc_antennas_sensor.set_value(config['incoherent-beam-antennas']) self._servers = self._parent._server_pool.allocate(mcast_config['num_workers_total']) server_str = ",".join(["{s.hostname}:{s.port}".format(s=server) for server in self._servers]) self._servers_sensor.set_value(server_str) self._nserver_sets_sensor.set_value(mcast_config['num_worker_sets']) self._nservers_per_set_sensor.set_value(mcast_config['num_workers_per_set']) self._cbc_mcast_groups = self._parent._ip_pool.allocate(mcast_config['num_mcast_groups']) self._cbc_mcast_groups_sensor.set_value(self._cbc_mcast_groups.format_katcp()) return cm @coroutine def get_ca_target_configuration(self, target): def ca_target_update_callback(received_timestamp, timestamp, status, value): # TODO, should we really reset all the beams or should we have # a mechanism to only update changed beams config_dict = json.loads(value) self.reset_beams() for target_string in config_dict.get('beams',[]): target = Target(target_string) self.add_beam(target) for tiling in config_dict.get('tilings',[]): target = Target(tiling['target']) #required freq = float(tiling.get('reference_frequency', self._cfreq_sensor.value())) nbeams = int(tiling['nbeams']) overlap = float(tiling.get('overlap', 0.5)) epoch = float(tiling.get('epoch', time.time())) self.add_tiling(target, nbeams, freq, overlap, epoch) yield self._ca_client.until_synced() try: response = yield self._ca_client.req.target_configuration_start(self._proxy_name, target.format_katcp()) except Exception as error: self.log.error("Request for target configuration to CA failed with error: {}".format(str(error))) raise error if not response.reply.reply_ok(): error = Exception(response.reply.arguments[1]) self.log.error("Request for target configuration to CA failed with error: {}".format(str(error))) raise error yield self._ca_client.until_synced() sensor = self._ca_client.sensor["{}_beam_position_configuration".format(self._proxy_name)] sensor.register_listener(ca_target_update_callback) self._ca_client.set_sampling_strategy(sensor.name, "event") def _beam_to_sensor_string(self, beam): return beam.target.format_katcp() @coroutine def target_start(self, target): self._phase_reference_sensor.set_value(target) if self._ca_client: yield self.get_ca_target_configuration(target) else: self.log.warning("No configuration authority is set, using default beam configuration") @coroutine def target_stop(self): if self._ca_client: sensor_name = "{}_beam_position_configuration".format(self._proxy_name) self._ca_client.set_sampling_strategy(sensor_name, "none") @coroutine def prepare(self, sb_id): """ @brief Prepare the beamformer for streaming @detail This method evaluates the current configuration creates a new DelayEngine and passes a prepare call to all allocated servers. """ if not self.idle: raise FbfProductStateError([self.IDLE], self.state) self.log.info("Preparing FBFUSE product") self._state_sensor.set_value(self.PREPARING) self.log.debug("Product moved to 'preparing' state") # Here we need to parse the streams and assign beams to streams: #mcast_addrs, mcast_port = parse_stream(self._streams['cbf.antenna_channelised_voltage']['i0.antenna-channelised-voltage']) if not self._ca_client: self.log.warning("No configuration authority found, using default configuration parameters") cm = self.set_sb_configuration(self._default_sb_config) else: #TODO: get the schedule block ID into this call from somewhere (configure?) try: config = yield self.get_ca_sb_configuration(sb_id) cm = self.set_sb_configuration(config) except Exception as error: self.log.error("Configuring from CA failed with error: {}".format(str(error))) self.log.warning("Reverting to default configuration") cm = self.set_sb_configuration(self._default_sb_config) cbc_antennas_names = parse_csv_antennas(self._cbc_antennas_sensor.value()) cbc_antennas = [self._antenna_map[name] for name in cbc_antennas_names] self._beam_manager = BeamManager(self._cbc_nbeams_sensor.value(), cbc_antennas) self._delay_config_server = DelayConfigurationServer("127.0.0.1", 0, self._beam_manager) self._delay_config_server.start() self.log.info("Started delay engine at: {}".format(self._delay_config_server.bind_address)) de_ip, de_port = self._delay_config_server.bind_address self._delay_config_server_sensor.set_value((de_ip, de_port)) # Need to tear down the beam sensors here # Here calculate the beam to multicast map self._beam_sensors = [] mcast_to_beam_map = {} groups = [ip for ip in self._cbc_mcast_groups] idxs = [beam.idx for beam in self._beam_manager.get_beams()] for group in groups: self.log.debug("Allocating beams to {}".format(str(group))) key = str(group) for _ in range(self._cbc_nbeams_per_group.value()): if not key in mcast_to_beam_map: mcast_to_beam_map[str(group)] = [] value = idxs.pop(0) self.log.debug("--> Allocated {} to {}".format(value, str(group))) mcast_to_beam_map[str(group)].append(value) self._cbc_mcast_groups_mapping_sensor.set_value(json.dumps(mcast_to_beam_map)) for beam in self._beam_manager.get_beams(): sensor = Sensor.string( "coherent-beam-{}".format(beam.idx), description="R.A. (deg), declination (deg) and source name for coherent beam with ID {}".format(beam.idx), default=self._beam_to_sensor_string(beam), initial_status=Sensor.UNKNOWN) beam.register_observer(lambda beam, sensor=sensor: sensor.set_value(self._beam_to_sensor_string(beam))) self._beam_sensors.append(sensor) self.add_sensor(sensor) self._parent.mass_inform(Message.inform('interface-changed')) #Here we actually start to prepare the remote workers ip_splits = self._streams.split(N_FENG_STREAMS_PER_WORKER) # This is assuming lower sideband and bandwidth is always +ve fbottom = self._feng_config['centre-frequency'] - self._feng_config['bandwidth']/2. coherent_beam_config = { 'tscrunch':self._cbc_tscrunch_sensor.value(), 'fscrunch':self._cbc_fscrunch_sensor.value(), 'antennas':self._cbc_antennas_sensor.value() } incoherent_beam_config = { 'tscrunch':self._ibc_tscrunch_sensor.value(), 'fscrunch':self._ibc_fscrunch_sensor.value(), 'antennas':self._ibc_antennas_sensor.value() } prepare_futures = [] for ii, (server, ip_range) in enumerate(zip(self._servers, ip_splits)): chan0_idx = cm.nchans_per_worker * ii chan0_freq = fbottom + chan0_idx * cm.channel_bandwidth future = server.prepare(ip_range.format_katcp(), cm.nchans_per_group, chan0_idx, chan0_freq, cm.channel_bandwidth, mcast_to_beam_map, self._feng_config['feng-antenna-map'], coherent_beam_config, incoherent_beam_config, de_ip, de_port) prepare_futures.append(future) failure_count = 0 for future in prepare_futures: try: yield future except Exception as error: log.error("Failed to configure server with error: {}".format(str(error))) failure_count += 1 if failure_count > 0: self._state_sensor.set_value(self.ERROR) self.log.info("Failed to prepare FBFUSE product") else: self._state_sensor.set_value(self.READY) self.log.info("Successfully prepared FBFUSE product") def deconfigure(self): """ @brief Deconfigure the product. To be called on a subarray deconfigure. @detail This is the final cleanup operation for the product, it should delete all sensors and ensure the release of all resource allocations. """ self.reset_sb_configuration() self.teardown_sensors() def capture_start(self): if not self.ready: raise FbfProductStateError([self.READY], self.state) self._state_sensor.set_value(self.STARTING) self.log.debug("Product moved to 'starting' state") """ futures = [] for server in self._servers: futures.append(server.req.start_capture()) for future in futures: try: response = yield future except: pass """ self._state_sensor.set_value(self.CAPTURING) self.log.debug("Product moved to 'capturing' state") def capture_stop(self): """ @brief Stops the beamformer servers streaming. @detail This should only be called on a schedule block reconfiguration if the same configuration persists between schedule blocks then it is preferable to continue streaming rather than stopping and starting again. """ if not self.capturing and not self.error: return self._state_sensor.set_value(self.STOPPING) self.target_stop() for server in self._servers: #yield server.req.deconfigure() pass self._state_sensor.set_value(self.IDLE) def add_beam(self, target): """ @brief Specify the parameters of one managed beam @param target A KATPOINT target object @return Returns the allocated Beam object """ valid_states = [self.READY, self.CAPTURING, self.STARTING] if not self.state in valid_states: raise FbfProductStateError(valid_states, self.state) return self._beam_manager.add_beam(target) def add_tiling(self, target, number_of_beams, reference_frequency, overlap, epoch): """ @brief Add a tiling to be managed @param target A KATPOINT target object @param reference_frequency The reference frequency at which to calculate the synthesised beam shape, and thus the tiling pattern. Typically this would be chosen to be the centre frequency of the current observation. @param overlap The desired overlap point between beams in the pattern. The overlap defines at what power point neighbouring beams in the tiling pattern will meet. For example an overlap point of 0.1 corresponds to beams overlapping only at their 10%-power points. Similarly a overlap of 0.5 corresponds to beams overlapping at their half-power points. [Note: This is currently a tricky parameter to use when values are close to zero. In future this may be define in sigma units or in multiples of the FWHM of the beam.] @returns The created Tiling object """ valid_states = [self.READY, self.CAPTURING, self.STARTING] if not self.state in valid_states: raise FbfProductStateError(valid_states, self.state) tiling = self._beam_manager.add_tiling(target, number_of_beams, reference_frequency, overlap) try: tiling.generate(self._katpoint_antennas, epoch) except Exception as error: self.log.error("Failed to generate tiling pattern with error: {}".format(str(error))) return tiling def reset_beams(self): """ @brief reset and deallocate all beams and tilings managed by this instance @note All tiling will be lost on this call and must be remade for subsequent observations """ valid_states = [self.READY, self.CAPTURING, self.STARTING] if not self.state in valid_states: raise FbfProductStateError(valid_states, self.state) self._beam_manager.reset()
class KatcpSidecar(object): def __init__(self, host, port): """ Constructs a new instance. :param host: The address of the server to sidecar :param port: The server port """ log.debug("Constructing sidecar for {}:{}".format(host, port)) self.rc = KATCPClientResource( dict(name="sidecar-client", address=(host, port), controlled=True)) self._update_callbacks = set() self._previous_sensors = set() @coroutine def start(self): """ @brief Start the sidecar """ @coroutine def _start(): log.debug("Waiting on synchronisation with server") yield self.rc.until_synced() log.debug("Client synced") log.debug("Requesting version info") response = yield self.rc.req.version_list() log.info("response: {}".format(response)) self.ioloop.add_callback(self.on_interface_changed) self.rc.start() self.ic = self.rc._inspecting_client self.ioloop = self.rc.ioloop self.ic.katcp_client.hook_inform( "interface-changed", lambda message: self.ioloop.add_callback( self.on_interface_changed)) self.ioloop.add_callback(_start) def stop(self): """ @brief Stop the sidecar """ self.rc.stop() @coroutine def on_interface_changed(self): """ @brief Synchronise with the sidecar'd servers new sensors """ log.debug("Waiting on synchronisation with server") yield self.rc.until_synced() log.debug("Client synced") current_sensors = set(self.rc.sensor.keys()) log.debug("Current sensor set: {}".format(current_sensors)) removed = self._previous_sensors.difference(current_sensors) log.debug("Sensors removed since last update: {}".format(removed)) added = current_sensors.difference(self._previous_sensors) log.debug("Sensors added since last update: {}".format(added)) for name in list(added): log.debug("Setting sampling strategy and callbacks on sensor '{}'". format(name)) self.rc.set_sampling_strategy(name, "auto") self.rc.set_sensor_listener(name, self.on_sensor_update) self._previous_sensors = current_sensors @coroutine def on_sensor_update(self, sensor, reading): """ @brief Callback to be executed on a sensor being updated @param sensor A KATCP Sensor Object @param reading The sensor reading """ log.debug("Recieved sensor update for sensor '{}': {}".format( sensor.name, repr(reading))) for callback in list(self._update_callbacks): try: callback(sensor, reading) except Exception as error: log.exception( "Failed to call update callback {} with error: {}".format( callback, str(error))) def add_sensor_update_callback(self, callback): """ @brief Add a sensor update callback. @param callback: The callback @note The callback must have a call signature of func(sensor, reading) """ self._update_callbacks.add(callback) def remove_sensor_update_callback(self, callback): """ @brief Remove a sensor update callback. @param callback: The callback """ self._update_callbacks.remove(callback)
class KATCPToIGUIConverter(object): def __init__(self, host, port, igui_host, igui_user, igui_pass, igui_device_id): """ @brief Class for katcp to igui converter. @param host KATCP host address @param port KATCP port number @param igui_host iGUI server hostname @param igui_user iGUI username @param igui_pass iGUI password @param igui_device_id iGUI device ID """ self.rc = KATCPClientResource( dict(name="test-client", address=(host, port), controlled=True)) self.host = host self.port = port self.igui_host = igui_host self.igui_user = igui_user self.igui_pass = igui_pass self.igui_group_id = None self.igui_device_id = igui_device_id self.igui_connection = IGUIConnection(self.igui_host, self.igui_user, self.igui_pass) self.igui_task_id = None self.igui_rxmap = None self.ioloop = None self.ic = None self.api_version = None self.implementation_version = None self.previous_sensors = set() def start(self): """ @brief Start the instance running @detail This call will trigger connection of the KATCPResource client and will login to the iGUI server. Once both connections are established the instance will retrieve a mapping of the iGUI receivers, devices and tasks and will try to identify the parent of the device_id provided in the constructor. @param self The object @return { description_of_the_return_value } """ @tornado.gen.coroutine def _start(): log.debug("Waiting on synchronisation with server") yield self.rc.until_synced() log.debug("Client synced") log.debug("Requesting version info") # This information can be used to get an iGUI device ID response = yield self.rc.req.version_list() log.info("response {}".format(response)) # for internal device KATCP server, response.informs[2].arguments return index out of range #_, api, implementation = response.informs[2].arguments #self.api_version = api #self.implementation_version = implementation #log.info("katcp-device API: {}".format(self.api_version)) #log.info("katcp-device implementation: {}".format(self.implementation_version)) self.ioloop.add_callback(self.update) log.debug("Starting {} instance".format(self.__class__.__name__)) # self.igui_connection.login() #self.igui_connection.login(self.igui_user, self.igui_pass) self.igui_rxmap = self.igui_connection.build_igui_representation() #log.debug(self.igui_rxmap) # Here we do a look up to find the parent of this device for rx in self.igui_rxmap: log.debug(rx.id) if self.igui_device_id in rx.devices._by_id.keys(): log.debug(self.igui_device_id) log.debug(rx.id) self.igui_rx_id = rx.id log.debug("Found Rx parent: {}".format(self.igui_rx_id)) break else: log.debug("Device '{}' is not a child of any receiver".format( self.igui_device_id)) raise IGUIMappingException( "Device '{}' is not a child of any receiver".format( self.igui_device_id)) #log.debug("iGUI representation:\n{}".format(self.igui_rxmap)) self.rc.start() self.ic = self.rc._inspecting_client self.ioloop = self.rc.ioloop self.ic.katcp_client.hook_inform( "interface-changed", lambda message: self.ioloop.add_callback(self.update)) self.ioloop.add_callback(_start) @tornado.gen.coroutine def update(self): """ @brief Synchronise with the KATCP servers sensors and register new listners """ log.debug("Waiting on synchronisation with server") yield self.rc.until_synced() log.debug("Client synced") current_sensors = set(self.rc.sensor.keys()) log.debug("Current sensor set: {}".format(current_sensors)) removed = self.previous_sensors.difference(current_sensors) log.debug("Sensors removed since last update: {}".format(removed)) added = current_sensors.difference(self.previous_sensors) log.debug("Sensors added since last update: {}".format(added)) for name in list(added): log.debug("Setting sampling strategy and callbacks on sensor '{}'". format(name)) # strat3 = ('event-rate', 2.0, 3.0) #event-rate doesn't work # self.rc.set_sampling_strategy(name, strat3) #KATCPSensorError: # Error setting strategy # not sure that auto means here self.rc.set_sampling_strategy(name, "auto") #self.rc.set_sampling_strategy(name, ["period", (10)]) #self.rc.set_sampling_strategy(name, "event") self.rc.set_sensor_listener(name, self._sensor_updated) self.previous_sensors = current_sensors def _sensor_updated(self, sensor, reading): """ @brief Callback to be executed on a sensor being updated @param sensor The sensor @param reading The sensor reading """ log.debug("Recieved sensor update for sensor '{}': {}".format( sensor.name, repr(reading))) try: rx = self.igui_rxmap.by_id(self.igui_rx_id) except KeyError: raise Exception("No iGUI receiver with ID {}".format( self.igui_rx_id)) try: device = rx.devices.by_id(self.igui_device_id) except KeyError: raise Exception("No iGUI device with ID {}".format( self.igui_device_id)) try: #self.igui_rxmap = self.igui_connection.build_igui_representation() #device = self.igui_rxmap.by_id(self.igui_rx_id).devices.by_id(self.igui_device_id) task = device.tasks.by_name(sensor.name) except KeyError: if (sensor.name[-3:] == 'PNG'): task = json.loads( self.igui_connection.create_task( device, (sensor.name, "NONE", "", "IMAGE", "GET_SET", "0", "0", "0", "-10000000000000000", "10000000000000000", "300"))) else: task = json.loads( self.igui_connection.create_task( device, (sensor.name, "NONE", "", "GETSET", "GET", "0", "0", "0", "-10000000000000000", "10000000000000000", "300"))) self.igui_task_id = str(task[0]['rx_task_id']) self.igui_connection.update_group_task_privileges( [self.igui_connection.igui_group_id, self.igui_task_id], "Y") self.igui_connection.update_group_task_privileges([ self.igui_connection.igui_group_id, self.igui_task_id, "update" ], "Y") self.igui_rxmap = self.igui_connection.build_igui_representation() device = self.igui_rxmap.by_id(self.igui_rx_id).devices.by_id( self.igui_device_id) task = device.tasks.by_id(self.igui_task_id) if (sensor.name[-3:] == 'PNG' ): # or some image type that we finally agreed on log.debug(sensor.name) log.debug(sensor.value) log.debug(len(sensor.value)) self.igui_connection.set_task_blob(task, reading.value) else: self.igui_connection.set_task_value(task, sensor.value) def stop(self): """ @brief Stop the client """ self.rc.stop()