def __proj_dependent_synapse_sdram(self, incoming_projections): """ Get the SDRAM used by synapse cores dependent on the projections :param list(~spynnaker.pyNN.models.Projection) incoming_projections: The projections to consider in the calculations :rtype: ~pacman.model.resources.MultiRegionSDRAM """ app_vertex = self._governed_app_vertex sdram = MultiRegionSDRAM() sdram.add_cost( PopulationSynapsesMachineVertexLead.SYNAPSE_REGIONS.pop_table, max( MasterPopTableAsBinarySearch.get_master_population_table_size( incoming_projections), BYTES_PER_WORD)) sdram.add_cost( PopulationSynapsesMachineVertexLead.SYNAPSE_REGIONS. connection_builder, max(app_vertex.get_synapse_expander_size(incoming_projections), BYTES_PER_WORD)) sdram.add_cost( PopulationSynapsesMachineVertexLead.SYNAPSE_REGIONS. bitfield_filter, max(get_estimated_sdram_for_bit_field_region(incoming_projections), BYTES_PER_WORD)) sdram.add_cost( PopulationSynapsesMachineVertexLead.SYNAPSE_REGIONS. bitfield_key_map, max(get_estimated_sdram_for_key_region(incoming_projections), BYTES_PER_WORD)) return sdram
def __get_synapse_constant_sdram(self, vertex_slice): """ Get the amount of fixed SDRAM used by synapse parts :param ~pacman.model.graphs.common.Slice vertex_slice: The slice of neurons to get the size of :rtype: ~pacman.model.resources.MultiRegionSDRAM """ sdram = MultiRegionSDRAM() app_vertex = self._governed_app_vertex sdram.add_cost( PopulationMachineVertex.SYNAPSE_REGIONS.synapse_params, app_vertex.get_synapse_params_size()) sdram.add_cost( PopulationMachineVertex.SYNAPSE_REGIONS.synapse_dynamics, app_vertex.get_synapse_dynamics_size(vertex_slice)) sdram.add_cost( PopulationMachineVertex.SYNAPSE_REGIONS.structural_dynamics, self.__structural_size(vertex_slice)) sdram.add_cost( PopulationMachineVertex.SYNAPSE_REGIONS.synaptic_matrix, self.__all_syn_block_size(vertex_slice)) sdram.add_cost( PopulationMachineVertex.SYNAPSE_REGIONS.direct_matrix, app_vertex.all_single_syn_size) sdram.add_cost( PopulationMachineVertex.SYNAPSE_REGIONS.pop_table, MasterPopTableAsBinarySearch.get_master_population_table_size( app_vertex.incoming_projections)) sdram.add_cost( PopulationMachineVertex.SYNAPSE_REGIONS.connection_builder, self.__synapse_expander_size()) sdram.merge(self.__bitfield_size()) return sdram
def test_get_max_row_length(dynamics_class, timing, weight, size, exception, max_size): MockSimulator.setup() if timing is not None and weight is not None: dynamics = dynamics_class(timing(), weight()) else: dynamics = dynamics_class() io = SynapseIORowBased() population_table = MasterPopTableAsBinarySearch() synapse_information = SynapseInformation(None, None, None, None, None, None, dynamics, 0) in_edge = ProjectionApplicationEdge(None, None, synapse_information) if exception is not None: with pytest.raises(exception) as exc_info: io._get_max_row_length(size, dynamics, population_table, in_edge, size) assert exc_info.value.max_size == max_size else: actual_size = io._get_max_row_length(size, dynamics, population_table, in_edge, size) assert actual_size == max_size
def __init__(self, post_vertex_slice, n_synapse_types, all_single_syn_sz, synaptic_matrix_region, direct_matrix_region, poptable_region, connection_builder_region, synaptic_matrix_ref=None, direct_matrix_ref=None, poptable_ref=None, connection_builder_ref=None): """ :param ~pacman.model.graphs.common.Slice post_vertex_slice: The slice of the post vertex that these matrices are for :param int n_synapse_types: The number of synapse types available :param int all_single_syn_sz: The space available for "direct" or "single" synapses :param int synaptic_matrix_region: The region where synaptic matrices are stored :param int direct_matrix_region: The region where "direct" or "single" synapses are stored :param int poptable_region: The region where the population table is stored :param int connection_builder_region: The region where the synapse generator information is stored :param synaptic_matrix_ref: The reference to the synaptic matrix region, or None if not referenceable :type synaptic_matrix_ref: int or None :param direct_matrix_ref: The reference to the direct matrix region, or None if not referenceable :type direct_matrix_ref: int or None :param poptable_ref: The reference to the pop table region, or None if not referenceable :type poptable_ref: int or None :param connection_builder_ref: The reference to the connection builder region, or None if not referenceable :type connection_builder_ref: int or None """ self.__post_vertex_slice = post_vertex_slice self.__n_synapse_types = n_synapse_types self.__all_single_syn_sz = all_single_syn_sz self.__synaptic_matrix_region = synaptic_matrix_region self.__direct_matrix_region = direct_matrix_region self.__poptable_region = poptable_region self.__connection_builder_region = connection_builder_region self.__synaptic_matrix_ref = synaptic_matrix_ref self.__direct_matrix_ref = direct_matrix_ref self.__poptable_ref = poptable_ref self.__connection_builder_ref = connection_builder_ref # Set up the master population table self.__poptable = MasterPopTableAsBinarySearch() # Map of (app_edge, synapse_info) to SynapticMatrixApp self.__matrices = dict() # Store locations of synaptic data and generated data self.__host_generated_block_addr = 0 self.__on_chip_generated_block_addr = 0 # Determine whether to generate on machine self.__gen_on_machine = False
class SynapticMatrices(object): """ Handler of synaptic matrices for a core of a population vertex """ __slots__ = [ # The slice of the post-vertex that these matrices are for "__post_vertex_slice", # The number of synapse types received "__n_synapse_types", # The maximum summed size of the "direct" or "single" matrices "__all_single_syn_sz", # The ID of the synaptic matrix region "__synaptic_matrix_region", # The ID of the "direct" or "single" matrix region "__direct_matrix_region", # The ID of the master population table region "__poptable_region", # The ID of the connection builder region "__connection_builder_region", # The master population table data structure "__poptable", # The sub-matrices for each incoming edge "__matrices", # The address within the synaptic matrix region after the last matrix # was written "__host_generated_block_addr", # The address within the synaptic matrix region after the last # generated matrix will be written "__on_chip_generated_block_addr", # Determine if any of the matrices can be generated on the machine "__gen_on_machine", # Reference to give the synaptic matrix "__synaptic_matrix_ref", # Reference to give the direct matrix "__direct_matrix_ref", # Reference to give the master population table "__poptable_ref", # Reference to give the connection builder "__connection_builder_ref" ] def __init__(self, post_vertex_slice, n_synapse_types, all_single_syn_sz, synaptic_matrix_region, direct_matrix_region, poptable_region, connection_builder_region, synaptic_matrix_ref=None, direct_matrix_ref=None, poptable_ref=None, connection_builder_ref=None): """ :param ~pacman.model.graphs.common.Slice post_vertex_slice: The slice of the post vertex that these matrices are for :param int n_synapse_types: The number of synapse types available :param int all_single_syn_sz: The space available for "direct" or "single" synapses :param int synaptic_matrix_region: The region where synaptic matrices are stored :param int direct_matrix_region: The region where "direct" or "single" synapses are stored :param int poptable_region: The region where the population table is stored :param int connection_builder_region: The region where the synapse generator information is stored :param synaptic_matrix_ref: The reference to the synaptic matrix region, or None if not referenceable :type synaptic_matrix_ref: int or None :param direct_matrix_ref: The reference to the direct matrix region, or None if not referenceable :type direct_matrix_ref: int or None :param poptable_ref: The reference to the pop table region, or None if not referenceable :type poptable_ref: int or None :param connection_builder_ref: The reference to the connection builder region, or None if not referenceable :type connection_builder_ref: int or None """ self.__post_vertex_slice = post_vertex_slice self.__n_synapse_types = n_synapse_types self.__all_single_syn_sz = all_single_syn_sz self.__synaptic_matrix_region = synaptic_matrix_region self.__direct_matrix_region = direct_matrix_region self.__poptable_region = poptable_region self.__connection_builder_region = connection_builder_region self.__synaptic_matrix_ref = synaptic_matrix_ref self.__direct_matrix_ref = direct_matrix_ref self.__poptable_ref = poptable_ref self.__connection_builder_ref = connection_builder_ref # Set up the master population table self.__poptable = MasterPopTableAsBinarySearch() # Map of (app_edge, synapse_info) to SynapticMatrixApp self.__matrices = dict() # Store locations of synaptic data and generated data self.__host_generated_block_addr = 0 self.__on_chip_generated_block_addr = 0 # Determine whether to generate on machine self.__gen_on_machine = False @property def host_generated_block_addr(self): """ The address within the synaptic region after the last block written by the on-host synaptic generation i.e. the start of the space that can be overwritten provided the synapse expander is run again :rtype: int """ return self.__host_generated_block_addr @property def on_chip_generated_matrix_size(self): """ The size of the space used by the generated matrix i.e. the space that can be overwritten provided the synapse expander is run again :rtype: int """ return (self.__on_chip_generated_block_addr - self.__host_generated_block_addr) def __app_matrix(self, app_edge, synapse_info): """ Get or create an application synaptic matrix object :param ProjectionApplicationEdge app_edge: The application edge to get the object for :param SynapseInformation synapse_info: The synapse information to get the object for :rtype: SynapticMatrixApp """ key = (app_edge, synapse_info) if key in self.__matrices: return self.__matrices[key] matrix = SynapticMatrixApp(self.__poptable, synapse_info, app_edge, self.__n_synapse_types, self.__all_single_syn_sz, self.__post_vertex_slice, self.__synaptic_matrix_region, self.__direct_matrix_region) self.__matrices[key] = matrix return matrix def write_synaptic_data(self, spec, incoming_projections, all_syn_block_sz, weight_scales, routing_info): """ Write the synaptic data for all incoming projections :param ~data_specification.DataSpecificationGenerator spec: The spec to write to :param list(~spynnaker8.models.Projection) incoming_projection: The projections to generate data for :param int all_syn_block_sz: The size in bytes of the space reserved for synapses :param list(float) weight_scales: The weight scale of each synapse :param ~pacman.model.routing_info.RoutingInfo routing_info: The routing information for all edges :param ~pacman.model.graphs.machine.MachineGraph machine_graph: The machine graph """ # If there are no synapses, there is nothing to do! if all_syn_block_sz == 0: return # Reserve the region spec.comment( "\nWriting Synaptic Matrix and Master Population Table:\n") spec.reserve_memory_region(region=self.__synaptic_matrix_region, size=all_syn_block_sz, label='SynBlocks', reference=self.__synaptic_matrix_ref) # Track writes inside the synaptic matrix region: block_addr = 0 self.__poptable.initialise_table() # Convert the data for convenience in_edges_by_app_edge, key_space_tracker = self.__in_edges_by_app_edge( incoming_projections, routing_info) # Set up for single synapses # The list is seeded with an empty array so we can just concatenate # later (as numpy doesn't let you concatenate nothing) single_synapses = [numpy.array([], dtype="uint32")] single_addr = 0 # Lets write some synapses spec.switch_write_focus(self.__synaptic_matrix_region) # Store a list of synapse info to be generated on the machine generate_on_machine = list() # For each machine edge in the vertex, create a synaptic list for app_edge, m_edges in in_edges_by_app_edge.items(): spec.comment("\nWriting matrix for edge:{}\n".format( app_edge.label)) app_key_info = self.__app_key_and_mask(m_edges, app_edge, routing_info, key_space_tracker) d_app_key_info = self.__delay_app_key_and_mask( m_edges, app_edge, routing_info, key_space_tracker) for synapse_info in app_edge.synapse_information: app_matrix = self.__app_matrix(app_edge, synapse_info) app_matrix.set_info(all_syn_block_sz, app_key_info, d_app_key_info, routing_info, weight_scales, m_edges) # If we can generate the connector on the machine, do so if app_matrix.can_generate_on_machine(single_addr): generate_on_machine.append(app_matrix) else: block_addr, single_addr = app_matrix.write_matrix( spec, block_addr, single_addr, single_synapses) self.__host_generated_block_addr = block_addr # Skip blocks that will be written on the machine, but add them # to the master population table generator_data = list() for app_matrix in generate_on_machine: block_addr = app_matrix.write_on_chip_matrix_data( generator_data, block_addr) self.__gen_on_machine = True self.__on_chip_generated_block_addr = block_addr # Finish the master population table self.__poptable.finish_master_pop_table(spec, self.__poptable_region, self.__poptable_ref) # Write the size and data of single synapses to the direct region single_data = numpy.concatenate(single_synapses) single_data_words = len(single_data) spec.reserve_memory_region(region=self.__direct_matrix_region, size=(single_data_words * BYTES_PER_WORD + DIRECT_MATRIX_HEADER_COST_BYTES), label='DirectMatrix', reference=self.__direct_matrix_ref) spec.switch_write_focus(self.__direct_matrix_region) spec.write_value(single_data_words * BYTES_PER_WORD) if single_data_words: spec.write_array(single_data) self.__write_synapse_expander_data_spec(spec, generator_data, weight_scales) def __write_synapse_expander_data_spec(self, spec, generator_data, weight_scales): """ Write the data spec for the synapse expander :param ~.DataSpecificationGenerator spec: The specification to write to :param list(GeneratorData) generator_data: The data to be written :param weight_scales: scaling of weights on each synapse :type weight_scales: list(int or float) """ if not generator_data: if self.__connection_builder_ref is not None: # If there is a reference, we still need a region to create spec.reserve_memory_region( region=self.__connection_builder_region, size=4, label="ConnectorBuilderRegion", reference=self.__connection_builder_ref) return n_bytes = (SYNAPSES_BASE_GENERATOR_SDRAM_USAGE_IN_BYTES + (self.__n_synapse_types * DataType.U3232.size)) for data in generator_data: n_bytes += data.size spec.reserve_memory_region(region=self.__connection_builder_region, size=n_bytes, label="ConnectorBuilderRegion", reference=self.__connection_builder_ref) spec.switch_write_focus(self.__connection_builder_region) spec.write_value(self.__synaptic_matrix_region) spec.write_value(len(generator_data)) spec.write_value(self.__post_vertex_slice.lo_atom) spec.write_value(self.__post_vertex_slice.n_atoms) spec.write_value(self.__n_synapse_types) spec.write_value(get_n_bits(self.__n_synapse_types)) n_neuron_id_bits = get_n_bits(self.__post_vertex_slice.n_atoms) spec.write_value(n_neuron_id_bits) for w in weight_scales: # if the weights are high enough and the population size large # enough, then weight_scales < 1 will result in a zero scale # if converted to an int, so we use U3232 here instead (as there # can be scales larger than U1616.max in conductance-based models) dtype = DataType.U3232 spec.write_value(data=min(w, dtype.max), data_type=dtype) items = list() for data in generator_data: items.extend(data.gen_data) spec.write_array(numpy.concatenate(items)) def __in_edges_by_app_edge(self, incoming_projections, routing_info): """ Convert a list of incoming projections to a dict of application edge -> list of machine edges, and a key tracker :param list(~spynnaker.pyNN.models.Projection) incoming_projections: The incoming projections :param RoutingInfo routing_info: Routing information for all edges :rtype: tuple(dict, KeySpaceTracker) """ in_edges_by_app_edge = defaultdict(OrderedSet) key_space_tracker = KeySpaceTracker() pre_vertices = set() for proj in incoming_projections: app_edge = proj._projection_edge # Skip if already done if app_edge in in_edges_by_app_edge: continue # Add all incoming machine edges for this slice for machine_edge in app_edge.machine_edges: if (machine_edge.post_vertex.vertex_slice == self.__post_vertex_slice): if machine_edge.pre_vertex in pre_vertices: continue pre_vertices.add(machine_edge.pre_vertex) rinfo = routing_info.get_routing_info_for_edge( machine_edge) key_space_tracker.allocate_keys(rinfo) in_edges_by_app_edge[app_edge].add(machine_edge) # Also go through the delay edges in case an undelayed edge # was filtered delay_edge = app_edge.delay_edge if delay_edge is not None: for machine_edge in delay_edge.machine_edges: if (machine_edge.post_vertex.vertex_slice == self.__post_vertex_slice): if machine_edge.pre_vertex in pre_vertices: continue pre_vertices.add(machine_edge.pre_vertex) rinfo = routing_info.get_routing_info_for_edge( machine_edge) key_space_tracker.allocate_keys(rinfo) undelayed_machine_edge = (app_edge.get_machine_edge( machine_edge.pre_vertex, machine_edge.post_vertex)) in_edges_by_app_edge[app_edge].add( undelayed_machine_edge) return in_edges_by_app_edge, key_space_tracker @staticmethod def __check_keys_adjacent(keys, mask_size): """ Check that a given list of keys and slices have no gaps between them :param list(tuple(int, Slice)) keys: A list of keys and slices to check :param mask_size: The number of 0s in the mask :rtype: bool """ key_increment = (1 << mask_size) last_key = None last_slice = None for i, (key, v_slice) in enumerate(keys): # If the first round, we can skip the checks and just store if last_key is not None: # Fail if next key is not adjacent to last key if (last_key + key_increment) != key: return False # Fail if this is not the last key and the number of atoms # don't match the other keys (last is OK to be different) elif ((i + 1) < len(keys) and last_slice.n_atoms != v_slice.n_atoms): return False # Fail if the atoms are not adjacent elif (last_slice.hi_atom + 1) != v_slice.lo_atom: return False # Store for the next round last_key = key last_slice = v_slice # Pass if nothing failed return True def __get_app_key_and_mask(self, keys, mask, n_stages, key_space_tracker): """ Get a key and mask for an incoming application vertex as a whole,\ or say it isn't possible (return None) :param list(tuple(int, Slice)) keys: The key and slice of each relevant machine vertex in the incoming application vertex :param int mask: The mask that covers all keys :param n_stages: The number of delay stages :param key_space_tracker: A key space tracker that has been filled in with all keys this vertex will receive :rtype: None or _AppKeyInfo """ # Can be merged only if keys are adjacent outside the mask keys = sorted(keys, key=lambda item: item[0]) mask_size = KeySpaceTracker.count_trailing_0s(mask) if not self.__check_keys_adjacent(keys, mask_size): return None # Get the key as the first key and the mask as the mask that covers # enough keys key = keys[0][0] n_extra_mask_bits = int(math.ceil(math.log(len(keys), 2))) core_mask = (2**n_extra_mask_bits) - 1 new_mask = mask & ~(core_mask << mask_size) # Final check because adjacent keys don't mean they all fit under a # single mask if key & new_mask != key: return None # Check that the key doesn't cover other keys that it shouldn't next_key = keys[-1][0] + (2**mask_size) max_key = key + (2**(mask_size + n_extra_mask_bits)) n_unused = max_key - (next_key & mask) if n_unused > 0 and key_space_tracker.is_allocated(next_key, n_unused): return None return _AppKeyInfo(key, new_mask, core_mask, mask_size, keys[0][1].n_atoms * n_stages) def __check_key_slices(self, n_atoms, slices, delay_stages=1): """ Check if a list of slices cover all n_atoms without any gaps :param int n_atoms: The total number of atoms expected :param list(Slice) slices: The list of slices to check :rtype: bool """ slices = sorted(slices, key=lambda s: s.lo_atom) slice_atoms = slices[-1].hi_atom - slices[0].lo_atom + 1 if slice_atoms != n_atoms: return False # Check that all slices are also there in between, and that all are # the same size (except the last one) next_high = 0 n_atoms_per_core = None last_slice = slices[-1] for s in slices: if s.lo_atom != next_high: return False if (n_atoms_per_core is not None and s != last_slice and n_atoms_per_core != s.n_atoms): return None next_high = s.hi_atom + 1 if n_atoms_per_core is None: n_atoms_per_core = s.n_atoms # If the number of atoms per core is too big, this can't be done if ((n_atoms_per_core * delay_stages) > self.__poptable.max_n_neurons_per_core): return False return True def __app_key_and_mask(self, m_edges, app_edge, routing_info, key_space_tracker): """ Get a key and mask for an incoming application vertex as a whole,\ or say it isn't possible (return None) :param list(PopulationMachineEdge) m_edges: The relevant machine edges of the application edge :param PopulationApplicationEdge app_edge: The application edge to get the key and mask of :param RoutingInfo routing_info: The routing information of all edges :param KeySpaceTracker key_space_tracker: A tracker pre-filled with the keys of all incoming edges """ # If there are too many pre-cores, give up now if len(m_edges) > self.__poptable.max_core_mask: return None # Work out if the keys allow the machine vertices to be merged mask = None keys = list() # Can be merged only if all the masks are the same pre_slices = list() for m_edge in m_edges: rinfo = routing_info.get_routing_info_for_edge(m_edge) vertex_slice = m_edge.pre_vertex.vertex_slice pre_slices.append(vertex_slice) # No routing info at all? Must have been filtered, so doesn't work if rinfo is None: return None # Mask is not the same as the last mask? Doesn't work if mask is not None and rinfo.first_mask != mask: return None mask = rinfo.first_mask keys.append((rinfo.first_key, vertex_slice)) if mask is None: return None if not self.__check_key_slices(app_edge.pre_vertex.n_atoms, pre_slices): return None return self.__get_app_key_and_mask(keys, mask, 1, key_space_tracker) def __delay_app_key_and_mask(self, m_edges, app_edge, routing_info, key_space_tracker): """ Get a key and mask for a whole incoming delayed application\ vertex, or say it isn't possible (return None) :param list(PopulationMachineEdge) m_edges: The relevant machine edges of the application edge :param PopulationApplicationEdge app_edge: The application edge to get the key and mask of :param RoutingInfo routing_info: The routing information of all edges :param KeySpaceTracker key_space_tracker: A tracker pre-filled with the keys of all incoming edges """ # Work out if the keys allow the machine vertices to be # merged mask = None keys = list() # Can be merged only if all the masks are the same pre_slices = list() for m_edge in m_edges: # If the edge doesn't have a delay edge, give up delayed_app_edge = m_edge.app_edge.delay_edge if delayed_app_edge is None: return None delayed_machine_edge = delayed_app_edge.get_machine_edge( m_edge.pre_vertex, m_edge.post_vertex) if delayed_machine_edge is None: return None rinfo = routing_info.get_routing_info_for_edge( delayed_machine_edge) vertex_slice = m_edge.pre_vertex.vertex_slice pre_slices.append(vertex_slice) # No routing info at all? Must have been filtered, so doesn't work if rinfo is None: return None # Mask is not the same as the last mask? Doesn't work if mask is not None and rinfo.first_mask != mask: return None mask = rinfo.first_mask keys.append((rinfo.first_key, vertex_slice)) if not self.__check_key_slices(app_edge.pre_vertex.n_atoms, pre_slices, app_edge.n_delay_stages): return None return self.__get_app_key_and_mask(keys, mask, app_edge.n_delay_stages, key_space_tracker) def get_connections_from_machine(self, transceiver, placement, app_edge, synapse_info): """ Get the synaptic connections from the machine :param ~spinnman.transceiver.Transceiver transceiver: Used to read the data from the machine :param ~pacman.model.placements.Placement placement: Where the vertices are on the machine :param ProjectionApplicationEdge app_edge: The application edge of the projection :param SynapseInformation synapse_info: The synapse information of the projection :return: A list of arrays of connections, each with dtype AbstractSynapseDynamics.NUMPY_CONNECTORS_DTYPE :rtype: ~numpy.ndarray """ matrix = self.__app_matrix(app_edge, synapse_info) return matrix.get_connections(transceiver, placement) def read_generated_connection_holders(self, transceiver, placement): """ Fill in any pre-run connection holders for data which is generated on the machine, after it has been generated :param ~spinnman.transceiver.Transceiver transceiver: How to read the data from the machine :param ~pacman.model.placements.Placement placement: where the data is to be read from """ for matrix in self.__matrices.values(): matrix.read_generated_connection_holders(transceiver, placement) def clear_connection_cache(self): """ Clear any values read from the machine """ for matrix in self.__matrices.values(): matrix.clear_connection_cache() @property def gen_on_machine(self): """ Whether any matrices need to be generated on the machine :rtype: bool """ return self.__gen_on_machine def get_index(self, app_edge, synapse_info, machine_edge): """ Get the index of an incoming projection in the population table :param ProjectionApplicationEdge app_edge: The application edge of the projection :param SynapseInformation synapse_info: The synapse information of the projection :param ~pacman.model.graphs.machine.MachineEdge machine_edge: The machine edge to get the index of """ matrix = self.__app_matrix(app_edge, synapse_info) return matrix.get_index(machine_edge)