class ParticipantController(object): def __init__(self, id, config_file, policy_file, logger): # participant id self.id = id # print ID for logging self.logger = logger # used to signal termination self.run = True self.prefix_lock = {} # Initialize participant params self.cfg = PConfig(config_file, self.id) # Vmac encoding mode # self.cfg.vmac_mode = config_file["vmac_mode"] # Dataplane mode---multi table or multi switch # self.cfg.dp_mode = config_file["dp_mode"] self.load_policies(policy_file) # The port 0 MAC is used for tagging outbound rules as belonging to us self.port0_mac = self.cfg.port0_mac self.nexthop_2_part = self.cfg.get_nexthop_2_part() # VNHs related params self.num_VNHs_in_use = 0 self.VNH_2_prefix = {} self.prefix_2_VNH = {} # Superset related params if self.cfg.isSupersetsMode(): self.supersets = SuperSets(self, self.cfg.vmac_options) else: # TODO: create similar class and variables for MDS self.mds = None # Keep track of flow rules pushed self.dp_pushed = [] # Keep track of flow rules which are scheduled to be pushed self.dp_queued = [] def start(self): # Start all clients/listeners/whatevs self.logger.info("Starting controller for participant") # ExaBGP Peering Instance self.bgp_instance = self.cfg.get_bgp_instance() self.logger.debug("Trace: Started controller for participant") # Route server client, Reference monitor client, Arp Proxy client self.xrs_client = self.cfg.get_xrs_client(self.logger) self.arp_client = self.cfg.get_arp_client(self.logger) self.refmon_client = self.cfg.get_refmon_client(self.logger) # class for building flow mod msgs to the reference monitor self.fm_builder = FlowModMsgBuilder(self.id, self.refmon_client.key) # Send flow rules for initial policies to the SDX's Reference Monitor self.initialize_dataplane() self.push_dp() # Start the event handler eh_socket = self.cfg.get_eh_info() self.listener_eh = Listener(eh_socket, authkey=None) self.start_eh() #ps_thread = Thread(target=self.start_eh) #ps_thread.daemon = True #ps_thread.start() def load_policies(self, policy_file): # Load policies from file with open(policy_file, 'r') as f: self.policies = json.load(f) port_count = len(self.cfg.ports) # sanitize the input policies if 'inbound' in self.policies: for policy in self.policies['inbound']: if 'action' not in policy: continue if 'fwd' in policy['action'] and int(policy['action']['fwd']) >= port_count: policy['action']['fwd'] = 0 def initialize_dataplane(self): "Read the config file and update the queued policy variable" self.logger.info("Initializing inbound rules") final_switch = "main-in" if self.cfg.isMultiTableMode(): final_switch = "main-out" self.init_vnh_assignment() rule_msgs = init_inbound_rules(self.id, self.policies, self.supersets, final_switch) self.logger.debug("Rule Messages INBOUND:: "+str(rule_msgs)) rule_msgs2 = init_outbound_rules(self, self.id, self.policies, self.supersets, final_switch) self.logger.debug("Rule Messages OUTBOUND:: "+str(rule_msgs2)) if 'changes' in rule_msgs2: if 'changes' not in rule_msgs: rule_msgs['changes'] = [] rule_msgs['changes'] += rule_msgs2['changes'] #TODO: Initialize Outbound Policies from RIB self.logger.debug("Rule Messages:: "+str(rule_msgs)) if 'changes' in rule_msgs: self.dp_queued.extend(rule_msgs["changes"]) def push_dp(self): ''' (1) Check if there are any policies queued to be pushed (2) Send the queued policies to reference monitor ''' self.logger.debug("Pushing current flow mod queue:") # it is crucial that dp_queued is traversed chronologically for flowmod in self.dp_queued: self.logger.debug("MOD: "+str(flowmod)) self.fm_builder.add_flow_mod(**flowmod) self.dp_pushed.append(flowmod) self.dp_queued = [] self.refmon_client.send(json.dumps(self.fm_builder.get_msg())) def stop(self): "Stop the Participants' SDN Controller" self.logger.info("Stopping Controller. "+str(self.logger_info)) # Signal Termination and close blocking listener self.run = False conn = Client(self.cfg.get_eh_info(), authkey=None) conn.send("terminate") conn.close() # TODO: confirm that this isn't silly self.xrs_client = None self.refmon_client = None self.arp_client = None # TODO: Think of better way of terminating this listener self.listener_eh.close() def start_eh(self): '''Socket listener for network events ''' self.logger.info("Event Handler started.") while self.run: self.logger.debug("EH waiting for connection...") conn_eh = self.listener_eh.accept() tmp = conn_eh.recv() if tmp != "terminate": self.logger.debug("EH established connection...") data = json.loads(tmp) self.logger.debug("Event received of type "+str(data.keys())) # Starting a thread for independently processing each incoming network event event_processor_thread = Thread(target = self.process_event, args = [data]) event_processor_thread.daemon = True event_processor_thread.start() # Send a message back to the sender. reply = "Event Received" conn_eh.send(reply) conn_eh.close() def process_event(self, data): "Locally process each incoming network event" if 'bgp' in data: self.logger.debug("Event Received: BGP Update.") route = data['bgp'] # Process the incoming BGP updates from XRS #self.logger.debug("BGP Route received: "+str(route)+" "+str(type(route))) self.process_bgp_route(route) elif 'policy' in data: # Process the event requesting change of participants' policies self.logger.debug("Event Received: Policy change.") change_info = data['policy'] self.process_policy_changes(change_info) elif 'arp' in data: (requester_srcmac, requested_vnh) = tuple(data['arp']) self.logger.debug("Event Received: ARP request for IP "+str(requested_vnh)) self.process_arp_request(requester_srcmac, requested_vnh) else: self.logger.warn("UNKNOWN EVENT TYPE RECEIVED: "+str(data)) def process_policy_changes(self, change_info): "Process the changes in participants' policies" # TODO: Implement the logic of dynamically changing participants' outbound and inbound policy ''' change_info = { 'removal_cookies' : [cookie1, ...], # Cookies of deleted policies 'new_policies' : { <policy file format> } } ''' # remove flow rules for the old policies removal_msgs = [] ''' for cookie in change_info['removal_cookies']: mod = {"rule_type":"outbound", "priority":0, "match":match_args , "action":{}, "cookie":cookie, "mod_type":"remove"} removal_msgs.append(mod) ''' self.dp_queued.extend(removal_msgs) # add flow rules for the new policies if self.cfg.isSupersetsMode(): dp_msgs = ss_process_policy_change(self.supersets, add_policies, remove_policies, policies, self.port_count, self.port0_mac) else: dp_msgs = [] self.dp_queued.extend(dp_msgs) self.push_dp() return 0 def process_arp_request(self, part_mac, vnh): vmac = "" if self.cfg.isSupersetsMode(): vmac = self.supersets.get_vmac(self, vnh) else: vmac = "whoa" # MDS vmac goes here arp_responses = list() # if this is gratuitous, send a reply to the part's ID if part_mac is None: gratuitous = True # set fields appropriately for gratuitous arps i = 0 for port in self.cfg.ports: eth_dst = vmac_part_port_match(self.id, i, self.supersets, False) arp_responses.append({'SPA': vnh, 'TPA': vnh, 'SHA': vmac, 'THA': vmac, 'eth_src': vmac, 'eth_dst': eth_dst}) i += 1 else: # if it wasn't gratuitous gratuitous = False # dig up the IP of the target participant for port in self.cfg.ports: if part_mac == port["MAC"]: part_ip = port["IP"] break # set field appropriately for arp responses arp_responses.append({'SPA': vnh, 'TPA': part_ip, 'SHA': vmac, 'THA': part_mac, 'eth_src': vmac, 'eth_dst': part_mac}) if gratuitous: self.logger.debug("Sending Gratuitious ARP: "+str(arp_responses)) else: self.logger.debug("Sending ARP Response: "+str(arp_responses)) for arp_response in arp_responses: self.arp_client.send(json.dumps(arp_response)) def getlock(self, prefixes): prefixes.sort() hsh = "-".join(prefixes) if hsh not in self.prefix_lock: #self.logger.debug("First Lock:: "+str(hsh)) self.prefix_lock[hsh] = RLock() #else: #self.logger.debug("Repeat :: "+str(hsh)) return self.prefix_lock[hsh] def process_bgp_route(self, route): "Process each incoming BGP advertisement" tstart = time.time() prefixes = get_prefixes_from_announcements(route) with self.getlock(prefixes): reply = '' # Map to update for each prefix in the route advertisement. updates = self.bgp_instance.update(route) #self.logger.debug("process_bgp_route:: "+str(updates)) # TODO: This step should be parallelized # TODO: The decision process for these prefixes is going to be same, we # should think about getting rid of such redundant computations. for update in updates: self.bgp_instance.decision_process_local(update) self.vnh_assignment(update) if TIMING: elapsed = time.time() - tstart self.logger.debug("Time taken for decision process: "+str(elapsed)) tstart = time.time() if self.cfg.isSupersetsMode(): ################## SUPERSET RESPONSE TO BGP ################## # update supersets "Map the set of BGP updates to a list of superset expansions." ss_changes, ss_changed_prefs = self.supersets.update_supersets(self, updates) if TIMING: elapsed = time.time() - tstart self.logger.debug("Time taken to update supersets: "+str(elapsed)) tstart = time.time() # ss_changed_prefs are prefixes for which the VMAC bits have changed # these prefixes must have gratuitous arps sent garp_required_vnhs = [self.prefix_2_VNH[prefix] for prefix in ss_changed_prefs] "If a recomputation event was needed, wipe out the flow rules." if ss_changes["type"] == "new": self.logger.debug("Wiping outbound rules.") wipe_msgs = msg_clear_all_outbound(self.policies, self.port0_mac) self.dp_queued.extend(wipe_msgs) #if a recomputation was needed, all VMACs must be reARPed # TODO: confirm reARPed is a word garp_required_vnhs = self.VNH_2_prefix.keys() if len(ss_changes['changes']) > 0: self.logger.debug("Supersets have changed: "+str(ss_changes)) "Map the superset changes to a list of new flow rules." flow_msgs = update_outbound_rules(ss_changes, self.policies, self.supersets, self.port0_mac) self.logger.debug("Flow msgs: "+str(flow_msgs)) "Dump the new rules into the dataplane queue." self.dp_queued.extend(flow_msgs) if TIMING: elapsed = time.time() - tstart self.logger.debug("Time taken to deal with ss_changes: "+str(elapsed)) tstart = time.time() ################## END SUPERSET RESPONSE ################## else: # TODO: similar logic for MDS self.logger.debug("Creating ctrlr messages for MDS scheme") self.push_dp() if TIMING: elapsed = time.time() - tstart self.logger.debug("Time taken to push dp msgs: "+str(elapsed)) tstart = time.time() changed_vnhs, announcements = self.bgp_instance.bgp_update_peers(updates, self.prefix_2_VNH, self.cfg.ports) """ Combine the VNHs which have changed BGP default routes with the VNHs which have changed supersets. """ changed_vnhs = set(changed_vnhs) changed_vnhs.update(garp_required_vnhs) # Send gratuitous ARP responses for all them for vnh in changed_vnhs: self.process_arp_request(None, vnh) # Tell Route Server that it needs to announce these routes for announcement in announcements: # TODO: Complete the logic for this function self.send_announcement(announcement) if TIMING: elapsed = time.time() - tstart self.logger.debug("Time taken to send garps/announcements: "+str(elapsed)) tstart = time.time() return reply def send_announcement(self, announcement): "Send the announcements to XRS" self.logger.debug("Sending announcements to XRS. "+str(type(announcement))) self.xrs_client.send(json.dumps(announcement)) def vnh_assignment(self, update): "Assign VNHs for the advertised prefixes" if self.cfg.isSupersetsMode(): " Superset" # TODO: Do we really need to assign a VNH for each advertised prefix? if ('announce' in update): prefix = update['announce'].prefix if (prefix not in self.prefix_2_VNH): # get next VNH and assign it the prefix self.num_VNHs_in_use += 1 vnh = str(self.cfg.VNHs[self.num_VNHs_in_use]) self.prefix_2_VNH[prefix] = vnh self.VNH_2_prefix[vnh] = prefix else: "Disjoint" # TODO: @Robert: Place your logic here for VNH assignment for MDS scheme self.logger.debug("VNH assignment called for disjoint vmac_mode") def init_vnh_assignment(self): "Assign VNHs for the advertised prefixes" if self.cfg.isSupersetsMode(): " Superset" # TODO: Do we really need to assign a VNH for each advertised prefix? #self.bgp_instance.rib["local"].dump() prefixes = self.bgp_instance.rib["local"].get_prefixes() #print 'init_vnh_assignment: prefixes:', prefixes #print 'init_vnh_assignment: prefix_2_VNH:', self.prefix_2_VNH for prefix in prefixes: if (prefix not in self.prefix_2_VNH): # get next VNH and assign it the prefix self.num_VNHs_in_use += 1 vnh = str(self.cfg.VNHs[self.num_VNHs_in_use]) self.prefix_2_VNH[prefix] = vnh self.VNH_2_prefix[vnh] = prefix else: "Disjoint" # TODO: @Robert: Place your logic here for VNH assignment for MDS scheme self.logger.debug("VNH assignment called for disjoint vmac_mode")