Beispiel #1
0
def main():
    results = Results(COMMON_CONFIG.line_length)
    results.add_to_summary(0, "=")
    results.add_to_summary(2, "STATUS", "SUBTEST")
    results.add_to_summary(0, "=")

    openstack_nodes = test_utils.get_nodes()
    installer_type = str(os.environ['INSTALLER_TYPE'].lower())
    # node.is_odl() doesn't work in Apex
    # https://jira.opnfv.org/browse/RELENG-192
    fuel_cmd = "sudo systemctl status opendaylight"
    apex_cmd = "sudo docker exec opendaylight_api " \
               "/opt/opendaylight/bin/status"
    health_cmd = "sudo docker ps -f name=opendaylight_api -f " \
                 "health=healthy -q"
    if installer_type in ["fuel"]:
        controllers = [
            node for node in openstack_nodes
            if "running" in node.run_cmd(fuel_cmd)
        ]
    elif installer_type in ["apex"]:
        controllers = [
            node for node in openstack_nodes if node.run_cmd(health_cmd)
            if "Running" in node.run_cmd(apex_cmd)
        ]

    msg = ("Verify that all OpenStack nodes OVS br-int have "
           "fail_mode set to secure")
    results.record_action(msg)
    results.add_to_summary(0, "-")
    if not controllers:
        msg = ("Controller (ODL) list is empty. Skipping rest of tests.")
        logger.info(msg)
        results.add_failure(msg)
        return results.compile_summary()
    else:
        msg = ("Controller (ODL) list is ready")
        logger.info(msg)
        results.add_success(msg)
    # Get fail_mode status on all nodes
    fail_mode_statuses = test_utils.is_fail_mode_secure()
    for node_name, status in fail_mode_statuses.iteritems():
        msg = 'Node {} br-int is fail_mode secure'.format(node_name)
        if status:
            results.add_success(msg)
        else:
            results.add_failure(msg)

    return results.compile_summary()
Beispiel #2
0
def gather_logs(name):
    openstack_nodes = test_utils.get_nodes()

    ft_utils.execute_command_raise('rm -rf /tmp/sdnvpn-logs/;'
                                   'mkdir -p /tmp/sdnvpn-logs/')
    for node in openstack_nodes:
        node.put_file('%s/../sh_utils/fetch-log-script.sh' % LIB_PATH,
                      '/tmp/fetch-log-script.sh')
        node.run_cmd('sudo bash /tmp/fetch-log-script.sh')
        node.get_file('/tmp/log_output.tar.gz',
                      '/tmp/log_output-%s.tar.gz' % node.get_dict()['name'])
        ft_utils.execute_command_raise('mkdir -p /tmp/sdnvpn-logs/')
        ft_utils.execute_command_raise(
            'cd /tmp/sdnvpn-logs/; '
            'tar -xzvf /tmp/log_output-%s.tar.gz --strip-components=1' %
            node.get_dict()['name'])

    ft_utils.execute_command_raise(
        'cd %s;tar czvf sdnvpn-logs-%s.tar.gz /tmp/sdnvpn-logs/' %
        (getattr(config.CONF, 'dir_results'), name))
Beispiel #3
0
def main():
    results = Results(COMMON_CONFIG.line_length)
    results.add_to_summary(0, "=")
    results.add_to_summary(2, "STATUS", "SUBTEST")
    results.add_to_summary(0, "=")

    openstack_nodes = test_utils.get_nodes()

    # node.is_odl() doesn't work in Apex
    # https://jira.opnfv.org/browse/RELENG-192
    controllers = [
        node for node in openstack_nodes
        if "running" in node.run_cmd("sudo systemctl status opendaylight")
    ]

    msg = ("Verify that all OpenStack nodes OVS br-int have "
           "fail_mode set to secure")
    results.record_action(msg)
    results.add_to_summary(0, "-")
    if not controllers:
        msg = ("Controller (ODL) list is empty. Skipping rest of tests.")
        logger.info(msg)
        results.add_failure(msg)
        return results.compile_summary()
    else:
        msg = ("Controller (ODL) list is ready")
        logger.info(msg)
        results.add_success(msg)
    # Get fail_mode status on all nodes
    fail_mode_statuses = test_utils.is_fail_mode_secure()
    for node_name, status in fail_mode_statuses.iteritems():
        msg = 'Node {} br-int is fail_mode secure'.format(node_name)
        if status:
            results.add_success(msg)
        else:
            results.add_failure(msg)

    return results.compile_summary()
Beispiel #4
0
def main():
    results = Results(COMMON_CONFIG.line_length)
    results.add_to_summary(0, "=")
    results.add_to_summary(2, "STATUS", "SUBTEST")
    results.add_to_summary(0, "=")

    openstack_nodes = test_utils.get_nodes()

    # node.is_odl() doesn't work in Apex
    # https://jira.opnfv.org/browse/RELENG-192
    controllers = [
        node for node in openstack_nodes
        if "running" in node.run_cmd("sudo systemctl status opendaylight")
    ]
    computes = [node for node in openstack_nodes if node.is_compute()]

    msg = ("Verify that OpenDaylight can start/communicate with zrpcd/Quagga")
    results.record_action(msg)
    results.add_to_summary(0, "-")
    if not controllers:
        msg = ("Controller (ODL) list is empty. Skipping rest of tests.")
        logger.info(msg)
        results.add_failure(msg)
        return results.compile_summary()
    else:
        msg = ("Controller (ODL) list is ready")
        logger.info(msg)
        results.add_success(msg)

    controller = controllers[0]  # We don't handle HA well
    get_ext_ip_cmd = "sudo ip a | grep br-ex | grep inet | awk '{print $2}'"
    ext_net_cidr = controller.run_cmd(get_ext_ip_cmd).strip().split('\n')
    ext_net_mask = ext_net_cidr[0].split('/')[1]
    controller_ext_ip = ext_net_cidr[0].split('/')[0]

    logger.info("Starting bgp speaker of controller at IP %s " %
                controller_ext_ip)
    logger.info("Checking if zrpcd is " "running on the controller node")

    output_zrpcd = controller.run_cmd("ps --no-headers -C " "zrpcd -o state")
    states = output_zrpcd.split()
    running = any([s != 'Z' for s in states])

    msg = ("zrpcd is running")

    if not running:
        logger.info("zrpcd is not running on the controller node")
        results.add_failure(msg)
    else:
        logger.info("zrpcd is running on the controller node")
        results.add_success(msg)

    results.add_to_summary(0, "-")

    # Ensure that ZRPCD ip & port are well configured within ODL
    add_client_conn_to_bgp = "bgp-connect -p 7644 -h 127.0.0.1 add"
    test_utils.run_odl_cmd(controller, add_client_conn_to_bgp)

    # Start bgp daemon
    start_quagga = "odl:configure-bgp -op start-bgp-server " \
                   "--as-num 100 --router-id {0}".format(controller_ext_ip)
    test_utils.run_odl_cmd(controller, start_quagga)

    logger.info("Checking if bgpd is running" " on the controller node")

    # Check if there is a non-zombie bgpd process
    output_bgpd = controller.run_cmd("ps --no-headers -C " "bgpd -o state")
    states = output_bgpd.split()
    running = any([s != 'Z' for s in states])

    msg = ("bgpd is running")
    if not running:
        logger.info("bgpd is not running on the controller node")
        results.add_failure(msg)
    else:
        logger.info("bgpd is running on the controller node")
        results.add_success(msg)

    results.add_to_summary(0, "-")

    # We should be able to restart the speaker
    # but the test is disabled because of buggy upstream
    # https://github.com/6WIND/zrpcd/issues/15
    # stop_quagga = 'odl:configure-bgp -op stop-bgp-server'
    # test_utils.run_odl_cmd(controller, stop_quagga)

    # logger.info("Checking if bgpd is still running"
    #             " on the controller node")

    # output_bgpd = controller.run_cmd("ps --no-headers -C " \
    #                                  "bgpd -o state")
    # states = output_bgpd.split()
    # running = any([s != 'Z' for s in states])

    # msg = ("bgpd is stopped")
    # if not running:
    #     logger.info("bgpd is not running on the controller node")
    #     results.add_success(msg)
    # else:
    #     logger.info("bgpd is still running on the controller node")
    #     results.add_failure(msg)

    # Taken from the sfc tests
    if not os.path.isfile(COMMON_CONFIG.ubuntu_image_path):
        logger.info("Downloading image")
        os_utils.download_url(
            "http://artifacts.opnfv.org/sdnvpn/"
            "ubuntu-16.04-server-cloudimg-amd64-disk1.img",
            "/home/opnfv/functest/data/")
    else:
        logger.info("Using old image")

    glance_client = os_utils.get_glance_client()
    nova_client = os_utils.get_nova_client()
    neutron_client = os_utils.get_neutron_client()

    (floatingip_ids, instance_ids, router_ids, network_ids, image_ids,
     subnet_ids, interfaces, bgpvpn_ids, flavor_ids) = ([] for i in range(9))

    try:
        _, flavor_id = test_utils.create_custom_flavor()
        flavor_ids.append(flavor_id)

        sg_id = os_utils.create_security_group_full(
            neutron_client, TESTCASE_CONFIG.secgroup_name,
            TESTCASE_CONFIG.secgroup_descr)
        test_utils.open_icmp(neutron_client, sg_id)
        test_utils.open_http_port(neutron_client, sg_id)

        test_utils.open_bgp_port(neutron_client, sg_id)
        net_id, subnet_1_id, router_1_id = test_utils.create_network(
            neutron_client, TESTCASE_CONFIG.net_1_name,
            TESTCASE_CONFIG.subnet_1_name, TESTCASE_CONFIG.subnet_1_cidr,
            TESTCASE_CONFIG.router_1_name)

        quagga_net_id, subnet_quagga_id, \
            router_quagga_id = test_utils.create_network(
                neutron_client,
                TESTCASE_CONFIG.quagga_net_name,
                TESTCASE_CONFIG.quagga_subnet_name,
                TESTCASE_CONFIG.quagga_subnet_cidr,
                TESTCASE_CONFIG.quagga_router_name)

        interfaces.append(tuple((router_1_id, subnet_1_id)))
        interfaces.append(tuple((router_quagga_id, subnet_quagga_id)))
        network_ids.extend([net_id, quagga_net_id])
        router_ids.extend([router_1_id, router_quagga_id])
        subnet_ids.extend([subnet_1_id, subnet_quagga_id])

        installer_type = str(os.environ['INSTALLER_TYPE'].lower())
        if installer_type == "fuel":
            disk = 'raw'
        elif installer_type == "apex":
            disk = 'qcow2'
        else:
            logger.error("Incompatible installer type")

        ubuntu_image_id = os_utils.create_glance_image(
            glance_client,
            COMMON_CONFIG.ubuntu_image_name,
            COMMON_CONFIG.ubuntu_image_path,
            disk,
            container="bare",
            public="public")

        image_ids.append(ubuntu_image_id)

        # NOTE(rski) The order of this seems a bit weird but
        # there is a reason for this, namely
        # https://jira.opnfv.org/projects/SDNVPN/issues/SDNVPN-99
        # so we create the quagga instance using cloud-init
        # and immediately give it a floating IP.
        # The cloud-init script should contain a small sleep for
        # this to work.
        # We also create the FIP first because it is used in the
        # cloud-init script.
        fip = os_utils.create_floating_ip(neutron_client)
        # fake_fip is needed to bypass NAT
        # see below for the reason why.
        fake_fip = os_utils.create_floating_ip(neutron_client)

        floatingip_ids.extend([fip['fip_id'], fake_fip['fip_id']])
        # pin quagga to some compute
        compute_node = nova_client.hypervisors.list()[0]
        quagga_compute_node = "nova:" + compute_node.hypervisor_hostname
        # Map the hypervisor used above to a compute handle
        # returned by releng's manager
        for comp in computes:
            if compute_node.host_ip in comp.run_cmd("sudo ip a"):
                compute = comp
                break
        quagga_bootstrap_script = quagga.gen_quagga_setup_script(
            controller_ext_ip, fake_fip['fip_addr'], ext_net_mask)

        quagga_vm = test_utils.create_instance(
            nova_client,
            TESTCASE_CONFIG.quagga_instance_name,
            ubuntu_image_id,
            quagga_net_id,
            sg_id,
            fixed_ip=TESTCASE_CONFIG.quagga_instance_ip,
            flavor=COMMON_CONFIG.custom_flavor_name,
            userdata=quagga_bootstrap_script,
            compute_node=quagga_compute_node)

        instance_ids.append(quagga_vm)

        fip_added = os_utils.add_floating_ip(nova_client, quagga_vm.id,
                                             fip['fip_addr'])

        msg = ("Assign a Floating IP to %s " %
               TESTCASE_CONFIG.quagga_instance_name)
        if fip_added:
            results.add_success(msg)
        else:
            results.add_failure(msg)
        test_utils.attach_instance_to_ext_br(quagga_vm, compute)

        try:
            testcase = "Bootstrap quagga inside an OpenStack instance"
            cloud_init_success = test_utils.wait_for_cloud_init(quagga_vm)
            if cloud_init_success:
                results.add_success(testcase)
            else:
                results.add_failure(testcase)
            results.add_to_summary(0, "=")

            results.add_to_summary(0, '-')
            results.add_to_summary(1, "Peer Quagga with OpenDaylight")
            results.add_to_summary(0, '-')

            neighbor = quagga.odl_add_neighbor(fake_fip['fip_addr'],
                                               controller_ext_ip, controller)
            peer = quagga.check_for_peering(controller)

        finally:
            test_utils.detach_instance_from_ext_br(quagga_vm, compute)

        if neighbor and peer:
            results.add_success("Peering with quagga")
        else:
            results.add_failure("Peering with quagga")

    except Exception as e:
        logger.error("exception occurred while executing testcase_3: %s", e)
        raise
    finally:
        test_utils.cleanup_nova(nova_client, instance_ids, flavor_ids)
        test_utils.cleanup_glance(glance_client, image_ids)
        test_utils.cleanup_neutron(neutron_client, floatingip_ids, bgpvpn_ids,
                                   interfaces, subnet_ids, router_ids,
                                   network_ids)

    return results.compile_summary()
Beispiel #5
0
def main():
    results = Results(COMMON_CONFIG.line_length)

    results.add_to_summary(0, "=")
    results.add_to_summary(2, "STATUS", "SUBTEST")
    results.add_to_summary(0, "=")

    nova_client = os_utils.get_nova_client()
    neutron_client = os_utils.get_neutron_client()
    glance_client = os_utils.get_glance_client()
    openstack_nodes = test_utils.get_nodes()

    (floatingip_ids, instance_ids, router_ids, network_ids, image_ids,
     subnet_ids, interfaces, bgpvpn_ids) = ([] for i in range(8))

    try:
        image_id = os_utils.create_glance_image(
            glance_client, TESTCASE_CONFIG.image_name,
            COMMON_CONFIG.image_path, disk=COMMON_CONFIG.image_format,
            container="bare", public='public')
        image_ids.append(image_id)

        network_1_id = test_utils.create_net(neutron_client,
                                             TESTCASE_CONFIG.net_1_name)
        subnet_1_id = test_utils.create_subnet(neutron_client,
                                               TESTCASE_CONFIG.subnet_1_name,
                                               TESTCASE_CONFIG.subnet_1_cidr,
                                               network_1_id)

        network_ids.append(network_1_id)
        subnet_ids.append(subnet_1_id)

        sg_id = os_utils.create_security_group_full(
            neutron_client, TESTCASE_CONFIG.secgroup_name,
            TESTCASE_CONFIG.secgroup_descr)

        # Check required number of compute nodes
        compute_hostname = (
            nova_client.hypervisors.list()[0].hypervisor_hostname)
        compute_nodes = [node for node in openstack_nodes
                         if node.is_compute()]

        av_zone_1 = "nova:" + compute_hostname
        # List of OVS bridges to get groups
        ovs_br = "br-int"
        # Get a list of flows and groups, before start topology
        initial_ovs_flows = len(test_utils.get_ovs_flows(compute_nodes,
                                                         [ovs_br]))
        initial_ovs_groups = len(test_utils.get_ovs_groups(compute_nodes,
                                                           [ovs_br]))

        # boot INSTANCES
        vm_2 = test_utils.create_instance(
            nova_client,
            TESTCASE_CONFIG.instance_2_name,
            image_id,
            network_1_id,
            sg_id,
            secgroup_name=TESTCASE_CONFIG.secgroup_name,
            compute_node=av_zone_1)

        vm_1 = test_utils.create_instance(
            nova_client,
            TESTCASE_CONFIG.instance_1_name,
            image_id,
            network_1_id,
            sg_id,
            secgroup_name=TESTCASE_CONFIG.secgroup_name,
            compute_node=av_zone_1)
        instance_ids.extend([vm_1.id, vm_2.id])

        # Wait for VMs to get ips.
        instances_up = test_utils.wait_for_instances_up(vm_1, vm_2)

        if not instances_up:
            logger.error("One or more instances is down")

        logger.info("Wait before subtest")
        test_utils.wait_before_subtest()
        # Get added OVS flows and groups
        added_ovs_flows = len(test_utils.get_ovs_flows(compute_nodes,
                                                       [ovs_br]))
        added_ovs_groups = len(test_utils.get_ovs_groups(compute_nodes,
                                                         [ovs_br]))
        # Check if flows and groups added successfully
        results.record_action("Check if new flows and groups were added "
                              "to OVS")

        msg = "New OVS flows added"
        results.add_to_summary(0, "-")
        if added_ovs_flows - initial_ovs_flows > 0:
            results.add_success(msg)
        else:
            results.add_failure(msg)
        results.add_to_summary(0, "=")

        msg = "New OVS groups added"
        results.add_to_summary(0, "-")
        if added_ovs_groups - initial_ovs_groups > 0:
            results.add_success(msg)
        else:
            results.add_failure(msg)
        results.add_to_summary(0, "=")

        get_ext_ip_cmd = "sudo ovs-vsctl get-controller {}".format(ovs_br)
        ovs_controller_conn = (compute_nodes[0].run_cmd(get_ext_ip_cmd).
                               strip().split('\n')[0])

        for compute_node in compute_nodes:
            # Disconnect OVS from controller
            compute_node.run_cmd("sudo ovs-vsctl del-controller {}".
                                 format(ovs_br))
            test_utils.wait_before_subtest()
            # Connect again OVS to Controller
            compute_node.run_cmd("sudo ovs-vsctl set-controller {} {}".
                                 format(ovs_br, ovs_controller_conn))

        logger.info("Wait before subtest resync type 1")
        test_utils.wait_before_subtest()
        # Get OVS flows added after the reconnection
        resynced_ovs_flows = len(test_utils.get_ovs_flows(
            compute_nodes, [ovs_br]))
        # Get OVS groups added after the reconnection
        resynced_ovs_groups = len(test_utils.get_ovs_groups(
            compute_nodes, [ovs_br]))

        record_action_msg = ("Check if flows/groups are reprogrammed in OVS "
                             "after its reconnection by del/set controller.")
        record_test_result(added_ovs_flows, resynced_ovs_flows,
                           added_ovs_groups, resynced_ovs_groups,
                           record_action_msg, results)

        for compute_node in compute_nodes:
            # Disconnect OVS from controller
            compute_node.run_cmd("sudo iptables -A OUTPUT -p tcp --dport 6653"
                                 " -j DROP")
            test_utils.wait_before_subtest()
            # Connect again OVS to Controller
            compute_node.run_cmd("sudo iptables -D OUTPUT -p tcp --dport 6653"
                                 " -j DROP")

        logger.info("Wait before subtest resync type 2")
        test_utils.wait_before_subtest()
        # Get OVS flows added after the reconnection
        resynced_ovs_flows = len(test_utils.get_ovs_flows(
            compute_nodes, [ovs_br]))
        # Get OVS groups added after the reconnection
        resynced_ovs_groups = len(test_utils.get_ovs_groups(
            compute_nodes, [ovs_br]))

        record_action_msg = ("Check if flows/groups are reprogrammed in OVS "
                             "after its reconnection by firewall rule for "
                             "OF port block/unblok")
        record_test_result(added_ovs_flows, resynced_ovs_flows,
                           added_ovs_groups, resynced_ovs_groups,
                           record_action_msg, results)

    except Exception as e:
        logger.error("exception occurred while executing testcase_12: %s", e)
        raise
    finally:
        # Cleanup topology
        test_utils.cleanup_nova(nova_client, instance_ids)
        test_utils.cleanup_glance(glance_client, image_ids)
        test_utils.cleanup_neutron(neutron_client, floatingip_ids, bgpvpn_ids,
                                   interfaces, subnet_ids, router_ids,
                                   network_ids)

    return results.compile_summary()
Beispiel #6
0
def main():
    conn = os_utils.get_os_connection()
    results = Results(COMMON_CONFIG.line_length, conn)
    results.add_to_summary(0, "=")
    results.add_to_summary(2, "STATUS", "SUBTEST")
    results.add_to_summary(0, "=")

    openstack_nodes = test_utils.get_nodes()
    installer_type = str(os.environ['INSTALLER_TYPE'].lower())

    # node.is_odl() doesn't work in Apex
    # https://jira.opnfv.org/browse/RELENG-192
    fuel_cmd = "sudo systemctl status opendaylight"
    apex_cmd = "sudo docker exec opendaylight_api " \
               "/opt/opendaylight/bin/status"
    health_cmd = "sudo docker ps -f name=opendaylight_api -f " \
                 "health=healthy -q"
    if installer_type in ["fuel"]:
        controllers = [
            node for node in openstack_nodes
            if "running" in node.run_cmd(fuel_cmd)
        ]
    elif installer_type in ["apex"]:
        controllers = [
            node for node in openstack_nodes if node.run_cmd(health_cmd)
            if "Running" in node.run_cmd(apex_cmd)
        ]

    computes = [node for node in openstack_nodes if node.is_compute()]

    msg = ("Verify that OpenDaylight can start/communicate with zrpcd/Quagga")
    results.record_action(msg)
    results.add_to_summary(0, "-")
    if not controllers:
        msg = ("Controller (ODL) list is empty. Skipping rest of tests.")
        logger.info(msg)
        results.add_failure(msg)
        return results.compile_summary()
    else:
        msg = ("Controller (ODL) list is ready")
        logger.info(msg)
        results.add_success(msg)

    logger.info("Checking if zrpcd is " "running on the controller nodes")

    for controller in controllers:
        output_zrpcd = controller.run_cmd("ps --no-headers -C "
                                          "zrpcd -o state")
        states = output_zrpcd.split()
        running = any([s != 'Z' for s in states])
        msg = ("zrpcd is running in {name}".format(name=controller.name))

        if not running:
            logger.info(
                "zrpcd is not running on the controller node {name}".format(
                    name=controller.name))
            results.add_failure(msg)
        else:
            logger.info(
                "zrpcd is running on the controller node {name}".format(
                    name=controller.name))
            results.add_success(msg)

        results.add_to_summary(0, "-")

    # Find the BGP entity owner in ODL because of this bug:
    # https://jira.opendaylight.org/browse/NETVIRT-1308
    msg = ("Found BGP entity owner")
    controller = test_utils.get_odl_bgp_entity_owner(controllers)
    if controller is None:
        logger.error("Failed to find the BGP entity owner")
        results.add_failure(msg)
    else:
        logger.info('BGP entity owner is {name}'.format(name=controller.name))
        results.add_success(msg)
    results.add_to_summary(0, "-")

    get_ext_ip_cmd = "sudo ip a | grep br-ex | grep inet | awk '{print $2}'"
    ext_net_cidr = controller.run_cmd(get_ext_ip_cmd).strip().split('\n')
    ext_net_mask = ext_net_cidr[0].split('/')[1]
    controller_ext_ip = ext_net_cidr[0].split('/')[0]

    logger.info("Starting bgp speaker of controller at IP %s " %
                controller_ext_ip)

    # Ensure that ZRPCD ip & port are well configured within ODL
    add_client_conn_to_bgp = "bgp-connect -p 7644 -h 127.0.0.1 add"
    test_utils.run_odl_cmd(controller, add_client_conn_to_bgp)

    # Start bgp daemon
    start_quagga = "odl:configure-bgp -op start-bgp-server " \
                   "--as-num 100 --router-id {0}".format(controller_ext_ip)
    test_utils.run_odl_cmd(controller, start_quagga)

    # we need to wait a bit until the bgpd is up
    time.sleep(5)

    logger.info("Checking if bgpd is running" " on the controller node")

    # Check if there is a non-zombie bgpd process
    output_bgpd = controller.run_cmd("ps --no-headers -C " "bgpd -o state")
    states = output_bgpd.split()
    running = any([s != 'Z' for s in states])

    msg = ("bgpd is running")
    if not running:
        logger.info("bgpd is not running on the controller node")
        results.add_failure(msg)
    else:
        logger.info("bgpd is running on the controller node")
        results.add_success(msg)

    results.add_to_summary(0, "-")

    # We should be able to restart the speaker
    # but the test is disabled because of buggy upstream
    # https://github.com/6WIND/zrpcd/issues/15
    # stop_quagga = 'odl:configure-bgp -op stop-bgp-server'
    # test_utils.run_odl_cmd(controller, stop_quagga)

    # logger.info("Checking if bgpd is still running"
    #             " on the controller node")

    # output_bgpd = controller.run_cmd("ps --no-headers -C " \
    #                                  "bgpd -o state")
    # states = output_bgpd.split()
    # running = any([s != 'Z' for s in states])

    # msg = ("bgpd is stopped")
    # if not running:
    #     logger.info("bgpd is not running on the controller node")
    #     results.add_success(msg)
    # else:
    #     logger.info("bgpd is still running on the controller node")
    #     results.add_failure(msg)

    # Taken from the sfc tests
    if not os.path.isfile(COMMON_CONFIG.ubuntu_image_path):
        logger.info("Downloading image")
        image_dest_path = '/'.join(
            COMMON_CONFIG.ubuntu_image_path.split('/')[:-1])
        os_utils.download_url(
            "http://artifacts.opnfv.org/sdnvpn/"
            "ubuntu-16.04-server-cloudimg-amd64-disk1.img", image_dest_path)
    else:
        logger.info("Using old image")

    conn = os_utils.get_os_connection()
    neutron_client = os_utils.get_neutron_client()

    (floatingip_ids, instance_ids, router_ids, network_ids, image_ids,
     subnet_ids, interfaces, bgpvpn_ids, flavor_ids) = ([] for i in range(9))
    quagga_vm = None
    fake_fip = None

    try:
        _, flavor_id = test_utils.create_custom_flavor()
        flavor_ids.append(flavor_id)

        sg_id = os_utils.create_security_group_full(
            neutron_client, TESTCASE_CONFIG.secgroup_name,
            TESTCASE_CONFIG.secgroup_descr)
        test_utils.open_icmp(neutron_client, sg_id)
        test_utils.open_http_port(neutron_client, sg_id)

        test_utils.open_bgp_port(neutron_client, sg_id)

        image_id = os_utils.create_glance_image(
            conn,
            TESTCASE_CONFIG.image_name,
            COMMON_CONFIG.image_path,
            disk=COMMON_CONFIG.image_format,
            container="bare",
            public='public')
        image_ids.append(image_id)

        net_1_id, subnet_1_id, router_1_id = test_utils.create_network(
            neutron_client, TESTCASE_CONFIG.net_1_name,
            TESTCASE_CONFIG.subnet_1_name, TESTCASE_CONFIG.subnet_1_cidr,
            TESTCASE_CONFIG.router_1_name)

        quagga_net_id, subnet_quagga_id, \
            router_quagga_id = test_utils.create_network(
                neutron_client,
                TESTCASE_CONFIG.quagga_net_name,
                TESTCASE_CONFIG.quagga_subnet_name,
                TESTCASE_CONFIG.quagga_subnet_cidr,
                TESTCASE_CONFIG.quagga_router_name)

        interfaces.append(tuple((router_1_id, subnet_1_id)))
        interfaces.append(tuple((router_quagga_id, subnet_quagga_id)))
        network_ids.extend([net_1_id, quagga_net_id])
        router_ids.extend([router_1_id, router_quagga_id])
        subnet_ids.extend([subnet_1_id, subnet_quagga_id])

        installer_type = str(os.environ['INSTALLER_TYPE'].lower())
        if installer_type == "fuel":
            disk = 'raw'
        elif installer_type == "apex":
            disk = 'qcow2'
        else:
            logger.error("Incompatible installer type")

        ubuntu_image_id = os_utils.create_glance_image(
            conn,
            COMMON_CONFIG.ubuntu_image_name,
            COMMON_CONFIG.ubuntu_image_path,
            disk,
            container="bare",
            public="public")

        image_ids.append(ubuntu_image_id)

        # NOTE(rski) The order of this seems a bit weird but
        # there is a reason for this, namely
        # https://jira.opnfv.org/projects/SDNVPN/issues/SDNVPN-99
        # so we create the quagga instance using cloud-init
        # and immediately give it a floating IP.
        # The cloud-init script should contain a small sleep for
        # this to work.
        # We also create the FIP first because it is used in the
        # cloud-init script.
        # fake_fip is needed to bypass NAT
        # see below for the reason why.
        fake_fip = os_utils.create_floating_ip(neutron_client)
        # pin quagga to some compute
        floatingip_ids.append(fake_fip['fip_id'])
        compute_node = conn.compute.hypervisors().next()
        compute_node = conn.compute.get_hypervisor(compute_node)
        quagga_compute_node = "nova:" + compute_node.name
        # Map the hypervisor used above to a compute handle
        # returned by releng's manager
        for comp in computes:
            if compute_node.host_ip in comp.run_cmd("sudo ip a"):
                compute = comp
                break
        quagga_bootstrap_script = quagga.gen_quagga_setup_script(
            controller_ext_ip, fake_fip['fip_addr'], ext_net_mask,
            TESTCASE_CONFIG.external_network_ip_prefix,
            TESTCASE_CONFIG.route_distinguishers,
            TESTCASE_CONFIG.import_targets, TESTCASE_CONFIG.export_targets)

        quagga_vm = test_utils.create_instance(
            conn,
            TESTCASE_CONFIG.quagga_instance_name,
            ubuntu_image_id,
            quagga_net_id,
            sg_id,
            fixed_ip=TESTCASE_CONFIG.quagga_instance_ip,
            flavor=COMMON_CONFIG.custom_flavor_name,
            userdata=quagga_bootstrap_script,
            compute_node=quagga_compute_node)

        instance_ids.append(quagga_vm.id)

        quagga_vm_port = test_utils.get_port(neutron_client, quagga_vm.id)
        fip_added = os_utils.attach_floating_ip(neutron_client,
                                                quagga_vm_port['id'])

        msg = ("Assign a Floating IP to %s " %
               TESTCASE_CONFIG.quagga_instance_name)
        if fip_added:
            results.add_success(msg)
            floatingip_ids.append(fip_added['floatingip']['id'])
        else:
            results.add_failure(msg)

        test_utils.attach_instance_to_ext_br(quagga_vm, compute)

        testcase = "Bootstrap quagga inside an OpenStack instance"
        cloud_init_success = test_utils.wait_for_cloud_init(conn, quagga_vm)
        if cloud_init_success:
            results.add_success(testcase)
        else:
            results.add_failure(testcase)
        results.add_to_summary(0, "=")

        results.add_to_summary(0, '-')
        results.add_to_summary(1, "Peer Quagga with OpenDaylight")
        results.add_to_summary(0, '-')

        neighbor = quagga.odl_add_neighbor(fake_fip['fip_addr'],
                                           controller_ext_ip, controller)
        peer = quagga.check_for_peering(controller)

        if neighbor and peer:
            results.add_success("Peering with quagga")
        else:
            results.add_failure("Peering with quagga")

        test_utils.add_quagga_external_gre_end_point(controllers,
                                                     fake_fip['fip_addr'])
        test_utils.wait_before_subtest()

        msg = ("Create VPN to define a VRF")
        results.record_action(msg)
        vpn_name = vpn_name = "sdnvpn-3"
        kwargs = {
            "import_targets": TESTCASE_CONFIG.import_targets,
            "export_targets": TESTCASE_CONFIG.export_targets,
            "route_targets": TESTCASE_CONFIG.route_targets,
            "route_distinguishers": TESTCASE_CONFIG.route_distinguishers,
            "name": vpn_name
        }
        bgpvpn = test_utils.create_bgpvpn(neutron_client, **kwargs)
        bgpvpn_id = bgpvpn['bgpvpn']['id']
        logger.debug("VPN1 created details: %s" % bgpvpn)
        bgpvpn_ids.append(bgpvpn_id)

        msg = ("Associate network '%s' to the VPN." %
               TESTCASE_CONFIG.net_1_name)
        results.record_action(msg)
        results.add_to_summary(0, "-")

        # create a vm and connect it with network1,
        # which is going to be bgpvpn associated
        userdata_common = test_utils.generate_ping_userdata(
            [TESTCASE_CONFIG.external_network_ip])

        compute_node = conn.compute.hypervisors().next()
        av_zone_1 = "nova:" + compute_node.name
        vm_bgpvpn = test_utils.create_instance(
            conn,
            TESTCASE_CONFIG.instance_1_name,
            image_id,
            net_1_id,
            sg_id,
            fixed_ip=TESTCASE_CONFIG.instance_1_ip,
            secgroup_name=TESTCASE_CONFIG.secgroup_name,
            compute_node=av_zone_1,
            userdata=userdata_common)
        instance_ids.append(vm_bgpvpn.id)

        # wait for VM to get IP
        instance_up = test_utils.wait_for_instances_up(vm_bgpvpn)
        if not instance_up:
            logger.error("One or more instances are down")

        test_utils.create_network_association(neutron_client, bgpvpn_id,
                                              net_1_id)

        test_utils.wait_before_subtest()

        msg = ("External IP prefix %s is exchanged with ODL" %
               TESTCASE_CONFIG.external_network_ip_prefix)
        fib_added = test_utils.is_fib_entry_present_on_odl(
            controllers, TESTCASE_CONFIG.external_network_ip_prefix,
            TESTCASE_CONFIG.route_distinguishers)
        if fib_added:
            results.add_success(msg)
        else:
            results.add_failure(msg)

        # TODO: uncomment the following once OVS is installed with > 2.8.3 and
        # underlay connectivity is established between vxlan overlay and
        # external network.
        # results.get_ping_status_target_ip(
        #    vm_bgpvpn,
        #    TESTCASE_CONFIG.external_network_name,
        #    TESTCASE_CONFIG.external_network_ip,
        #    expected="PASS",
        #    timeout=300)

        results.add_to_summary(0, "=")
        logger.info("\n%s" % results.summary)

    except Exception as e:
        logger.error("exception occurred while executing testcase_3: %s", e)
        raise
    finally:
        if quagga_vm is not None:
            test_utils.detach_instance_from_ext_br(quagga_vm, compute)
        test_utils.cleanup_nova(conn, instance_ids, flavor_ids)
        test_utils.cleanup_glance(conn, image_ids)
        test_utils.cleanup_neutron(neutron_client, floatingip_ids, bgpvpn_ids,
                                   interfaces, subnet_ids, router_ids,
                                   network_ids)
        if fake_fip is not None:
            bgp_nbr_disconnect_cmd = ("bgp-nbr -i %s -a 200 del" %
                                      fake_fip['fip_addr'])
            test_utils.run_odl_cmd(controller, bgp_nbr_disconnect_cmd)
        bgp_server_stop_cmd = ("bgp-rtr -r %s -a 100 del" % controller_ext_ip)
        odl_zrpc_disconnect_cmd = "bgp-connect -p 7644 -h 127.0.0.1 del"
        test_utils.run_odl_cmd(controller, bgp_server_stop_cmd)
        test_utils.run_odl_cmd(controller, odl_zrpc_disconnect_cmd)

    return results.compile_summary()
Beispiel #7
0
def main():
    results = Results(COMMON_CONFIG.line_length)

    results.add_to_summary(0, "=")
    results.add_to_summary(2, "STATUS", "SUBTEST")
    results.add_to_summary(0, "=")

    nova_client = os_utils.get_nova_client()
    neutron_client = os_utils.get_neutron_client()
    glance_client = os_utils.get_glance_client()
    openstack_nodes = test_utils.get_nodes()

    (floatingip_ids, instance_ids, router_ids, network_ids, image_ids,
     subnet_ids, interfaces, bgpvpn_ids) = ([] for i in range(8))

    try:
        image_id = os_utils.create_glance_image(
            glance_client, TESTCASE_CONFIG.image_name,
            COMMON_CONFIG.image_path, disk=COMMON_CONFIG.image_format,
            container="bare", public='public')
        image_ids.append(image_id)

        network_1_id = test_utils.create_net(neutron_client,
                                             TESTCASE_CONFIG.net_1_name)
        subnet_1_id = test_utils.create_subnet(neutron_client,
                                               TESTCASE_CONFIG.subnet_1_name,
                                               TESTCASE_CONFIG.subnet_1_cidr,
                                               network_1_id)

        network_ids.append(network_1_id)
        subnet_ids.append(subnet_1_id)

        sg_id = os_utils.create_security_group_full(
            neutron_client, TESTCASE_CONFIG.secgroup_name,
            TESTCASE_CONFIG.secgroup_descr)

        # Check required number of compute nodes
        compute_hostname = (
            nova_client.hypervisors.list()[0].hypervisor_hostname)
        compute_nodes = [node for node in openstack_nodes
                         if node.is_compute()]

        av_zone_1 = "nova:" + compute_hostname
        # List of OVS bridges to get groups
        ovs_br = "br-int"
        # Get a list of groups, before start topology
        initial_ovs_groups = test_utils.get_ovs_groups(compute_nodes,
                                                       [ovs_br])

        # boot INSTANCES
        vm_2 = test_utils.create_instance(
            nova_client,
            TESTCASE_CONFIG.instance_2_name,
            image_id,
            network_1_id,
            sg_id,
            secgroup_name=TESTCASE_CONFIG.secgroup_name,
            compute_node=av_zone_1)

        vm_1 = test_utils.create_instance(
            nova_client,
            TESTCASE_CONFIG.instance_1_name,
            image_id,
            network_1_id,
            sg_id,
            secgroup_name=TESTCASE_CONFIG.secgroup_name,
            compute_node=av_zone_1)
        instance_ids.extend([vm_1.id, vm_2.id])

        # Wait for VMs to get ips.
        instances_up = test_utils.wait_for_instances_up(vm_1, vm_2)

        if not instances_up:
            logger.error("One or more instances is down")
            # TODO: Handle this appropriately

        logger.info("Wait before subtest")
        test_utils.wait_before_subtest()
        # Get added OVS groups
        added_ovs_groups = (len(initial_ovs_groups) -
                            len(test_utils.get_ovs_groups(
                                compute_nodes, [ovs_br])))
        # Check if group added successfully
        results.record_action("Check if a new group was added to OVS")
        msg = "New OVS group added"
        results.add_to_summary(0, "-")
        if added_ovs_groups != 0:
            results.add_success(msg)
        else:
            results.add_failure(msg)
        results.add_to_summary(0, "=")
        # Backup OVS controller connection info.
        # To support HA changes should be made here.
        get_ext_ip_cmd = "sudo ovs-vsctl get-controller {}".format(ovs_br)
        ovs_controller_conn = (compute_nodes[0].run_cmd(get_ext_ip_cmd).
                               strip().split('\n')[0])
        # Disconnect OVS from controller
        for compute_node in compute_nodes:
            compute_node.run_cmd("sudo ovs-vsctl del-controller {}".
                                 format(ovs_br))
    except Exception as e:
        logger.error("exception occurred while executing testcase_1: %s", e)
        raise
    finally:
        # Cleanup topology
        test_utils.cleanup_nova(nova_client, instance_ids)
        test_utils.cleanup_glance(glance_client, image_ids)
        test_utils.cleanup_neutron(neutron_client, floatingip_ids, bgpvpn_ids,
                                   interfaces, subnet_ids, router_ids,
                                   network_ids)
    # Connect again OVS to Controller
    for compute_node in compute_nodes:
        compute_node.run_cmd("sudo ovs-vsctl set-controller {} {}".
                             format(ovs_br, ovs_controller_conn))
    logger.info("Wait before subtest")
    test_utils.wait_before_subtest()
    # Get OVS groups added after the reconnection
    added_ovs_groups = (len(initial_ovs_groups) -
                        len(test_utils.get_ovs_groups(
                            compute_nodes, [ovs_br])))

    # Check if group removed successfully
    results.record_action("Check if group was removed from OVS "
                          "after deleting the topology.")
    msg = ""
    # After removing the topology, groups must be equal to the initial
    if added_ovs_groups != 0:
        msg += " Additional group was not deleted from OVS"
    results.add_to_summary(0, "-")
    if len(msg) == 0:
        msg = "Group was deleted from ovs"
        results.add_success(msg)
    else:
        results.add_failure(msg)

    return results.compile_summary()