def generate_trust_ids(users_to_delete):
    """
    From a list of users to delete, generate a file with a trustid for each
    user. The user is acting as the trustor, delegating in a trustee, which
    will impersonate it to delete its resources.

    :param users_to_delete: a list of trustors.
    :return: this function does not return anything. It creates a file.
    """
    global logger

    osclients = OpenStackClients()
    users_trusted_ids = open('users_trusted_ids.txt', 'w')
    check_users = CheckUsers()

    # Use an alternative URL that allow direct access to the keystone admin
    # endpoint, because the registered one uses an internal IP address.

    osclients.override_endpoint(
        'identity', osclients.region, 'admin', KEYSTONE_ENDPOINT)

    trust_factory = TrustFactory(osclients)
    lines = users_to_delete.readlines()
    total = len(lines)
    count = 0
    if 'TRUSTEE_USER' in env:
        trustee = env['TRUSTEE_USER']
    else:
        trustee = TRUSTEE

    for user in lines:
        user = user.strip()
        if user == '':
            continue
        try:
            count += 1
            (username, trust_id) = trust_factory.create_trust_admin(
                user, trustee)
            users_trusted_ids.write(username + ',' + trust_id + '\n')
            msg = 'Generated trustid for user {0} ({1}/{2})'
            logger.info(msg.format(user, count, total))
        except Exception, e:
            msg = 'Failed getting trust-id from trustor {0}. Reason: {1}'
            logger.error(msg.format(user, str(e)))
Example #2
0
class OpenStackMap(object):
    """
    This class build a map from the resources (VMs, networks, images,
    volumes, users, tenants, roles...) in an OpenStack infrastructure.

    This map is optionally cached on disk and can be accessed offline or can be
    used as a metadata snapshot of an OpenStack infrastructure (e.g. to
    analyse resource use and build statistics)
    """

    # objects strategy see the __init__ method documentation
    DIRECT_OBJECTS, NO_CACHE_OBJECTS, REFRESH_OBJECTS, USE_CACHE_OBJECTS, USE_CACHE_OBJECTS_ONLY = \
        range(5)

    # If use_wrapper is True, dictionaries are wrapped to allow access to
    # resource['field'] also as resource.field. This is not used when
    # objects_strategy is DIRECT_OBJECTS
    use_wrapper = True

    load_filters = True

    resources_region = ['vms', 'images', 'routers', 'networks', 'subnets', 'ports', 'floatingips',
                        'security_groups', 'volumes', 'volume_backups', 'volume_snapshots']

    def __init__(
            self, persistence_dir='~/openstackmap', region=None, auth_url=None,
            objects_strategy=USE_CACHE_OBJECTS, auto_load=True):
        """
        Constructor
        :param persistence_dir: The path where the data is saved. Ignored if
           objects_strategy is DIRECT_OBJECTS or NO_CACHE_OBJECTS
        :param region: the initial region (if undefined, use OS_REGION_NAME)
        :param auth_url: the keystone URI
        :param objects_strategy: sets the strategy about the object maps
            contained here. It can be:
         * DIRECT_OBJECTS is using the objects as they are got from
           the API. Be careful because methods as delete are available! The
           objects are not cached.
         * NO_CACHE_OBJECTS is using the objects converted to dictionaries, so
           methods to do operations are not available. The objects are not
           cached.
         * REFRESH_OBJECTS is using the objects converted to dictionaries. The
           objects are cached: the new version replace the old one.
         * USE_CACHE_OBJECTS is using the objects converted to dictionaries. If
           a cached copy of the objects are available, it is used.
         * USE_CACHE_OBJECTS_ONLY is using the objects converted to dictionaries.
           This strategy used cached objects only. It never contacts with the
           servers, even when the object is not available in the local cache.

        :param auto_load: if True, invoke self.load_all()
         Note that neutron objects returned by the API are already dictionaries
        """

        self.logger = logging.getLogger(__name__)

        if auth_url:
            self.osclients = OpenStackClients(auth_url=auth_url)
        else:
            self.osclients = OpenStackClients()

        if region:
            self.osclients.set_region(region)
        else:
            if 'OS_REGION_NAME' not in env:
                raise Exception('Region parameter must be provided or '
                                'OS_REGION_NAME variable must be defined.')
            else:
                region = env['OS_REGION_NAME']

        if 'KEYSTONE_ADMIN_ENDPOINT' in os.environ:
            self.osclients.override_endpoint(
                'identity', self.osclients.region, 'admin',
                os.environ['KEYSTONE_ADMIN_ENDPOINT'])

        self.objects_strategy = objects_strategy

        self.persistence_dir = os.path.expanduser(persistence_dir)
        self.pers_region = self.persistence_dir + '/' + region
        self.pers_keystone = self.persistence_dir + '/keystone'

        if objects_strategy not in (OpenStackMap.DIRECT_OBJECTS,
                                    OpenStackMap.NO_CACHE_OBJECTS):
            if not os.path.exists(self.persistence_dir):
                os.mkdir(self.persistence_dir)

            if not os.path.exists(self.pers_keystone):
                os.mkdir(self.pers_keystone)

            if not os.path.exists(self.pers_region):
                os.mkdir(self.pers_region)

        self._init_resource_maps()

        if auto_load:
            self.load_all()

        self.region_map = dict()

    def _init_resource_maps(self):
        """init all the resources that will be available
        as empty dictionaries"""
        # Keystone resources
        self.users = dict()
        self.users_by_name = dict()
        self.roles = dict()
        self.tenants = dict()
        self.tenants_by_name = dict()
        self.roles_a = list()
        self.roles_by_project = dict()
        self.roles_by_user = dict()
        self.filters = dict()
        self.filters_by_project = dict()

        # Glance resources
        self.images = dict()
        # Neutron resources
        self.networks = dict()
        self.subnets = dict()
        self.routers = dict()
        self.floatingips = dict()
        self.floatingips_by_ip = dict()
        self.ports = dict()
        self.security_groups = dict()
        # Nova resources
        self.vms = dict()
        self.flavors = dict()
        # Cinder resources
        self.volumes = dict()
        self.volume_backups = dict()
        self.volume_snapshots = dict()

    def _load(self, name):
        """Load the resources persisted with pickle. This resources (e.g.
        networks, vms, images...) are saved independently for each region.

        This method is called for load_cinder, load_nova, load_glance,
        load_neutron...

        :param name: the resource name
        :return: a dictionary of objects (dictionaries) indexed by id
        """
        objects = pickle.load(open(self.pers_region + '/' + name +
                                   '.pickle', 'rb'))
        return self._convert(objects)

    def _load_fkeystone(self, name):
        """Load the keystone objects persisted with pickle. This resources are
        shared among the regions. This method is used in load_keystone

        :param name: the resource name
        :return: a list/dictionary of objects (dictionaries)
        """
        objects = pickle.load(open(self.pers_keystone + '/' + name +
                                   '.pickle', 'rb'))
        return self._convert(objects)

    def _convert(self, objects):
        """if use_wrapper, convert objects from dictionary to __E class;
        this class allows accessing  a['key'] also as a.key"""
        if self.use_wrapper and len(objects) > 0:
            class __E(dict):
                def __init__(self, d=dict()):
                    self.__dict__ = self
                    dict.__init__(self, d)

            if isinstance(objects, dict):
                if not isinstance(objects.values()[0], dict):
                    return objects
                return dict((key, __E(objects[key])) for key in objects)
            else:
                return list(__E(object) for object in objects)

        else:
            return objects

    def _get_keystone_data(self):
        """get data from keystone server"""
        dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS
        save = dict_object and \
            self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS
        keystone = self.osclients.get_keystoneclientv3()
        roles = keystone.roles.list()
        users = keystone.users.list()
        tenants = keystone.projects.list()
        roles_a = keystone.role_assignments.list()

        if OpenStackMap.load_filters:
            ef = {'service_type': 'identity', 'interface': 'public'}
            resp = keystone.session.get('/OS-EP-FILTER/endpoint_groups',
                                        endpoint_filter=ef)
            filters = resp.json()['endpoint_groups']
            projects_by_filter = dict()
            filters_by_project = dict()
            for f in filters:
                filter_id = f['id']
                resp = keystone.session.get(
                    '/OS-EP-FILTER/endpoint_groups/' + filter_id +
                    '/projects', endpoint_filter=ef)
                projects = resp.json()['projects']
                projects_by_filter[filter_id] = projects
                for project in projects:
                    if project['id'] not in filters_by_project:
                        filters_by_project[project['id']] = list()
                    filters_by_project[project['id']].append(filter_id)

        else:
            filters = list()
            filters_by_project = dict()

        roles_by_user = dict()
        roles_by_project = dict()

        for roleasig in roles_a:
            try:
                userid = roleasig.user['id']
                if 'project' in roleasig.scope:
                    projectid = roleasig.scope['project']['id']
                else:
                    projectid = str(roleasig.scope)
                roleid = roleasig.role['id']
                if userid not in roles_by_user:
                    roles_by_user[userid] = list()
            except AttributeError:
                # Discard roles not assigned to users
                continue

            if projectid not in roles_by_project:
                roles_by_project[projectid] = list()
            roles_by_user[userid].append((roleid, projectid))
            roles_by_project[projectid].append((roleid, userid))

        self.roles_by_user = roles_by_user
        self.roles_by_project = roles_by_project

        tenants_by_name = dict()
        users_by_name = dict()

        filters = dict((f['id'], f) for f in filters)
        if dict_object:
            roles = dict((role.id, role.to_dict()) for role in roles)
            users = dict((user.id, user.to_dict()) for user in users)
            tenants = dict((tenant.id, tenant.to_dict()) for tenant in tenants)
            roles_a = list(asig.to_dict() for asig in roles_a)
            for tenant in tenants.values():
                tenants_by_name[tenant['name']] = tenant
            for user in users.values():
                users_by_name[user['name']] = user
        else:
            self.roles = dict((role.id, role) for role in roles)
            self.users = dict((user.id, user) for user in users)
            self.tenants = dict((tenant.id, tenant) for tenant in tenants)
            self.roles_a = roles_a

            tenants_by_name = dict()
            for tenant in tenants:
                if getattr(tenant, 'name', None):
                    tenants_by_name[tenant.name] = tenant

            for user in users:
                if getattr(user, 'name', None):
                    users_by_name[user.name] = user

            self.tenants_by_name = tenants_by_name
            self.users_by_name = users_by_name
            self.filters = filters
            self.filters_by_project = filters_by_project
            return

        if save:
            with open(self.pers_keystone + '/roles.pickle', 'wb') as f:
                pickle.dump(roles, f, protocol=-1)

            with open(self.pers_keystone + '/users.pickle', 'wb') as f:
                pickle.dump(users, f, protocol=-1)

            with open(self.pers_keystone + '/users_by_name.pickle', 'wb') as f:
                pickle.dump(users_by_name, f, protocol=-1)

            with open(self.pers_keystone + '/tenants.pickle', 'wb') as f:
                pickle.dump(tenants, f, protocol=-1)

            with open(self.pers_keystone + '/tenants_by_name.pickle', 'wb') as\
                    f:
                pickle.dump(tenants_by_name, f)

            with open(self.pers_keystone + '/roles_a.pickle', 'wb') as f:
                pickle.dump(roles_a, f, protocol=-1)

            with open(self.pers_keystone + '/roles_by_user.pickle', 'wb') as f:
                pickle.dump(roles_by_user, f, protocol=-1)

            with open(self.pers_keystone + '/roles_by_project.pickle', 'wb') as\
                    f:
                pickle.dump(roles_by_project, f, protocol=-1)

            with open(self.pers_keystone + '/filters.pickle', 'wb') as f:
                pickle.dump(filters, f, protocol=-1)

            with open(self.pers_keystone + '/filters_by_project.pickle', 'wb') \
                    as f:
                pickle.dump(filters_by_project, f, protocol=-1)

        self.roles = self._convert(roles)
        self.users = self._convert(users)
        self.users_by_name = self._convert(users_by_name)
        self.tenants = self._convert(tenants)
        self.tenants_by_name = self._convert(tenants_by_name)
        self.roles_a = self._convert(roles_a)
        self.filters = self._convert(filters)
        self.filters_by_project = self._convert(filters_by_project)

    def _get_nova_data(self):
        """ get data from nova"""
        dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS
        save = dict_object and \
            self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS

        nova = self.osclients.get_novaclient()
        vms = nova.servers.list(search_opts={'all_tenants': 1})

        if dict_object:
            vms = dict((vm.id, vm.to_dict()) for vm in vms)
        else:
            self.vms = dict((vm.id, vm) for vm in vms)
            return

        if save:
            with open(self.pers_region + '/vms.pickle', 'wb') as f:
                pickle.dump(vms, f, protocol=-1)

        self.vms = self._convert(vms)

    def _get_cinder_data(self):
        """get data from cinder"""
        dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS
        save = dict_object and \
            self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS

        cinder = self.osclients.get_cinderclientv1()
        volumes = cinder.volumes.list(search_opts={'all_tenants': 1})
        snapshots = cinder.volume_snapshots.list(
            search_opts={'all_tenants': 1})
        backups = cinder.backups.list(search_opts={'all_tenants': 1})

        if dict_object:
            volumes = dict((volume.id, volume.__dict__) for volume in volumes)
            snapshots = dict((snapshot.id, snapshot.__dict__) for snapshot in
                             snapshots)
            backups = dict((backup.id, backup.__dict__) for backup in backups)
        else:
            self.volumes = dict((volume.id, volume) for volume in volumes)
            self.volume_snapshots = dict((snapshot.id, snapshot) for snapshot
                                         in snapshots)
            self.volume_backups = dict((backup.id, backup) for backup in
                                       backups)
            return

        if save:
            with open(self.pers_region + '/volumes.pickle', 'wb') as f:
                pickle.dump(volumes, f, protocol=-1)
            with open(self.pers_region + '/volume_snapshots.pickle', 'wb') as \
                    f:
                pickle.dump(snapshots, f, protocol=-1)
            with open(self.pers_region + '/volume_backups.pickle', 'wb') as f:
                pickle.dump(backups, f, protocol=-1)

        self.volumes = self._convert(volumes)
        self.volume_snapshots = self._convert(snapshots)
        self.volume_backups = self._convert(backups)

    def _get_glance_data(self):
        """get data from glance"""
        dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS
        save = dict_object and \
            self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS

        glance = self.osclients.get_glanceclient()
        images = glance.images.findall()

        if dict_object:
            images = dict((image.id, image.to_dict()) for image in
                          glance.images.findall())
        else:
            self.images = dict((image.id, image) for image in images)
            return

        if save:
            with open(self.pers_region + '/images.pickle', 'wb') as f:
                pickle.dump(images, f, protocol=-1)

        self.images = self._convert(images)

    def _get_neutron_data(self):
        """get network data from neutron"""
        dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS
        save = dict_object and \
            self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS

        neutron = self.osclients.get_neutronclient()
        networks = neutron.list_networks()['networks']
        subnets = neutron.list_subnets()['subnets']
        routers = neutron.list_routers()['routers']
        sec_grps = neutron.list_security_groups()['security_groups']
        floatingips = neutron.list_floatingips()['floatingips']
        ports = neutron.list_ports()['ports']

        nets = dict((network['id'], network) for network in networks)
        snets = dict((subnet['id'], subnet) for subnet in subnets)
        routers = dict((router['id'], router) for router in routers)
        floatingips = dict((floatingip['id'], floatingip) for floatingip in
                           floatingips)
        sec_grps = dict((sg['id'], sg) for sg in sec_grps)
        ports = dict((port['id'], port) for port in ports)

        if dict_object:
            # neutron objects are already dicts, so they do not need conversion
            pass
        else:
            self.networks = nets
            self.subnets = snets
            self.routers = routers
            self.floatingips = floatingips
            self.security_groups = sec_grps
            self.ports = ports
            return

        if save:
            with open(self.pers_region + '/networks.pickle', 'wb') as f:
                pickle.dump(nets, f, protocol=-1)

            with open(self.pers_region + '/subnets.pickle', 'wb') as f:
                pickle.dump(snets, f, protocol=-1)

            with open(self.pers_region + '/routers.pickle', 'wb') as f:
                pickle.dump(routers, f, protocol=-1)

            with open(self.pers_region + '/floatingips.pickle', 'wb') as f:
                pickle.dump(floatingips, f, protocol=-1)

            with open(self.pers_region + '/security_groups.pickle', 'wb') as f:
                pickle.dump(sec_grps, f, protocol=-1)

            with open(self.pers_region + '/ports.pickle', 'wb') as f:
                pickle.dump(ports, f, protocol=-1)

        self.networks = self._convert(nets)
        self.subnets = self._convert(snets)
        self.routers = self._convert(routers)
        self.floatingips = self._convert(floatingips)
        self.security_groups = self._convert(sec_grps)
        self.ports = self._convert(ports)

    def load_nova(self):
        """load nova data: vms"""
        if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or
            self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \
                os.path.exists(self.pers_region + '/vms.pickle'):
            self.vms = self._load('vms')
        else:
            if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY:
                raise Exception('Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about nova')

            self._get_nova_data()

    def load_glance(self):
        """load glance data: images"""
        if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or
            self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \
                os.path.exists(self.pers_region + '/images.pickle'):
            self.images = self._load('images')
        else:
            if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY:
                raise Exception('Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about glance')
            self._get_glance_data()

    def load_neutron(self):
        """load neutron (network) data: networks, subnets, routers,
           floatingips, security_groups, ports"""
        if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or
            self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \
                os.path.exists(self.pers_region + '/networks.pickle'):
            self.networks = self._load('networks')
            # legacy
            if os.path.exists(self.pers_region + '/subnetworks.pickle'):
                self.subnets = self._load('subnetworks')
            else:
                self.subnets = self._load('subnets')

            self.routers = self._load('routers')
            self.floatingips = self._load('floatingips')
            # legacy
            if os.path.exists(self.pers_region + '/securitygroups.pickle'):
                self.security_groups = self._load('securitygroups')
            else:
                self.security_groups = self._load('security_groups')

            self.ports = self._load('ports')
        else:
            if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY:
                raise Exception('Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about neutron')
            self._get_neutron_data()

        # make indexes
        self.floatingips_by_ip = dict((f['floating_ip_address'], f)
                                      for f in self.floatingips.values()
                                      if 'floating_ip_address' in f)

    def load_keystone(self):
        """load keystone data: users, tenants, roles, roles_a,
           users_by_name, tenants_by_name, roles_by_project, roles_by_user
        """
        if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or
            self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \
                os.path.exists(self.pers_keystone + '/users.pickle'):
            self.users = self._load_fkeystone('users')
            self.users_by_name = self._load_fkeystone('users_by_name')
            self.tenants = self._load_fkeystone('tenants')
            self.tenants_by_name = self._load_fkeystone('tenants_by_name')
            # legacy code
            if os.path.exists(self.pers_keystone + '/asignments.pickle'):
                self.roles_a = self._load_fkeystone('asignments')
            else:
                self.roles_a = self._load_fkeystone('roles_a')
            self.roles = self._load_fkeystone('roles')
            self.roles_by_project = self._load_fkeystone('roles_by_project')
            self.roles_by_user = self._load_fkeystone('roles_by_user')
            self.filters = self._load_fkeystone('filters')
            # legacy code
            if os.path.exists(self.pers_keystone + '/filters_byproject.pickle'):
                self.filters_by_project = self._load_fkeystone('filters_byproject')
            else:
                self.filters_by_project = self._load_fkeystone('filters_by_project')

        else:
            if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY:
                raise Exception('Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about keystone')

            self._get_keystone_data()

    def load_cinder(self):
        """load cinder data: volumes, volume_backups, volume_snapshots
        """
        if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or
            self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \
                os.path.exists(self.pers_region + '/volumes.pickle'):
            self.volumes = self._load('volumes')
            self.volume_backups = self._load('volume_backups')
            self.volume_snapshots = self._load('volume_snapshots')
        else:
            if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY:
                raise Exception('Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about cinder')

            self._get_cinder_data()

    def load_all(self):
        """load all data"""
        region = self.osclients.region
        if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY:
            if os.path.exists(self.pers_region + '/vms.pickle'):
                self.load_nova()

            if os.path.exists(self.pers_region + '/networks.pickle'):
                self.load_neutron()

            if os.path.exists(self.pers_region + '/images.pickle'):
                self.load_glance()

            if os.path.exists(self.pers_region + '/volumes.pickle'):
                self.load_cinder()
        else:
            if region in self.osclients.get_regions('compute'):
                self.load_nova()

            if region in self.osclients.get_regions('network'):
                self.load_neutron()

            if region in self.osclients.get_regions('image'):
                self.load_glance()

            if region in self.osclients.get_regions('volume'):
                self.load_cinder()

        self.load_keystone()

    def change_region(self, region, auto_load=True):
        """change region and clean maps. Optionally load the maps.
        :param region: the new region
        :param auto_load: True to invoke load_all
        :return: nothing
        """
        self.pers_region = self.persistence_dir + '/' + region
        if not os.path.exists(self.pers_region) and self.objects_strategy\
           not in (OpenStackMap.DIRECT_OBJECTS, OpenStackMap.NO_CACHE_OBJECTS):
            os.mkdir(self.pers_region)
        self.osclients.set_region(region)
        self._init_resource_maps()
        if auto_load:
            self.load_all()

    def preload_regions(self, regions=None, all_regions_excluded=None):
        """Method to preload the data of the specified regions. If
        regions is None, use all the available regions in the federation, but
        the specified in all_regions_excluded.

        The data for each region will be available at the region_map dictionary.

        It must be noted that this method could affect the current region
        and the values of the direct maps (vms, networks...). This is because it
        calls change_region/load_<service_name> methods. If regions is provided, then
        the last region in the list will be the new current region"""

        if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY:
            regions_in_disk = set(os.listdir(self.persistence_dir))
            if 'keystone' in regions_in_disk:
                regions_in_disk.remove('keystone')

            regions_compute = set()
            regions_network = set()
            regions_image = set()
            regions_volume = set()

            for region in regions_in_disk:
                path = self.persistence_dir + os.path.sep + region
                if os.path.exists(path + os.path.sep + 'vms.pickle'):
                    regions_compute.add(region)

                if os.path.exists(path + os.path.sep + 'networks.pickle'):
                    regions_network.add(region)

                if os.path.exists(path + os.path.sep + 'images.pickle'):
                    regions_image.add(region)

                if os.path.exists(path + os.path.sep + 'volumes.pickle'):
                    regions_volume.add(region)

        else:
            regions_compute = self.osclients.get_regions('compute')
            regions_network = self.osclients.get_regions('network')
            regions_image = self.osclients.get_regions('image')
            regions_volume = self.osclients.get_regions('volume')

        if not regions:
            """ not regions specified, so all existing regions with some
                resource (compute, network, image or volume) are considered.
                If all_regions_excluded is defined, then these regions are excluded
            """
            all_regions = regions_compute.union(regions_network).union(
                regions_image).union(regions_volume)
            if all_regions_excluded:
                all_regions.difference_update(all_regions_excluded)
            # Move the current region to the end of the list.
            # This is because the direct map dictionaries (vms, networks, etc.)
            # and the region field are updated with the change_region() call.

            if self.osclients.region in all_regions:
                all_regions.remove(self.osclients.region)
            regions = list(all_regions)
            regions.append(self.osclients.region)

        for region in regions:
            try:
                self.logger.info('Creating map of region ' + region)
                self.change_region(region, False)
                if region in regions_compute:
                    self.load_nova()
                if region in regions_network:
                    self.load_neutron()
                if region in regions_image:
                    self.load_glance()
                if region in regions_volume:
                    self.load_cinder()
                region_map = dict()
                for resource in self.resources_region:
                    region_map[resource] = getattr(self, resource)
                self.region_map[region] = region_map

            except Exception, e:
                msg = 'Failed the creation of the map of {0}. Cause: {1}'
                self.logger.error(msg.format(region, str(e)))
                # Remove dir if it exists and is empty.
                dir = os.path.join(self.persistence_dir, region)
                if os.path.isdir(dir) and len(os.listdir(dir)) == 0:
                    os.rmdir(dir)

        self.load_keystone()
Example #3
0
class OpenStackMap(object):
    """
    This class build a map from the resources (VMs, networks, images,
    volumes, users, tenants, roles...) in an OpenStack infrastructure.

    This map is optionally cached on disk and can be accessed offline or can be
    used as a metadata snapshot of an OpenStack infrastructure (e.g. to
    analyse resource use and build statistics)
    """

    # objects strategy see the __init__ method documentation
    DIRECT_OBJECTS, NO_CACHE_OBJECTS, REFRESH_OBJECTS, USE_CACHE_OBJECTS, USE_CACHE_OBJECTS_ONLY = \
        range(5)

    # If use_wrapper is True, dictionaries are wrapped to allow access to
    # resource['field'] also as resource.field. This is not used when
    # objects_strategy is DIRECT_OBJECTS
    use_wrapper = True

    load_filters = True

    resources_region = [
        'vms', 'images', 'routers', 'networks', 'subnets', 'ports',
        'floatingips', 'security_groups', 'volumes', 'volume_backups',
        'volume_snapshots'
    ]

    def __init__(self,
                 persistence_dir='~/openstackmap',
                 region=None,
                 auth_url=None,
                 objects_strategy=USE_CACHE_OBJECTS,
                 auto_load=True):
        """
        Constructor
        :param persistence_dir: The path where the data is saved. Ignored if
           objects_strategy is DIRECT_OBJECTS or NO_CACHE_OBJECTS
        :param region: the initial region (if undefined, use OS_REGION_NAME)
        :param auth_url: the keystone URI
        :param objects_strategy: sets the strategy about the object maps
            contained here. It can be:
         * DIRECT_OBJECTS is using the objects as they are got from
           the API. Be careful because methods as delete are available! The
           objects are not cached.
         * NO_CACHE_OBJECTS is using the objects converted to dictionaries, so
           methods to do operations are not available. The objects are not
           cached.
         * REFRESH_OBJECTS is using the objects converted to dictionaries. The
           objects are cached: the new version replace the old one.
         * USE_CACHE_OBJECTS is using the objects converted to dictionaries. If
           a cached copy of the objects are available, it is used.
         * USE_CACHE_OBJECTS_ONLY is using the objects converted to dictionaries.
           This strategy used cached objects only. It never contacts with the
           servers, even when the object is not available in the local cache.

        :param auto_load: if True, invoke self.load_all()
         Note that neutron objects returned by the API are already dictionaries
        """

        self.logger = logging.getLogger(__name__)

        if auth_url:
            self.osclients = OpenStackClients(auth_url=auth_url)
        else:
            self.osclients = OpenStackClients()

        if region:
            self.osclients.set_region(region)
        else:
            if 'OS_REGION_NAME' not in env:
                raise Exception('Region parameter must be provided or '
                                'OS_REGION_NAME variable must be defined.')
            else:
                region = env['OS_REGION_NAME']

        if 'KEYSTONE_ADMIN_ENDPOINT' in os.environ:
            self.osclients.override_endpoint(
                'identity', self.osclients.region, 'admin',
                os.environ['KEYSTONE_ADMIN_ENDPOINT'])

        self.objects_strategy = objects_strategy

        self.persistence_dir = os.path.expanduser(persistence_dir)
        self.pers_region = self.persistence_dir + '/' + region
        self.pers_keystone = self.persistence_dir + '/keystone'

        if objects_strategy not in (OpenStackMap.DIRECT_OBJECTS,
                                    OpenStackMap.NO_CACHE_OBJECTS):
            if not os.path.exists(self.persistence_dir):
                os.mkdir(self.persistence_dir)

            if not os.path.exists(self.pers_keystone):
                os.mkdir(self.pers_keystone)

            if not os.path.exists(self.pers_region):
                os.mkdir(self.pers_region)

        self._init_resource_maps()

        if auto_load:
            self.load_all()

        self.region_map = dict()

    def _init_resource_maps(self):
        """init all the resources that will be available
        as empty dictionaries"""
        # Keystone resources
        self.users = dict()
        self.users_by_name = dict()
        self.roles = dict()
        self.tenants = dict()
        self.tenants_by_name = dict()
        self.roles_a = list()
        self.roles_by_project = dict()
        self.roles_by_user = dict()
        self.filters = dict()
        self.filters_by_project = dict()

        # Glance resources
        self.images = dict()
        # Neutron resources
        self.networks = dict()
        self.subnets = dict()
        self.routers = dict()
        self.floatingips = dict()
        self.floatingips_by_ip = dict()
        self.ports = dict()
        self.security_groups = dict()
        # Nova resources
        self.vms = dict()
        self.flavors = dict()
        # Cinder resources
        self.volumes = dict()
        self.volume_backups = dict()
        self.volume_snapshots = dict()

    def _load(self, name):
        """Load the resources persisted with pickle. This resources (e.g.
        networks, vms, images...) are saved independently for each region.

        This method is called for load_cinder, load_nova, load_glance,
        load_neutron...

        :param name: the resource name
        :return: a dictionary of objects (dictionaries) indexed by id
        """
        objects = pickle.load(
            open(self.pers_region + '/' + name + '.pickle', 'rb'))
        return self._convert(objects)

    def _load_fkeystone(self, name):
        """Load the keystone objects persisted with pickle. This resources are
        shared among the regions. This method is used in load_keystone

        :param name: the resource name
        :return: a list/dictionary of objects (dictionaries)
        """
        objects = pickle.load(
            open(self.pers_keystone + '/' + name + '.pickle', 'rb'))
        return self._convert(objects)

    def _convert(self, objects):
        """if use_wrapper, convert objects from dictionary to __E class;
        this class allows accessing  a['key'] also as a.key"""
        if self.use_wrapper and len(objects) > 0:

            class __E(dict):
                def __init__(self, d=dict()):
                    self.__dict__ = self
                    dict.__init__(self, d)

            if isinstance(objects, dict):
                if not isinstance(objects.values()[0], dict):
                    return objects
                return dict((key, __E(objects[key])) for key in objects)
            else:
                return list(__E(object) for object in objects)

        else:
            return objects

    def _get_keystone_data(self):
        """get data from keystone server"""
        dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS
        save = dict_object and \
            self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS
        keystone = self.osclients.get_keystoneclientv3()
        roles = keystone.roles.list()
        users = keystone.users.list()
        tenants = keystone.projects.list()
        roles_a = keystone.role_assignments.list()

        if OpenStackMap.load_filters:
            ef = {'service_type': 'identity', 'interface': 'public'}
            resp = keystone.session.get('/OS-EP-FILTER/endpoint_groups',
                                        endpoint_filter=ef)
            filters = resp.json()['endpoint_groups']
            projects_by_filter = dict()
            filters_by_project = dict()
            for f in filters:
                filter_id = f['id']
                resp = keystone.session.get('/OS-EP-FILTER/endpoint_groups/' +
                                            filter_id + '/projects',
                                            endpoint_filter=ef)
                projects = resp.json()['projects']
                projects_by_filter[filter_id] = projects
                for project in projects:
                    if project['id'] not in filters_by_project:
                        filters_by_project[project['id']] = list()
                    filters_by_project[project['id']].append(filter_id)

        else:
            filters = list()
            filters_by_project = dict()

        roles_by_user = dict()
        roles_by_project = dict()

        for roleasig in roles_a:
            try:
                userid = roleasig.user['id']
                if 'project' in roleasig.scope:
                    projectid = roleasig.scope['project']['id']
                else:
                    projectid = str(roleasig.scope)
                roleid = roleasig.role['id']
                if userid not in roles_by_user:
                    roles_by_user[userid] = list()
            except AttributeError:
                # Discard roles not assigned to users
                continue

            if projectid not in roles_by_project:
                roles_by_project[projectid] = list()
            roles_by_user[userid].append((roleid, projectid))
            roles_by_project[projectid].append((roleid, userid))

        self.roles_by_user = roles_by_user
        self.roles_by_project = roles_by_project

        tenants_by_name = dict()
        users_by_name = dict()

        filters = dict((f['id'], f) for f in filters)
        if dict_object:
            roles = dict((role.id, role.to_dict()) for role in roles)
            users = dict((user.id, user.to_dict()) for user in users)
            tenants = dict((tenant.id, tenant.to_dict()) for tenant in tenants)
            roles_a = list(asig.to_dict() for asig in roles_a)
            for tenant in tenants.values():
                tenants_by_name[tenant['name']] = tenant
            for user in users.values():
                users_by_name[user['name']] = user
        else:
            self.roles = dict((role.id, role) for role in roles)
            self.users = dict((user.id, user) for user in users)
            self.tenants = dict((tenant.id, tenant) for tenant in tenants)
            self.roles_a = roles_a

            tenants_by_name = dict()
            for tenant in tenants:
                if getattr(tenant, 'name', None):
                    tenants_by_name[tenant.name] = tenant

            for user in users:
                if getattr(user, 'name', None):
                    users_by_name[user.name] = user

            self.tenants_by_name = tenants_by_name
            self.users_by_name = users_by_name
            self.filters = filters
            self.filters_by_project = filters_by_project
            return

        if save:
            with open(self.pers_keystone + '/roles.pickle', 'wb') as f:
                pickle.dump(roles, f, protocol=-1)

            with open(self.pers_keystone + '/users.pickle', 'wb') as f:
                pickle.dump(users, f, protocol=-1)

            with open(self.pers_keystone + '/users_by_name.pickle', 'wb') as f:
                pickle.dump(users_by_name, f, protocol=-1)

            with open(self.pers_keystone + '/tenants.pickle', 'wb') as f:
                pickle.dump(tenants, f, protocol=-1)

            with open(self.pers_keystone + '/tenants_by_name.pickle', 'wb') as\
                    f:
                pickle.dump(tenants_by_name, f)

            with open(self.pers_keystone + '/roles_a.pickle', 'wb') as f:
                pickle.dump(roles_a, f, protocol=-1)

            with open(self.pers_keystone + '/roles_by_user.pickle', 'wb') as f:
                pickle.dump(roles_by_user, f, protocol=-1)

            with open(self.pers_keystone + '/roles_by_project.pickle', 'wb') as\
                    f:
                pickle.dump(roles_by_project, f, protocol=-1)

            with open(self.pers_keystone + '/filters.pickle', 'wb') as f:
                pickle.dump(filters, f, protocol=-1)

            with open(self.pers_keystone + '/filters_by_project.pickle', 'wb') \
                    as f:
                pickle.dump(filters_by_project, f, protocol=-1)

        self.roles = self._convert(roles)
        self.users = self._convert(users)
        self.users_by_name = self._convert(users_by_name)
        self.tenants = self._convert(tenants)
        self.tenants_by_name = self._convert(tenants_by_name)
        self.roles_a = self._convert(roles_a)
        self.filters = self._convert(filters)
        self.filters_by_project = self._convert(filters_by_project)

    def _get_nova_data(self):
        """ get data from nova"""
        dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS
        save = dict_object and \
            self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS

        nova = self.osclients.get_novaclient()
        vms = nova.servers.list(search_opts={'all_tenants': 1})

        if dict_object:
            vms = dict((vm.id, vm.to_dict()) for vm in vms)
        else:
            self.vms = dict((vm.id, vm) for vm in vms)
            return

        if save:
            with open(self.pers_region + '/vms.pickle', 'wb') as f:
                pickle.dump(vms, f, protocol=-1)

        self.vms = self._convert(vms)

    def _get_cinder_data(self):
        """get data from cinder"""
        dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS
        save = dict_object and \
            self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS

        cinder = self.osclients.get_cinderclientv1()
        volumes = cinder.volumes.list(search_opts={'all_tenants': 1})
        snapshots = cinder.volume_snapshots.list(
            search_opts={'all_tenants': 1})
        backups = cinder.backups.list(search_opts={'all_tenants': 1})

        if dict_object:
            volumes = dict((volume.id, volume.__dict__) for volume in volumes)
            snapshots = dict(
                (snapshot.id, snapshot.__dict__) for snapshot in snapshots)
            backups = dict((backup.id, backup.__dict__) for backup in backups)
        else:
            self.volumes = dict((volume.id, volume) for volume in volumes)
            self.volume_snapshots = dict(
                (snapshot.id, snapshot) for snapshot in snapshots)
            self.volume_backups = dict(
                (backup.id, backup) for backup in backups)
            return

        if save:
            with open(self.pers_region + '/volumes.pickle', 'wb') as f:
                pickle.dump(volumes, f, protocol=-1)
            with open(self.pers_region + '/volume_snapshots.pickle', 'wb') as \
                    f:
                pickle.dump(snapshots, f, protocol=-1)
            with open(self.pers_region + '/volume_backups.pickle', 'wb') as f:
                pickle.dump(backups, f, protocol=-1)

        self.volumes = self._convert(volumes)
        self.volume_snapshots = self._convert(snapshots)
        self.volume_backups = self._convert(backups)

    def _get_glance_data(self):
        """get data from glance"""
        dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS
        save = dict_object and \
            self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS

        glance = self.osclients.get_glanceclient()
        images = glance.images.findall()

        if dict_object:
            images = dict((image.id, image.to_dict())
                          for image in glance.images.findall())
        else:
            self.images = dict((image.id, image) for image in images)
            return

        if save:
            with open(self.pers_region + '/images.pickle', 'wb') as f:
                pickle.dump(images, f, protocol=-1)

        self.images = self._convert(images)

    def _get_neutron_data(self):
        """get network data from neutron"""
        dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS
        save = dict_object and \
            self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS

        neutron = self.osclients.get_neutronclient()
        networks = neutron.list_networks()['networks']
        subnets = neutron.list_subnets()['subnets']
        routers = neutron.list_routers()['routers']
        sec_grps = neutron.list_security_groups()['security_groups']
        floatingips = neutron.list_floatingips()['floatingips']
        ports = neutron.list_ports()['ports']

        nets = dict((network['id'], network) for network in networks)
        snets = dict((subnet['id'], subnet) for subnet in subnets)
        routers = dict((router['id'], router) for router in routers)
        floatingips = dict(
            (floatingip['id'], floatingip) for floatingip in floatingips)
        sec_grps = dict((sg['id'], sg) for sg in sec_grps)
        ports = dict((port['id'], port) for port in ports)

        if dict_object:
            # neutron objects are already dicts, so they do not need conversion
            pass
        else:
            self.networks = nets
            self.subnets = snets
            self.routers = routers
            self.floatingips = floatingips
            self.security_groups = sec_grps
            self.ports = ports
            return

        if save:
            with open(self.pers_region + '/networks.pickle', 'wb') as f:
                pickle.dump(nets, f, protocol=-1)

            with open(self.pers_region + '/subnets.pickle', 'wb') as f:
                pickle.dump(snets, f, protocol=-1)

            with open(self.pers_region + '/routers.pickle', 'wb') as f:
                pickle.dump(routers, f, protocol=-1)

            with open(self.pers_region + '/floatingips.pickle', 'wb') as f:
                pickle.dump(floatingips, f, protocol=-1)

            with open(self.pers_region + '/security_groups.pickle', 'wb') as f:
                pickle.dump(sec_grps, f, protocol=-1)

            with open(self.pers_region + '/ports.pickle', 'wb') as f:
                pickle.dump(ports, f, protocol=-1)

        self.networks = self._convert(nets)
        self.subnets = self._convert(snets)
        self.routers = self._convert(routers)
        self.floatingips = self._convert(floatingips)
        self.security_groups = self._convert(sec_grps)
        self.ports = self._convert(ports)

    def load_nova(self):
        """load nova data: vms"""
        if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or
            self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \
                os.path.exists(self.pers_region + '/vms.pickle'):
            self.vms = self._load('vms')
        else:
            if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY:
                raise Exception(
                    'Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about nova'
                )

            self._get_nova_data()

    def load_glance(self):
        """load glance data: images"""
        if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or
            self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \
                os.path.exists(self.pers_region + '/images.pickle'):
            self.images = self._load('images')
        else:
            if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY:
                raise Exception(
                    'Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about glance'
                )
            self._get_glance_data()

    def load_neutron(self):
        """load neutron (network) data: networks, subnets, routers,
           floatingips, security_groups, ports"""
        if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or
            self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \
                os.path.exists(self.pers_region + '/networks.pickle'):
            self.networks = self._load('networks')
            # legacy
            if os.path.exists(self.pers_region + '/subnetworks.pickle'):
                self.subnets = self._load('subnetworks')
            else:
                self.subnets = self._load('subnets')

            self.routers = self._load('routers')
            self.floatingips = self._load('floatingips')
            # legacy
            if os.path.exists(self.pers_region + '/securitygroups.pickle'):
                self.security_groups = self._load('securitygroups')
            else:
                self.security_groups = self._load('security_groups')

            self.ports = self._load('ports')
        else:
            if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY:
                raise Exception(
                    'Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about neutron'
                )
            self._get_neutron_data()

        # make indexes
        self.floatingips_by_ip = dict((f['floating_ip_address'], f)
                                      for f in self.floatingips.values()
                                      if 'floating_ip_address' in f)

    def load_keystone(self):
        """load keystone data: users, tenants, roles, roles_a,
           users_by_name, tenants_by_name, roles_by_project, roles_by_user
        """
        if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or
            self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \
                os.path.exists(self.pers_keystone + '/users.pickle'):
            self.users = self._load_fkeystone('users')
            self.users_by_name = self._load_fkeystone('users_by_name')
            self.tenants = self._load_fkeystone('tenants')
            self.tenants_by_name = self._load_fkeystone('tenants_by_name')
            # legacy code
            if os.path.exists(self.pers_keystone + '/asignments.pickle'):
                self.roles_a = self._load_fkeystone('asignments')
            else:
                self.roles_a = self._load_fkeystone('roles_a')
            self.roles = self._load_fkeystone('roles')
            self.roles_by_project = self._load_fkeystone('roles_by_project')
            self.roles_by_user = self._load_fkeystone('roles_by_user')
            self.filters = self._load_fkeystone('filters')
            # legacy code
            if os.path.exists(self.pers_keystone +
                              '/filters_byproject.pickle'):
                self.filters_by_project = self._load_fkeystone(
                    'filters_byproject')
            else:
                self.filters_by_project = self._load_fkeystone(
                    'filters_by_project')

        else:
            if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY:
                raise Exception(
                    'Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about keystone'
                )

            self._get_keystone_data()

    def load_cinder(self):
        """load cinder data: volumes, volume_backups, volume_snapshots
        """
        if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or
            self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \
                os.path.exists(self.pers_region + '/volumes.pickle'):
            self.volumes = self._load('volumes')
            self.volume_backups = self._load('volume_backups')
            self.volume_snapshots = self._load('volume_snapshots')
        else:
            if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY:
                raise Exception(
                    'Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about cinder'
                )

            self._get_cinder_data()

    def load_all(self):
        """load all data"""
        region = self.osclients.region
        if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY:
            if os.path.exists(self.pers_region + '/vms.pickle'):
                self.load_nova()

            if os.path.exists(self.pers_region + '/networks.pickle'):
                self.load_neutron()

            if os.path.exists(self.pers_region + '/images.pickle'):
                self.load_glance()

            if os.path.exists(self.pers_region + '/volumes.pickle'):
                self.load_cinder()
        else:
            if region in self.osclients.get_regions('compute'):
                self.load_nova()

            if region in self.osclients.get_regions('network'):
                self.load_neutron()

            if region in self.osclients.get_regions('image'):
                self.load_glance()

            if region in self.osclients.get_regions('volume'):
                self.load_cinder()

        self.load_keystone()

    def change_region(self, region, auto_load=True):
        """change region and clean maps. Optionally load the maps.
        :param region: the new region
        :param auto_load: True to invoke load_all
        :return: nothing
        """
        self.pers_region = self.persistence_dir + '/' + region
        if not os.path.exists(self.pers_region) and self.objects_strategy\
           not in (OpenStackMap.DIRECT_OBJECTS, OpenStackMap.NO_CACHE_OBJECTS):
            os.mkdir(self.pers_region)
        self.osclients.set_region(region)
        self._init_resource_maps()
        if auto_load:
            self.load_all()

    def preload_regions(self, regions=None, all_regions_excluded=None):
        """Method to preload the data of the specified regions. If
        regions is None, use all the available regions in the federation, but
        the specified in all_regions_excluded.

        The data for each region will be available at the region_map dictionary.

        It must be noted that this method could affect the current region
        and the values of the direct maps (vms, networks...). This is because it
        calls change_region/load_<service_name> methods. If regions is provided, then
        the last region in the list will be the new current region"""

        if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY:
            regions_in_disk = set(os.listdir(self.persistence_dir))
            if 'keystone' in regions_in_disk:
                regions_in_disk.remove('keystone')

            regions_compute = set()
            regions_network = set()
            regions_image = set()
            regions_volume = set()

            for region in regions_in_disk:
                path = self.persistence_dir + os.path.sep + region
                if os.path.exists(path + os.path.sep + 'vms.pickle'):
                    regions_compute.add(region)

                if os.path.exists(path + os.path.sep + 'networks.pickle'):
                    regions_network.add(region)

                if os.path.exists(path + os.path.sep + 'images.pickle'):
                    regions_image.add(region)

                if os.path.exists(path + os.path.sep + 'volumes.pickle'):
                    regions_volume.add(region)

        else:
            regions_compute = self.osclients.get_regions('compute')
            regions_network = self.osclients.get_regions('network')
            regions_image = self.osclients.get_regions('image')
            regions_volume = self.osclients.get_regions('volume')

        if not regions:
            """ not regions specified, so all existing regions with some
                resource (compute, network, image or volume) are considered.
                If all_regions_excluded is defined, then these regions are excluded
            """
            all_regions = regions_compute.union(regions_network).union(
                regions_image).union(regions_volume)
            if all_regions_excluded:
                all_regions.difference_update(all_regions_excluded)
            # Move the current region to the end of the list.
            # This is because the direct map dictionaries (vms, networks, etc.)
            # and the region field are updated with the change_region() call.

            if self.osclients.region in all_regions:
                all_regions.remove(self.osclients.region)
            regions = list(all_regions)
            regions.append(self.osclients.region)

        for region in regions:
            try:
                self.logger.info('Creating map of region ' + region)
                self.change_region(region, False)
                if region in regions_compute:
                    self.load_nova()
                if region in regions_network:
                    self.load_neutron()
                if region in regions_image:
                    self.load_glance()
                if region in regions_volume:
                    self.load_cinder()
                region_map = dict()
                for resource in self.resources_region:
                    region_map[resource] = getattr(self, resource)
                self.region_map[region] = region_map

            except Exception, e:
                msg = 'Failed the creation of the map of {0}. Cause: {1}'
                self.logger.error(msg.format(region, str(e)))
                # Remove dir if it exists and is empty.
                dir = os.path.join(self.persistence_dir, region)
                if os.path.isdir(dir) and len(os.listdir(dir)) == 0:
                    os.rmdir(dir)

        self.load_keystone()
Example #4
0
class UserResources(object):
    """Class to list, delete user resources. Also provides a method to stop all
    the VM of the tenant.

    This class works creating a instance with the credential of the user owner
    of the resources. It does not use an admin credential!!!"""
    def __init__(self, username, password, tenant_id=None, tenant_name=None,
                 trust_id=None):
        """
        Constructor of the class. Tenant_id or tenant_name or tust_id must be
        provided.

        There are two ways of using this class:
        -passing the user, password, tenant_id or tenant_name of the user whose
        resources are being deleted
        -passing the user and password of the trustee, and the trust_id
        generated to impersonate the trustor.

        :param username: the user name whose resources are deleted
          or the user name of the trustee
        :param password: the password of the user whose resources are deleted
          or the password of the trustee
        :param tenant_id: the tenant id of the user whose resources must be
        deleted.
        :param tenant_name: the tenant name of the user whose resources must be
         deleted
        :param trust_id: the trust_id used to impersonate the user whose
        resources must be deleted
        :return: nothing
        """

        self.logger = logging.getLogger(__name__)
        self.clients = OpenStackClients()

        if tenant_id:
            self.clients.set_credential(username, password,
                                        tenant_id=tenant_id)
        elif tenant_name:
            self.clients.set_credential(username, password,
                                        tenant_name=tenant_name)
        elif trust_id:
            self.clients.set_credential(username, password, trust_id=trust_id)
            self.trust_id = trust_id
        else:
            raise(
                'Either tenant_id or tenant_name or trust_id must be provided')

        region = self.clients.region
        self.clients.override_endpoint(
            'identity', region, 'admin', settings.KEYSTONE_ENDPOINT)

        self.user_id = self.clients.get_session().get_user_id()
        session = self.clients.get_session()
        self.user_name = username
        self.nova = NovaResources(self.clients)
        self.cinder = CinderResources(self.clients)
        self.glance = GlanceResources(self.clients)

        try:
            self.neutron = NeutronResources(self.clients)
        except Exception:
            # The region does not support Neutron
            # It would be better to check the endpoint
            self.neutron = None

        try:
            self.blueprints = BluePrintResources(self.clients)
        except Exception:
            # The region does not support PaaS Manager
            # It would be better to check the endpoint
            self.blueprints = None

        try:
            self.swift = SwiftResources(self.clients)
        except Exception:
            # The region does not support Swift
            # It would be better to check the endpoint
            self.swift = None

        # Images in use is a set used to avoid deleting formerly glance images
        # in use by other tenants
        self.imagesinuse = set()

        # Regions the user has access
        self.regions_available = set()
        self.regions_available.update(self.clients.get_regions('compute'))

    def change_region(self, region):
        """
        change the region. All the clients need to be updated, but the
        session does not.
        :param region: the name of the region
        :return: nothing.
        """
        self.clients.set_region(region)
        self.clients.override_endpoint(
            'identity', region, 'admin', settings.KEYSTONE_ENDPOINT)

        self.nova.on_region_changed()
        self.glance.on_region_changed()

        try:
            if self.swift:
                self.swift.on_region_changed()
            else:
                self.swift = SwiftResources(self.clients)
        except Exception:
            # The region does not support swift
            self.swift = None

        self.cinder.on_region_changed()

        try:
            if self.blueprints:
                self.blueprint.on_region_changed()
            else:
                self.blueprints = BluePrintResources(self.clients)
        except Exception:
            # The region has not configured paas manager
            self.blueprints = None

        try:
            if self.neutron:
                self.neutron.on_region_changed()
            else:
                self.neutron = NeutronResources(self.clients)
        except Exception:
            # The region does not support neutron
            self.neutron = None

    def delete_tenant_resources_pri_1(self):
        """Delete here all the elements that do not depend of others are
        deleted first"""

        try:
            self.nova.delete_user_keypairs()
        except Exception, e:
            msg = 'Deletion of keypairs failed. Reason: '
            self.logger.error(msg + str(e))

        # Snapshots must be deleted before the volumes, because a snapshot
        # depends of a volume.
        try:
            self.cinder.delete_tenant_volume_snapshots()
        except Exception, e:
            msg = 'Deletion of volume snaphosts failed. Reason: '
            self.logger.error(msg + str(e))
class TestOSClientsOverrideEndpoint(TestCase):
    """Class to test the endpoint override feature"""

    def setUp(self):
        d = defaultdict(list)
        d['catalog'].append(service)
        self.access = d
        self.osclients = OpenStackClients()
        self.url = 'http://fake.org:9090'
        self.original_url = service['endpoints'][1]['url']

    def restore_catalog(self):
        """restore catalog"""
        service['endpoints'][1]['url'] = self.original_url

    def tearDown(self):
        """restore objects"""
        self.restore_catalog()

    def override_endpoint(self):
        """method that override the endpoint"""
        self.osclients.override_endpoint('object-store', 'Spain2', 'admin', self.url)

    def assertOverrideEndpoint(self):
        """check that the override has been done"""
        self.assertEquals(self.osclients.get_admin_endpoint('object-store', 'Spain2'), self.url)

    def test_override_endpoint_session(self):
        """test that invoking override endpoint does not create a session"""
        self.override_endpoint()

        self.assertFalse(self.osclients._session_v2)
        self.assertFalse(self.osclients._session_v3)

    def test_override_endpoint(self):
        """check that a session catalog is overriden"""
        mock = MagicMock()
        config = {'auth.get_access.return_value': self.access}
        mock.configure_mock(**config)
        self.osclients._session_v3 = mock
        self.override_endpoint()
        self.assertOverrideEndpoint()

    @patch('utils.osclients.session')
    def test_override_endpoint_multiple(self, mock):
        """test that override works with an already created session and then
        with a new one without invoking the method again"""
        config = {'Session.return_value.auth.get_access.return_value': self.access}
        mock.configure_mock(**config)
        session = self.osclients.get_session()
        self.override_endpoint()
        self.assertOverrideEndpoint()

        # invalidate and create a new session; ensure than catalog is again
        # the original. Setting a new token invalidate the session. The new
        # one is created at the invocation of get_admin_endpoint.
        self.restore_catalog()
        self.osclients.set_token('faketoken')

        # check again
        self.assertOverrideEndpoint()
            {'name': usertocheck.name})
        return True
    else:
        return False

logger.debug('Getting expired users')
(next_to_expire, expired_users) = ExpiredUsers(
    username=env['OS_USERNAME'], password=env['OS_PASSWORD'],
    tenant=env['OS_TENANT_NAME']).get_yellow_red_users()

osclients = OpenStackClients()

# Use an alternative URL that allow direct access to the keystone admin
# endpoint, because the registered one uses an internal IP address.

osclients.override_endpoint(
    'identity', osclients.region, 'admin', settings.KEYSTONE_ENDPOINT)

keystone = osclients.get_keystoneclientv3()


# build users map
logger.debug('Building user map')
users_by_id = dict()
for user in keystone.users.list():
    users_by_id[user.id] = user

with open('users_to_delete.txt', 'w') as fich_delete:
    logger.debug('Generating user delete list')
    for user_id in expired_users:
        if not is_user_protected(users_by_id[user_id]):
            fich_delete.write(user_id + "\n")