def test_set_region(self): """test_set_region check that we could change the region after create the client""" # FIRST CHECK: Region is recovered from ENV osclients = OpenStackClients() self.assertEqual(osclients.region, self.OS_REGION_NAME) # Check that region is updated to a new Value. region = "Budapest" osclients.set_region(region) self.assertEqual(osclients.region, region)
class OpenStackMap(object): """ This class build a map from the resources (VMs, networks, images, volumes, users, tenants, roles...) in an OpenStack infrastructure. This map is optionally cached on disk and can be accessed offline or can be used as a metadata snapshot of an OpenStack infrastructure (e.g. to analyse resource use and build statistics) """ # objects strategy see the __init__ method documentation DIRECT_OBJECTS, NO_CACHE_OBJECTS, REFRESH_OBJECTS, USE_CACHE_OBJECTS, USE_CACHE_OBJECTS_ONLY = \ range(5) # If use_wrapper is True, dictionaries are wrapped to allow access to # resource['field'] also as resource.field. This is not used when # objects_strategy is DIRECT_OBJECTS use_wrapper = True load_filters = True resources_region = ['vms', 'images', 'routers', 'networks', 'subnets', 'ports', 'floatingips', 'security_groups', 'volumes', 'volume_backups', 'volume_snapshots'] def __init__( self, persistence_dir='~/openstackmap', region=None, auth_url=None, objects_strategy=USE_CACHE_OBJECTS, auto_load=True): """ Constructor :param persistence_dir: The path where the data is saved. Ignored if objects_strategy is DIRECT_OBJECTS or NO_CACHE_OBJECTS :param region: the initial region (if undefined, use OS_REGION_NAME) :param auth_url: the keystone URI :param objects_strategy: sets the strategy about the object maps contained here. It can be: * DIRECT_OBJECTS is using the objects as they are got from the API. Be careful because methods as delete are available! The objects are not cached. * NO_CACHE_OBJECTS is using the objects converted to dictionaries, so methods to do operations are not available. The objects are not cached. * REFRESH_OBJECTS is using the objects converted to dictionaries. The objects are cached: the new version replace the old one. * USE_CACHE_OBJECTS is using the objects converted to dictionaries. If a cached copy of the objects are available, it is used. * USE_CACHE_OBJECTS_ONLY is using the objects converted to dictionaries. This strategy used cached objects only. It never contacts with the servers, even when the object is not available in the local cache. :param auto_load: if True, invoke self.load_all() Note that neutron objects returned by the API are already dictionaries """ self.logger = logging.getLogger(__name__) if auth_url: self.osclients = OpenStackClients(auth_url=auth_url) else: self.osclients = OpenStackClients() if region: self.osclients.set_region(region) else: if 'OS_REGION_NAME' not in env: raise Exception('Region parameter must be provided or ' 'OS_REGION_NAME variable must be defined.') else: region = env['OS_REGION_NAME'] if 'KEYSTONE_ADMIN_ENDPOINT' in os.environ: self.osclients.override_endpoint( 'identity', self.osclients.region, 'admin', os.environ['KEYSTONE_ADMIN_ENDPOINT']) self.objects_strategy = objects_strategy self.persistence_dir = os.path.expanduser(persistence_dir) self.pers_region = self.persistence_dir + '/' + region self.pers_keystone = self.persistence_dir + '/keystone' if objects_strategy not in (OpenStackMap.DIRECT_OBJECTS, OpenStackMap.NO_CACHE_OBJECTS): if not os.path.exists(self.persistence_dir): os.mkdir(self.persistence_dir) if not os.path.exists(self.pers_keystone): os.mkdir(self.pers_keystone) if not os.path.exists(self.pers_region): os.mkdir(self.pers_region) self._init_resource_maps() if auto_load: self.load_all() self.region_map = dict() def _init_resource_maps(self): """init all the resources that will be available as empty dictionaries""" # Keystone resources self.users = dict() self.users_by_name = dict() self.roles = dict() self.tenants = dict() self.tenants_by_name = dict() self.roles_a = list() self.roles_by_project = dict() self.roles_by_user = dict() self.filters = dict() self.filters_by_project = dict() # Glance resources self.images = dict() # Neutron resources self.networks = dict() self.subnets = dict() self.routers = dict() self.floatingips = dict() self.floatingips_by_ip = dict() self.ports = dict() self.security_groups = dict() # Nova resources self.vms = dict() self.flavors = dict() # Cinder resources self.volumes = dict() self.volume_backups = dict() self.volume_snapshots = dict() def _load(self, name): """Load the resources persisted with pickle. This resources (e.g. networks, vms, images...) are saved independently for each region. This method is called for load_cinder, load_nova, load_glance, load_neutron... :param name: the resource name :return: a dictionary of objects (dictionaries) indexed by id """ objects = pickle.load(open(self.pers_region + '/' + name + '.pickle', 'rb')) return self._convert(objects) def _load_fkeystone(self, name): """Load the keystone objects persisted with pickle. This resources are shared among the regions. This method is used in load_keystone :param name: the resource name :return: a list/dictionary of objects (dictionaries) """ objects = pickle.load(open(self.pers_keystone + '/' + name + '.pickle', 'rb')) return self._convert(objects) def _convert(self, objects): """if use_wrapper, convert objects from dictionary to __E class; this class allows accessing a['key'] also as a.key""" if self.use_wrapper and len(objects) > 0: class __E(dict): def __init__(self, d=dict()): self.__dict__ = self dict.__init__(self, d) if isinstance(objects, dict): if not isinstance(objects.values()[0], dict): return objects return dict((key, __E(objects[key])) for key in objects) else: return list(__E(object) for object in objects) else: return objects def _get_keystone_data(self): """get data from keystone server""" dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS save = dict_object and \ self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS keystone = self.osclients.get_keystoneclientv3() roles = keystone.roles.list() users = keystone.users.list() tenants = keystone.projects.list() roles_a = keystone.role_assignments.list() if OpenStackMap.load_filters: ef = {'service_type': 'identity', 'interface': 'public'} resp = keystone.session.get('/OS-EP-FILTER/endpoint_groups', endpoint_filter=ef) filters = resp.json()['endpoint_groups'] projects_by_filter = dict() filters_by_project = dict() for f in filters: filter_id = f['id'] resp = keystone.session.get( '/OS-EP-FILTER/endpoint_groups/' + filter_id + '/projects', endpoint_filter=ef) projects = resp.json()['projects'] projects_by_filter[filter_id] = projects for project in projects: if project['id'] not in filters_by_project: filters_by_project[project['id']] = list() filters_by_project[project['id']].append(filter_id) else: filters = list() filters_by_project = dict() roles_by_user = dict() roles_by_project = dict() for roleasig in roles_a: try: userid = roleasig.user['id'] if 'project' in roleasig.scope: projectid = roleasig.scope['project']['id'] else: projectid = str(roleasig.scope) roleid = roleasig.role['id'] if userid not in roles_by_user: roles_by_user[userid] = list() except AttributeError: # Discard roles not assigned to users continue if projectid not in roles_by_project: roles_by_project[projectid] = list() roles_by_user[userid].append((roleid, projectid)) roles_by_project[projectid].append((roleid, userid)) self.roles_by_user = roles_by_user self.roles_by_project = roles_by_project tenants_by_name = dict() users_by_name = dict() filters = dict((f['id'], f) for f in filters) if dict_object: roles = dict((role.id, role.to_dict()) for role in roles) users = dict((user.id, user.to_dict()) for user in users) tenants = dict((tenant.id, tenant.to_dict()) for tenant in tenants) roles_a = list(asig.to_dict() for asig in roles_a) for tenant in tenants.values(): tenants_by_name[tenant['name']] = tenant for user in users.values(): users_by_name[user['name']] = user else: self.roles = dict((role.id, role) for role in roles) self.users = dict((user.id, user) for user in users) self.tenants = dict((tenant.id, tenant) for tenant in tenants) self.roles_a = roles_a tenants_by_name = dict() for tenant in tenants: if getattr(tenant, 'name', None): tenants_by_name[tenant.name] = tenant for user in users: if getattr(user, 'name', None): users_by_name[user.name] = user self.tenants_by_name = tenants_by_name self.users_by_name = users_by_name self.filters = filters self.filters_by_project = filters_by_project return if save: with open(self.pers_keystone + '/roles.pickle', 'wb') as f: pickle.dump(roles, f, protocol=-1) with open(self.pers_keystone + '/users.pickle', 'wb') as f: pickle.dump(users, f, protocol=-1) with open(self.pers_keystone + '/users_by_name.pickle', 'wb') as f: pickle.dump(users_by_name, f, protocol=-1) with open(self.pers_keystone + '/tenants.pickle', 'wb') as f: pickle.dump(tenants, f, protocol=-1) with open(self.pers_keystone + '/tenants_by_name.pickle', 'wb') as\ f: pickle.dump(tenants_by_name, f) with open(self.pers_keystone + '/roles_a.pickle', 'wb') as f: pickle.dump(roles_a, f, protocol=-1) with open(self.pers_keystone + '/roles_by_user.pickle', 'wb') as f: pickle.dump(roles_by_user, f, protocol=-1) with open(self.pers_keystone + '/roles_by_project.pickle', 'wb') as\ f: pickle.dump(roles_by_project, f, protocol=-1) with open(self.pers_keystone + '/filters.pickle', 'wb') as f: pickle.dump(filters, f, protocol=-1) with open(self.pers_keystone + '/filters_by_project.pickle', 'wb') \ as f: pickle.dump(filters_by_project, f, protocol=-1) self.roles = self._convert(roles) self.users = self._convert(users) self.users_by_name = self._convert(users_by_name) self.tenants = self._convert(tenants) self.tenants_by_name = self._convert(tenants_by_name) self.roles_a = self._convert(roles_a) self.filters = self._convert(filters) self.filters_by_project = self._convert(filters_by_project) def _get_nova_data(self): """ get data from nova""" dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS save = dict_object and \ self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS nova = self.osclients.get_novaclient() vms = nova.servers.list(search_opts={'all_tenants': 1}) if dict_object: vms = dict((vm.id, vm.to_dict()) for vm in vms) else: self.vms = dict((vm.id, vm) for vm in vms) return if save: with open(self.pers_region + '/vms.pickle', 'wb') as f: pickle.dump(vms, f, protocol=-1) self.vms = self._convert(vms) def _get_cinder_data(self): """get data from cinder""" dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS save = dict_object and \ self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS cinder = self.osclients.get_cinderclientv1() volumes = cinder.volumes.list(search_opts={'all_tenants': 1}) snapshots = cinder.volume_snapshots.list( search_opts={'all_tenants': 1}) backups = cinder.backups.list(search_opts={'all_tenants': 1}) if dict_object: volumes = dict((volume.id, volume.__dict__) for volume in volumes) snapshots = dict((snapshot.id, snapshot.__dict__) for snapshot in snapshots) backups = dict((backup.id, backup.__dict__) for backup in backups) else: self.volumes = dict((volume.id, volume) for volume in volumes) self.volume_snapshots = dict((snapshot.id, snapshot) for snapshot in snapshots) self.volume_backups = dict((backup.id, backup) for backup in backups) return if save: with open(self.pers_region + '/volumes.pickle', 'wb') as f: pickle.dump(volumes, f, protocol=-1) with open(self.pers_region + '/volume_snapshots.pickle', 'wb') as \ f: pickle.dump(snapshots, f, protocol=-1) with open(self.pers_region + '/volume_backups.pickle', 'wb') as f: pickle.dump(backups, f, protocol=-1) self.volumes = self._convert(volumes) self.volume_snapshots = self._convert(snapshots) self.volume_backups = self._convert(backups) def _get_glance_data(self): """get data from glance""" dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS save = dict_object and \ self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS glance = self.osclients.get_glanceclient() images = glance.images.findall() if dict_object: images = dict((image.id, image.to_dict()) for image in glance.images.findall()) else: self.images = dict((image.id, image) for image in images) return if save: with open(self.pers_region + '/images.pickle', 'wb') as f: pickle.dump(images, f, protocol=-1) self.images = self._convert(images) def _get_neutron_data(self): """get network data from neutron""" dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS save = dict_object and \ self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS neutron = self.osclients.get_neutronclient() networks = neutron.list_networks()['networks'] subnets = neutron.list_subnets()['subnets'] routers = neutron.list_routers()['routers'] sec_grps = neutron.list_security_groups()['security_groups'] floatingips = neutron.list_floatingips()['floatingips'] ports = neutron.list_ports()['ports'] nets = dict((network['id'], network) for network in networks) snets = dict((subnet['id'], subnet) for subnet in subnets) routers = dict((router['id'], router) for router in routers) floatingips = dict((floatingip['id'], floatingip) for floatingip in floatingips) sec_grps = dict((sg['id'], sg) for sg in sec_grps) ports = dict((port['id'], port) for port in ports) if dict_object: # neutron objects are already dicts, so they do not need conversion pass else: self.networks = nets self.subnets = snets self.routers = routers self.floatingips = floatingips self.security_groups = sec_grps self.ports = ports return if save: with open(self.pers_region + '/networks.pickle', 'wb') as f: pickle.dump(nets, f, protocol=-1) with open(self.pers_region + '/subnets.pickle', 'wb') as f: pickle.dump(snets, f, protocol=-1) with open(self.pers_region + '/routers.pickle', 'wb') as f: pickle.dump(routers, f, protocol=-1) with open(self.pers_region + '/floatingips.pickle', 'wb') as f: pickle.dump(floatingips, f, protocol=-1) with open(self.pers_region + '/security_groups.pickle', 'wb') as f: pickle.dump(sec_grps, f, protocol=-1) with open(self.pers_region + '/ports.pickle', 'wb') as f: pickle.dump(ports, f, protocol=-1) self.networks = self._convert(nets) self.subnets = self._convert(snets) self.routers = self._convert(routers) self.floatingips = self._convert(floatingips) self.security_groups = self._convert(sec_grps) self.ports = self._convert(ports) def load_nova(self): """load nova data: vms""" if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \ os.path.exists(self.pers_region + '/vms.pickle'): self.vms = self._load('vms') else: if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY: raise Exception('Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about nova') self._get_nova_data() def load_glance(self): """load glance data: images""" if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \ os.path.exists(self.pers_region + '/images.pickle'): self.images = self._load('images') else: if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY: raise Exception('Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about glance') self._get_glance_data() def load_neutron(self): """load neutron (network) data: networks, subnets, routers, floatingips, security_groups, ports""" if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \ os.path.exists(self.pers_region + '/networks.pickle'): self.networks = self._load('networks') # legacy if os.path.exists(self.pers_region + '/subnetworks.pickle'): self.subnets = self._load('subnetworks') else: self.subnets = self._load('subnets') self.routers = self._load('routers') self.floatingips = self._load('floatingips') # legacy if os.path.exists(self.pers_region + '/securitygroups.pickle'): self.security_groups = self._load('securitygroups') else: self.security_groups = self._load('security_groups') self.ports = self._load('ports') else: if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY: raise Exception('Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about neutron') self._get_neutron_data() # make indexes self.floatingips_by_ip = dict((f['floating_ip_address'], f) for f in self.floatingips.values() if 'floating_ip_address' in f) def load_keystone(self): """load keystone data: users, tenants, roles, roles_a, users_by_name, tenants_by_name, roles_by_project, roles_by_user """ if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \ os.path.exists(self.pers_keystone + '/users.pickle'): self.users = self._load_fkeystone('users') self.users_by_name = self._load_fkeystone('users_by_name') self.tenants = self._load_fkeystone('tenants') self.tenants_by_name = self._load_fkeystone('tenants_by_name') # legacy code if os.path.exists(self.pers_keystone + '/asignments.pickle'): self.roles_a = self._load_fkeystone('asignments') else: self.roles_a = self._load_fkeystone('roles_a') self.roles = self._load_fkeystone('roles') self.roles_by_project = self._load_fkeystone('roles_by_project') self.roles_by_user = self._load_fkeystone('roles_by_user') self.filters = self._load_fkeystone('filters') # legacy code if os.path.exists(self.pers_keystone + '/filters_byproject.pickle'): self.filters_by_project = self._load_fkeystone('filters_byproject') else: self.filters_by_project = self._load_fkeystone('filters_by_project') else: if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY: raise Exception('Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about keystone') self._get_keystone_data() def load_cinder(self): """load cinder data: volumes, volume_backups, volume_snapshots """ if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \ os.path.exists(self.pers_region + '/volumes.pickle'): self.volumes = self._load('volumes') self.volume_backups = self._load('volume_backups') self.volume_snapshots = self._load('volume_snapshots') else: if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY: raise Exception('Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about cinder') self._get_cinder_data() def load_all(self): """load all data""" region = self.osclients.region if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY: if os.path.exists(self.pers_region + '/vms.pickle'): self.load_nova() if os.path.exists(self.pers_region + '/networks.pickle'): self.load_neutron() if os.path.exists(self.pers_region + '/images.pickle'): self.load_glance() if os.path.exists(self.pers_region + '/volumes.pickle'): self.load_cinder() else: if region in self.osclients.get_regions('compute'): self.load_nova() if region in self.osclients.get_regions('network'): self.load_neutron() if region in self.osclients.get_regions('image'): self.load_glance() if region in self.osclients.get_regions('volume'): self.load_cinder() self.load_keystone() def change_region(self, region, auto_load=True): """change region and clean maps. Optionally load the maps. :param region: the new region :param auto_load: True to invoke load_all :return: nothing """ self.pers_region = self.persistence_dir + '/' + region if not os.path.exists(self.pers_region) and self.objects_strategy\ not in (OpenStackMap.DIRECT_OBJECTS, OpenStackMap.NO_CACHE_OBJECTS): os.mkdir(self.pers_region) self.osclients.set_region(region) self._init_resource_maps() if auto_load: self.load_all() def preload_regions(self, regions=None, all_regions_excluded=None): """Method to preload the data of the specified regions. If regions is None, use all the available regions in the federation, but the specified in all_regions_excluded. The data for each region will be available at the region_map dictionary. It must be noted that this method could affect the current region and the values of the direct maps (vms, networks...). This is because it calls change_region/load_<service_name> methods. If regions is provided, then the last region in the list will be the new current region""" if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY: regions_in_disk = set(os.listdir(self.persistence_dir)) if 'keystone' in regions_in_disk: regions_in_disk.remove('keystone') regions_compute = set() regions_network = set() regions_image = set() regions_volume = set() for region in regions_in_disk: path = self.persistence_dir + os.path.sep + region if os.path.exists(path + os.path.sep + 'vms.pickle'): regions_compute.add(region) if os.path.exists(path + os.path.sep + 'networks.pickle'): regions_network.add(region) if os.path.exists(path + os.path.sep + 'images.pickle'): regions_image.add(region) if os.path.exists(path + os.path.sep + 'volumes.pickle'): regions_volume.add(region) else: regions_compute = self.osclients.get_regions('compute') regions_network = self.osclients.get_regions('network') regions_image = self.osclients.get_regions('image') regions_volume = self.osclients.get_regions('volume') if not regions: """ not regions specified, so all existing regions with some resource (compute, network, image or volume) are considered. If all_regions_excluded is defined, then these regions are excluded """ all_regions = regions_compute.union(regions_network).union( regions_image).union(regions_volume) if all_regions_excluded: all_regions.difference_update(all_regions_excluded) # Move the current region to the end of the list. # This is because the direct map dictionaries (vms, networks, etc.) # and the region field are updated with the change_region() call. if self.osclients.region in all_regions: all_regions.remove(self.osclients.region) regions = list(all_regions) regions.append(self.osclients.region) for region in regions: try: self.logger.info('Creating map of region ' + region) self.change_region(region, False) if region in regions_compute: self.load_nova() if region in regions_network: self.load_neutron() if region in regions_image: self.load_glance() if region in regions_volume: self.load_cinder() region_map = dict() for resource in self.resources_region: region_map[resource] = getattr(self, resource) self.region_map[region] = region_map except Exception, e: msg = 'Failed the creation of the map of {0}. Cause: {1}' self.logger.error(msg.format(region, str(e))) # Remove dir if it exists and is empty. dir = os.path.join(self.persistence_dir, region) if os.path.isdir(dir) and len(os.listdir(dir)) == 0: os.rmdir(dir) self.load_keystone()
class OpenStackMap(object): """ This class build a map from the resources (VMs, networks, images, volumes, users, tenants, roles...) in an OpenStack infrastructure. This map is optionally cached on disk and can be accessed offline or can be used as a metadata snapshot of an OpenStack infrastructure (e.g. to analyse resource use and build statistics) """ # objects strategy see the __init__ method documentation DIRECT_OBJECTS, NO_CACHE_OBJECTS, REFRESH_OBJECTS, USE_CACHE_OBJECTS, USE_CACHE_OBJECTS_ONLY = \ range(5) # If use_wrapper is True, dictionaries are wrapped to allow access to # resource['field'] also as resource.field. This is not used when # objects_strategy is DIRECT_OBJECTS use_wrapper = True load_filters = True resources_region = [ 'vms', 'images', 'routers', 'networks', 'subnets', 'ports', 'floatingips', 'security_groups', 'volumes', 'volume_backups', 'volume_snapshots' ] def __init__(self, persistence_dir='~/openstackmap', region=None, auth_url=None, objects_strategy=USE_CACHE_OBJECTS, auto_load=True): """ Constructor :param persistence_dir: The path where the data is saved. Ignored if objects_strategy is DIRECT_OBJECTS or NO_CACHE_OBJECTS :param region: the initial region (if undefined, use OS_REGION_NAME) :param auth_url: the keystone URI :param objects_strategy: sets the strategy about the object maps contained here. It can be: * DIRECT_OBJECTS is using the objects as they are got from the API. Be careful because methods as delete are available! The objects are not cached. * NO_CACHE_OBJECTS is using the objects converted to dictionaries, so methods to do operations are not available. The objects are not cached. * REFRESH_OBJECTS is using the objects converted to dictionaries. The objects are cached: the new version replace the old one. * USE_CACHE_OBJECTS is using the objects converted to dictionaries. If a cached copy of the objects are available, it is used. * USE_CACHE_OBJECTS_ONLY is using the objects converted to dictionaries. This strategy used cached objects only. It never contacts with the servers, even when the object is not available in the local cache. :param auto_load: if True, invoke self.load_all() Note that neutron objects returned by the API are already dictionaries """ self.logger = logging.getLogger(__name__) if auth_url: self.osclients = OpenStackClients(auth_url=auth_url) else: self.osclients = OpenStackClients() if region: self.osclients.set_region(region) else: if 'OS_REGION_NAME' not in env: raise Exception('Region parameter must be provided or ' 'OS_REGION_NAME variable must be defined.') else: region = env['OS_REGION_NAME'] if 'KEYSTONE_ADMIN_ENDPOINT' in os.environ: self.osclients.override_endpoint( 'identity', self.osclients.region, 'admin', os.environ['KEYSTONE_ADMIN_ENDPOINT']) self.objects_strategy = objects_strategy self.persistence_dir = os.path.expanduser(persistence_dir) self.pers_region = self.persistence_dir + '/' + region self.pers_keystone = self.persistence_dir + '/keystone' if objects_strategy not in (OpenStackMap.DIRECT_OBJECTS, OpenStackMap.NO_CACHE_OBJECTS): if not os.path.exists(self.persistence_dir): os.mkdir(self.persistence_dir) if not os.path.exists(self.pers_keystone): os.mkdir(self.pers_keystone) if not os.path.exists(self.pers_region): os.mkdir(self.pers_region) self._init_resource_maps() if auto_load: self.load_all() self.region_map = dict() def _init_resource_maps(self): """init all the resources that will be available as empty dictionaries""" # Keystone resources self.users = dict() self.users_by_name = dict() self.roles = dict() self.tenants = dict() self.tenants_by_name = dict() self.roles_a = list() self.roles_by_project = dict() self.roles_by_user = dict() self.filters = dict() self.filters_by_project = dict() # Glance resources self.images = dict() # Neutron resources self.networks = dict() self.subnets = dict() self.routers = dict() self.floatingips = dict() self.floatingips_by_ip = dict() self.ports = dict() self.security_groups = dict() # Nova resources self.vms = dict() self.flavors = dict() # Cinder resources self.volumes = dict() self.volume_backups = dict() self.volume_snapshots = dict() def _load(self, name): """Load the resources persisted with pickle. This resources (e.g. networks, vms, images...) are saved independently for each region. This method is called for load_cinder, load_nova, load_glance, load_neutron... :param name: the resource name :return: a dictionary of objects (dictionaries) indexed by id """ objects = pickle.load( open(self.pers_region + '/' + name + '.pickle', 'rb')) return self._convert(objects) def _load_fkeystone(self, name): """Load the keystone objects persisted with pickle. This resources are shared among the regions. This method is used in load_keystone :param name: the resource name :return: a list/dictionary of objects (dictionaries) """ objects = pickle.load( open(self.pers_keystone + '/' + name + '.pickle', 'rb')) return self._convert(objects) def _convert(self, objects): """if use_wrapper, convert objects from dictionary to __E class; this class allows accessing a['key'] also as a.key""" if self.use_wrapper and len(objects) > 0: class __E(dict): def __init__(self, d=dict()): self.__dict__ = self dict.__init__(self, d) if isinstance(objects, dict): if not isinstance(objects.values()[0], dict): return objects return dict((key, __E(objects[key])) for key in objects) else: return list(__E(object) for object in objects) else: return objects def _get_keystone_data(self): """get data from keystone server""" dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS save = dict_object and \ self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS keystone = self.osclients.get_keystoneclientv3() roles = keystone.roles.list() users = keystone.users.list() tenants = keystone.projects.list() roles_a = keystone.role_assignments.list() if OpenStackMap.load_filters: ef = {'service_type': 'identity', 'interface': 'public'} resp = keystone.session.get('/OS-EP-FILTER/endpoint_groups', endpoint_filter=ef) filters = resp.json()['endpoint_groups'] projects_by_filter = dict() filters_by_project = dict() for f in filters: filter_id = f['id'] resp = keystone.session.get('/OS-EP-FILTER/endpoint_groups/' + filter_id + '/projects', endpoint_filter=ef) projects = resp.json()['projects'] projects_by_filter[filter_id] = projects for project in projects: if project['id'] not in filters_by_project: filters_by_project[project['id']] = list() filters_by_project[project['id']].append(filter_id) else: filters = list() filters_by_project = dict() roles_by_user = dict() roles_by_project = dict() for roleasig in roles_a: try: userid = roleasig.user['id'] if 'project' in roleasig.scope: projectid = roleasig.scope['project']['id'] else: projectid = str(roleasig.scope) roleid = roleasig.role['id'] if userid not in roles_by_user: roles_by_user[userid] = list() except AttributeError: # Discard roles not assigned to users continue if projectid not in roles_by_project: roles_by_project[projectid] = list() roles_by_user[userid].append((roleid, projectid)) roles_by_project[projectid].append((roleid, userid)) self.roles_by_user = roles_by_user self.roles_by_project = roles_by_project tenants_by_name = dict() users_by_name = dict() filters = dict((f['id'], f) for f in filters) if dict_object: roles = dict((role.id, role.to_dict()) for role in roles) users = dict((user.id, user.to_dict()) for user in users) tenants = dict((tenant.id, tenant.to_dict()) for tenant in tenants) roles_a = list(asig.to_dict() for asig in roles_a) for tenant in tenants.values(): tenants_by_name[tenant['name']] = tenant for user in users.values(): users_by_name[user['name']] = user else: self.roles = dict((role.id, role) for role in roles) self.users = dict((user.id, user) for user in users) self.tenants = dict((tenant.id, tenant) for tenant in tenants) self.roles_a = roles_a tenants_by_name = dict() for tenant in tenants: if getattr(tenant, 'name', None): tenants_by_name[tenant.name] = tenant for user in users: if getattr(user, 'name', None): users_by_name[user.name] = user self.tenants_by_name = tenants_by_name self.users_by_name = users_by_name self.filters = filters self.filters_by_project = filters_by_project return if save: with open(self.pers_keystone + '/roles.pickle', 'wb') as f: pickle.dump(roles, f, protocol=-1) with open(self.pers_keystone + '/users.pickle', 'wb') as f: pickle.dump(users, f, protocol=-1) with open(self.pers_keystone + '/users_by_name.pickle', 'wb') as f: pickle.dump(users_by_name, f, protocol=-1) with open(self.pers_keystone + '/tenants.pickle', 'wb') as f: pickle.dump(tenants, f, protocol=-1) with open(self.pers_keystone + '/tenants_by_name.pickle', 'wb') as\ f: pickle.dump(tenants_by_name, f) with open(self.pers_keystone + '/roles_a.pickle', 'wb') as f: pickle.dump(roles_a, f, protocol=-1) with open(self.pers_keystone + '/roles_by_user.pickle', 'wb') as f: pickle.dump(roles_by_user, f, protocol=-1) with open(self.pers_keystone + '/roles_by_project.pickle', 'wb') as\ f: pickle.dump(roles_by_project, f, protocol=-1) with open(self.pers_keystone + '/filters.pickle', 'wb') as f: pickle.dump(filters, f, protocol=-1) with open(self.pers_keystone + '/filters_by_project.pickle', 'wb') \ as f: pickle.dump(filters_by_project, f, protocol=-1) self.roles = self._convert(roles) self.users = self._convert(users) self.users_by_name = self._convert(users_by_name) self.tenants = self._convert(tenants) self.tenants_by_name = self._convert(tenants_by_name) self.roles_a = self._convert(roles_a) self.filters = self._convert(filters) self.filters_by_project = self._convert(filters_by_project) def _get_nova_data(self): """ get data from nova""" dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS save = dict_object and \ self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS nova = self.osclients.get_novaclient() vms = nova.servers.list(search_opts={'all_tenants': 1}) if dict_object: vms = dict((vm.id, vm.to_dict()) for vm in vms) else: self.vms = dict((vm.id, vm) for vm in vms) return if save: with open(self.pers_region + '/vms.pickle', 'wb') as f: pickle.dump(vms, f, protocol=-1) self.vms = self._convert(vms) def _get_cinder_data(self): """get data from cinder""" dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS save = dict_object and \ self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS cinder = self.osclients.get_cinderclientv1() volumes = cinder.volumes.list(search_opts={'all_tenants': 1}) snapshots = cinder.volume_snapshots.list( search_opts={'all_tenants': 1}) backups = cinder.backups.list(search_opts={'all_tenants': 1}) if dict_object: volumes = dict((volume.id, volume.__dict__) for volume in volumes) snapshots = dict( (snapshot.id, snapshot.__dict__) for snapshot in snapshots) backups = dict((backup.id, backup.__dict__) for backup in backups) else: self.volumes = dict((volume.id, volume) for volume in volumes) self.volume_snapshots = dict( (snapshot.id, snapshot) for snapshot in snapshots) self.volume_backups = dict( (backup.id, backup) for backup in backups) return if save: with open(self.pers_region + '/volumes.pickle', 'wb') as f: pickle.dump(volumes, f, protocol=-1) with open(self.pers_region + '/volume_snapshots.pickle', 'wb') as \ f: pickle.dump(snapshots, f, protocol=-1) with open(self.pers_region + '/volume_backups.pickle', 'wb') as f: pickle.dump(backups, f, protocol=-1) self.volumes = self._convert(volumes) self.volume_snapshots = self._convert(snapshots) self.volume_backups = self._convert(backups) def _get_glance_data(self): """get data from glance""" dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS save = dict_object and \ self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS glance = self.osclients.get_glanceclient() images = glance.images.findall() if dict_object: images = dict((image.id, image.to_dict()) for image in glance.images.findall()) else: self.images = dict((image.id, image) for image in images) return if save: with open(self.pers_region + '/images.pickle', 'wb') as f: pickle.dump(images, f, protocol=-1) self.images = self._convert(images) def _get_neutron_data(self): """get network data from neutron""" dict_object = self.objects_strategy != OpenStackMap.DIRECT_OBJECTS save = dict_object and \ self.objects_strategy != OpenStackMap.NO_CACHE_OBJECTS neutron = self.osclients.get_neutronclient() networks = neutron.list_networks()['networks'] subnets = neutron.list_subnets()['subnets'] routers = neutron.list_routers()['routers'] sec_grps = neutron.list_security_groups()['security_groups'] floatingips = neutron.list_floatingips()['floatingips'] ports = neutron.list_ports()['ports'] nets = dict((network['id'], network) for network in networks) snets = dict((subnet['id'], subnet) for subnet in subnets) routers = dict((router['id'], router) for router in routers) floatingips = dict( (floatingip['id'], floatingip) for floatingip in floatingips) sec_grps = dict((sg['id'], sg) for sg in sec_grps) ports = dict((port['id'], port) for port in ports) if dict_object: # neutron objects are already dicts, so they do not need conversion pass else: self.networks = nets self.subnets = snets self.routers = routers self.floatingips = floatingips self.security_groups = sec_grps self.ports = ports return if save: with open(self.pers_region + '/networks.pickle', 'wb') as f: pickle.dump(nets, f, protocol=-1) with open(self.pers_region + '/subnets.pickle', 'wb') as f: pickle.dump(snets, f, protocol=-1) with open(self.pers_region + '/routers.pickle', 'wb') as f: pickle.dump(routers, f, protocol=-1) with open(self.pers_region + '/floatingips.pickle', 'wb') as f: pickle.dump(floatingips, f, protocol=-1) with open(self.pers_region + '/security_groups.pickle', 'wb') as f: pickle.dump(sec_grps, f, protocol=-1) with open(self.pers_region + '/ports.pickle', 'wb') as f: pickle.dump(ports, f, protocol=-1) self.networks = self._convert(nets) self.subnets = self._convert(snets) self.routers = self._convert(routers) self.floatingips = self._convert(floatingips) self.security_groups = self._convert(sec_grps) self.ports = self._convert(ports) def load_nova(self): """load nova data: vms""" if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \ os.path.exists(self.pers_region + '/vms.pickle'): self.vms = self._load('vms') else: if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY: raise Exception( 'Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about nova' ) self._get_nova_data() def load_glance(self): """load glance data: images""" if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \ os.path.exists(self.pers_region + '/images.pickle'): self.images = self._load('images') else: if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY: raise Exception( 'Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about glance' ) self._get_glance_data() def load_neutron(self): """load neutron (network) data: networks, subnets, routers, floatingips, security_groups, ports""" if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \ os.path.exists(self.pers_region + '/networks.pickle'): self.networks = self._load('networks') # legacy if os.path.exists(self.pers_region + '/subnetworks.pickle'): self.subnets = self._load('subnetworks') else: self.subnets = self._load('subnets') self.routers = self._load('routers') self.floatingips = self._load('floatingips') # legacy if os.path.exists(self.pers_region + '/securitygroups.pickle'): self.security_groups = self._load('securitygroups') else: self.security_groups = self._load('security_groups') self.ports = self._load('ports') else: if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY: raise Exception( 'Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about neutron' ) self._get_neutron_data() # make indexes self.floatingips_by_ip = dict((f['floating_ip_address'], f) for f in self.floatingips.values() if 'floating_ip_address' in f) def load_keystone(self): """load keystone data: users, tenants, roles, roles_a, users_by_name, tenants_by_name, roles_by_project, roles_by_user """ if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \ os.path.exists(self.pers_keystone + '/users.pickle'): self.users = self._load_fkeystone('users') self.users_by_name = self._load_fkeystone('users_by_name') self.tenants = self._load_fkeystone('tenants') self.tenants_by_name = self._load_fkeystone('tenants_by_name') # legacy code if os.path.exists(self.pers_keystone + '/asignments.pickle'): self.roles_a = self._load_fkeystone('asignments') else: self.roles_a = self._load_fkeystone('roles_a') self.roles = self._load_fkeystone('roles') self.roles_by_project = self._load_fkeystone('roles_by_project') self.roles_by_user = self._load_fkeystone('roles_by_user') self.filters = self._load_fkeystone('filters') # legacy code if os.path.exists(self.pers_keystone + '/filters_byproject.pickle'): self.filters_by_project = self._load_fkeystone( 'filters_byproject') else: self.filters_by_project = self._load_fkeystone( 'filters_by_project') else: if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY: raise Exception( 'Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about keystone' ) self._get_keystone_data() def load_cinder(self): """load cinder data: volumes, volume_backups, volume_snapshots """ if (self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS or self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY) and \ os.path.exists(self.pers_region + '/volumes.pickle'): self.volumes = self._load('volumes') self.volume_backups = self._load('volume_backups') self.volume_snapshots = self._load('volume_snapshots') else: if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY: raise Exception( 'Strategy is USE_CACHE_OBJECTS_ONLY but there are not cached data about cinder' ) self._get_cinder_data() def load_all(self): """load all data""" region = self.osclients.region if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY: if os.path.exists(self.pers_region + '/vms.pickle'): self.load_nova() if os.path.exists(self.pers_region + '/networks.pickle'): self.load_neutron() if os.path.exists(self.pers_region + '/images.pickle'): self.load_glance() if os.path.exists(self.pers_region + '/volumes.pickle'): self.load_cinder() else: if region in self.osclients.get_regions('compute'): self.load_nova() if region in self.osclients.get_regions('network'): self.load_neutron() if region in self.osclients.get_regions('image'): self.load_glance() if region in self.osclients.get_regions('volume'): self.load_cinder() self.load_keystone() def change_region(self, region, auto_load=True): """change region and clean maps. Optionally load the maps. :param region: the new region :param auto_load: True to invoke load_all :return: nothing """ self.pers_region = self.persistence_dir + '/' + region if not os.path.exists(self.pers_region) and self.objects_strategy\ not in (OpenStackMap.DIRECT_OBJECTS, OpenStackMap.NO_CACHE_OBJECTS): os.mkdir(self.pers_region) self.osclients.set_region(region) self._init_resource_maps() if auto_load: self.load_all() def preload_regions(self, regions=None, all_regions_excluded=None): """Method to preload the data of the specified regions. If regions is None, use all the available regions in the federation, but the specified in all_regions_excluded. The data for each region will be available at the region_map dictionary. It must be noted that this method could affect the current region and the values of the direct maps (vms, networks...). This is because it calls change_region/load_<service_name> methods. If regions is provided, then the last region in the list will be the new current region""" if self.objects_strategy == OpenStackMap.USE_CACHE_OBJECTS_ONLY: regions_in_disk = set(os.listdir(self.persistence_dir)) if 'keystone' in regions_in_disk: regions_in_disk.remove('keystone') regions_compute = set() regions_network = set() regions_image = set() regions_volume = set() for region in regions_in_disk: path = self.persistence_dir + os.path.sep + region if os.path.exists(path + os.path.sep + 'vms.pickle'): regions_compute.add(region) if os.path.exists(path + os.path.sep + 'networks.pickle'): regions_network.add(region) if os.path.exists(path + os.path.sep + 'images.pickle'): regions_image.add(region) if os.path.exists(path + os.path.sep + 'volumes.pickle'): regions_volume.add(region) else: regions_compute = self.osclients.get_regions('compute') regions_network = self.osclients.get_regions('network') regions_image = self.osclients.get_regions('image') regions_volume = self.osclients.get_regions('volume') if not regions: """ not regions specified, so all existing regions with some resource (compute, network, image or volume) are considered. If all_regions_excluded is defined, then these regions are excluded """ all_regions = regions_compute.union(regions_network).union( regions_image).union(regions_volume) if all_regions_excluded: all_regions.difference_update(all_regions_excluded) # Move the current region to the end of the list. # This is because the direct map dictionaries (vms, networks, etc.) # and the region field are updated with the change_region() call. if self.osclients.region in all_regions: all_regions.remove(self.osclients.region) regions = list(all_regions) regions.append(self.osclients.region) for region in regions: try: self.logger.info('Creating map of region ' + region) self.change_region(region, False) if region in regions_compute: self.load_nova() if region in regions_network: self.load_neutron() if region in regions_image: self.load_glance() if region in regions_volume: self.load_cinder() region_map = dict() for resource in self.resources_region: region_map[resource] = getattr(self, resource) self.region_map[region] = region_map except Exception, e: msg = 'Failed the creation of the map of {0}. Cause: {1}' self.logger.error(msg.format(region, str(e))) # Remove dir if it exists and is empty. dir = os.path.join(self.persistence_dir, region) if os.path.isdir(dir) and len(os.listdir(dir)) == 0: os.rmdir(dir) self.load_keystone()
class UserResources(object): """Class to list, delete user resources. Also provides a method to stop all the VM of the tenant. This class works creating a instance with the credential of the user owner of the resources. It does not use an admin credential!!!""" def __init__(self, username, password, tenant_id=None, tenant_name=None, trust_id=None): """ Constructor of the class. Tenant_id or tenant_name or tust_id must be provided. There are two ways of using this class: -passing the user, password, tenant_id or tenant_name of the user whose resources are being deleted -passing the user and password of the trustee, and the trust_id generated to impersonate the trustor. :param username: the user name whose resources are deleted or the user name of the trustee :param password: the password of the user whose resources are deleted or the password of the trustee :param tenant_id: the tenant id of the user whose resources must be deleted. :param tenant_name: the tenant name of the user whose resources must be deleted :param trust_id: the trust_id used to impersonate the user whose resources must be deleted :return: nothing """ self.logger = logging.getLogger(__name__) self.clients = OpenStackClients() if tenant_id: self.clients.set_credential(username, password, tenant_id=tenant_id) elif tenant_name: self.clients.set_credential(username, password, tenant_name=tenant_name) elif trust_id: self.clients.set_credential(username, password, trust_id=trust_id) self.trust_id = trust_id else: raise( 'Either tenant_id or tenant_name or trust_id must be provided') region = self.clients.region self.clients.override_endpoint( 'identity', region, 'admin', settings.KEYSTONE_ENDPOINT) self.user_id = self.clients.get_session().get_user_id() session = self.clients.get_session() self.user_name = username self.nova = NovaResources(self.clients) self.cinder = CinderResources(self.clients) self.glance = GlanceResources(self.clients) try: self.neutron = NeutronResources(self.clients) except Exception: # The region does not support Neutron # It would be better to check the endpoint self.neutron = None try: self.blueprints = BluePrintResources(self.clients) except Exception: # The region does not support PaaS Manager # It would be better to check the endpoint self.blueprints = None try: self.swift = SwiftResources(self.clients) except Exception: # The region does not support Swift # It would be better to check the endpoint self.swift = None # Images in use is a set used to avoid deleting formerly glance images # in use by other tenants self.imagesinuse = set() # Regions the user has access self.regions_available = set() self.regions_available.update(self.clients.get_regions('compute')) def change_region(self, region): """ change the region. All the clients need to be updated, but the session does not. :param region: the name of the region :return: nothing. """ self.clients.set_region(region) self.clients.override_endpoint( 'identity', region, 'admin', settings.KEYSTONE_ENDPOINT) self.nova.on_region_changed() self.glance.on_region_changed() try: if self.swift: self.swift.on_region_changed() else: self.swift = SwiftResources(self.clients) except Exception: # The region does not support swift self.swift = None self.cinder.on_region_changed() try: if self.blueprints: self.blueprint.on_region_changed() else: self.blueprints = BluePrintResources(self.clients) except Exception: # The region has not configured paas manager self.blueprints = None try: if self.neutron: self.neutron.on_region_changed() else: self.neutron = NeutronResources(self.clients) except Exception: # The region does not support neutron self.neutron = None def delete_tenant_resources_pri_1(self): """Delete here all the elements that do not depend of others are deleted first""" try: self.nova.delete_user_keypairs() except Exception, e: msg = 'Deletion of keypairs failed. Reason: ' self.logger.error(msg + str(e)) # Snapshots must be deleted before the volumes, because a snapshot # depends of a volume. try: self.cinder.delete_tenant_volume_snapshots() except Exception, e: msg = 'Deletion of volume snaphosts failed. Reason: ' self.logger.error(msg + str(e))
class ServersFacade(object): def __init__(self, target): """Create a new Facade for the specified target (a target is shared between regions using the same credential)""" self.osclients = OpenStackClients(target['keystone_url']) self.osclients.set_credential(target['user'], target['password'], target['tenant']) if target.get('use_keystone_v3', False): self.osclients.set_keystone_version(True) else: self.osclients.set_keystone_version(False) self.session = self.osclients.get_session() self.target = target # This is a default value self.images_dir = '/var/lib/glance/images' self.logger = logger_cli def _get_glanceclient(self, region): """helper method, to get a glanceclient for the region""" self.osclients.set_region(region) return self.osclients.get_glanceclient() def get_regions(self): """It returns the list of regions on the specified target. :return: a list of region names. """ return self.osclients.get_regions('image') def get_imagelist(self, regionobj): """return a image list from the glance of the specified region :param regionobj: The GlanceSyncRegion object of the region to list :return: a list of GlanceSyncImage objects """ client = self._get_glanceclient(regionobj.region) try: target = regionobj.target # We need a Pool to implement a timeout. Unfortunately # setting client.images.client.timeout does nothing. if 'list_images_timeout' in target: timeout = target['list_images_timeout'] else: timeout = _default_timeout pool = Pool(1) result = pool.apply_async(_getrawimagelist, (client,)) images = result.get(timeout=timeout) image_list = list() for image in images: i = GlanceSyncImage( image['name'], image['id'], regionobj.fullname, image['owner'], image['is_public'], image['checksum'], image['size'], image['status'], image['properties'], image) image_list.append(i) except TimeoutError: msg = regionobj.fullname + \ ': Timeout while retrieving image list.' self.logger.error(msg) raise GlanceFacadeException(msg) except Exception, e: cause = str(e) if not cause: cause = repr(e) msg = regionobj.fullname + \ ': Error retrieving image list. Cause: ' + cause self.logger.error(msg) raise GlanceFacadeException(msg) return image_list
class ServersFacade(object): def __init__(self, target): """Create a new Facade for the specified target (a target is shared between regions using the same credential)""" self.osclients = OpenStackClients(target['keystone_url']) self.osclients.set_credential(target['user'], target['password'], target['tenant']) if target.get('use_keystone_v3', False): self.osclients.set_keystone_version(True) else: self.osclients.set_keystone_version(False) self.session = self.osclients.get_session() self.target = target # This is a default value self.images_dir = '/var/lib/glance/images' self.logger = logger_cli def _get_glanceclient(self, region): """helper method, to get a glanceclient for the region""" self.osclients.set_region(region) return self.osclients.get_glanceclient() def get_regions(self): """It returns the list of regions on the specified target. :return: a list of region names. """ return self.osclients.get_regions('image') def get_imagelist(self, regionobj): """return a image list from the glance of the specified region :param regionobj: The GlanceSyncRegion object of the region to list :return: a list of GlanceSyncImage objects """ client = self._get_glanceclient(regionobj.region) try: target = regionobj.target # We need a Pool to implement a timeout. Unfortunately # setting client.images.client.timeout does nothing. if 'list_images_timeout' in target: timeout = target['list_images_timeout'] else: timeout = _default_timeout pool = Pool(1) result = pool.apply_async(_getrawimagelist, (client, )) images = result.get(timeout=timeout) image_list = list() for image in images: i = GlanceSyncImage(image['name'], image['id'], regionobj.fullname, image['owner'], image['is_public'], image['checksum'], image['size'], image['status'], image['properties'], image) image_list.append(i) except TimeoutError: msg = regionobj.fullname + \ ': Timeout while retrieving image list.' self.logger.error(msg) raise GlanceFacadeException(msg) except Exception, e: cause = str(e) if not cause: cause = repr(e) msg = regionobj.fullname + \ ': Error retrieving image list. Cause: ' + cause self.logger.error(msg) raise GlanceFacadeException(msg) return image_list