def main(options): parser = ArgumentParser(prog='delete_jobs_G5k') parser.add_argument('-j', '--job_ids', dest='job_ids', type=str, help='Grid5000 job IDs') args = parser.parse_args() if args.job_ids: job_ids = g5k_experimenter.parse_job_ids(args.job_ids) print('Jobs will be deleted:') else: sites = get_g5k_sites() job_ids = get_current_oar_jobs(frontends=sites) print('All your running jobs:') print(''.join(['%s:%s\n' % (site, job_id) for job_id, site in job_ids])) decision = input('Do you want to delete those jobs [y/n]? ') if decision.lower().strip() == 'y': oardel(job_ids) print('Delete jobs successfully!') else: print('Bye bye!')
def _remove_excluded(planning, excluded_resources): """This function remove elements from planning""" # first removing the site for element in excluded_resources: if element in get_g5k_sites() and element in planning: del planning[element] # then removing specific clusters for site_pl in planning.values(): for res in list(site_pl): if res in excluded_resources: del site_pl[res] continue for element in list(site_pl[res]): if element in excluded_resources: del site_pl[res][element]
def show_resources(resources, msg='Resources', max_resources=None, queues='default'): """Print the resources in a fancy way""" if not max_resources: max_resources = {} total_hosts = 0 log = style.log_header(msg) + '\n' for site in get_g5k_sites(): site_added = False if site in resources: log += style.log_header(site).ljust(20) + ' ' + str( resources[site]) if site in max_resources: log += '/' + str(max_resources[site]) log += ' ' site_added = True for cluster in get_site_clusters(site, queues=queues): if len(list(set(get_site_clusters(site)) & set(resources.keys()))) > 0 \ and not site_added: log += style.log_header(site).ljust(20) if site in max_resources: log += '/' + str(max_resources[site]) log += ' ' site_added = True if cluster in resources: log += style.emph(cluster) + ': ' + str(resources[cluster]) if cluster in max_resources: log += '/' + str(max_resources[cluster]) log += ' ' total_hosts += resources[cluster] if site_added: log += '\n' if 'grid5000' in resources: log += style.log_header('Grid5000').ljust(20) + str( resources['grid5000']) if "grid5000" in max_resources: log += '/' + str(max_resources["grid5000"]) elif total_hosts > 0: log += style.log_header('Total ').ljust(20) + str(total_hosts) logger.info(log)
def get_job_by_name(job_name, sites=None): """ """ logger.detail('Looking for a job named %s', style.emph(job_name)) if not sites: sites = get_g5k_sites() oargrid_jobs = get_current_oargrid_jobs() if len(oargrid_jobs) > 0: for g_job in oargrid_jobs: for job in get_oargrid_job_oar_jobs(g_job): info = get_oar_job_info(job[0], job[1]) if info['name'] == job_name: logger.info('Oargridjob %s found !', style.emph(g_job)) return g_job, None running_jobs = get_current_oar_jobs(sites) for job in running_jobs: info = get_oar_job_info(job[0], job[1]) if info['name'] == job_name: logger.info('Job %s found on site %s !', style.emph(job[0]), style.host(job[1])) return job return None, None
def _set_colors(): colors = {} colors['busy'] = '#666666' rgb_colors = [(x[0]/255., x[1]/255., x[2]/255.) for x in \ [(255., 122., 122.), (255., 204., 122.), (255., 255., 122.), (255., 246., 153.), (204., 255., 122.), (122., 255., 122.), (122., 255., 255.), (122., 204., 255.), (204., 188., 255.), (255., 188., 255.)]] i_site = 0 for site in sorted(get_g5k_sites()): colors[site] = rgb_colors[i_site] i_cluster = 0 for cluster in sorted(get_site_clusters(site, queues=None)): min_index = colors[site].index(min(colors[site])) color = [0., 0., 0.] for i in range(3): color[i] = min(colors[site][i], 1.) if i == min_index: color[i] += i_cluster * 0.12 colors[cluster] = tuple(color) i_cluster += 1 i_site += 1 return colors
from wikitools import wiki, page, wikifile from os import environ, mkdir from execo import logger from networkx import is_isomorphic from networkx.readwrite import json_graph from execo_g5k.topology import g5k_graph, treemap from argparse import ArgumentParser from execo_g5k.api_utils import get_g5k_sites parser = ArgumentParser(prog=sys.argv[0], description='Update topology maps on the' + ' Grid\'5000 wiki') parser.add_argument('site', help='Choose the site') args = parser.parse_args() site = args.site if site not in get_g5k_sites(): logger.error('%s is not a valid G5K site') _json_dir = environ['HOME'] + '/.execo/topology/' try: mkdir(_json_dir) except: pass logger.setLevel('WARNING') g = g5k_graph([site]) logger.setLevel('INFO') try: with open(_json_dir + site + '.json', 'r') as infile: old_json = json.load(infile) g_old = json_graph.node_link_graph(old_json)
def get_vms_slot(vms, elements, slots, excluded_elements=None): """Return a slot with enough RAM and CPU """ chosen_slot = None mem = vms[0]['mem'] cpu = vms[0]['n_cpu'] req_ram = sum([vm['mem'] for vm in vms]) req_cpu = sum([vm['n_cpu'] for vm in vms]) / 3 logger.debug('RAM %s CPU %s', req_ram, req_cpu) for element in excluded_elements: if element in get_g5k_sites(): excluded_elements += [ cluster for cluster in get_site_clusters(element) if cluster not in excluded_elements ] if 'grid5000' in elements: clusters = [ cluster for cluster in get_g5k_clusters() if cluster not in excluded_elements and get_cluster_site not in excluded_elements ] else: clusters = [ element for element in elements if element in get_g5k_clusters() and element not in excluded_elements ] for element in elements: if element in get_g5k_sites(): clusters += [ cluster for cluster in get_site_clusters(element) if cluster not in clusters and cluster not in excluded_elements ] for slot in slots: hosts = [] for element in slot[2]: if str(element) in clusters: n_hosts = slot[2][element] for i in range(n_hosts): hosts.append(Host(str(element + '-1.' + \ get_cluster_site(element) + '.grid5000.fr'))) attr = get_CPU_RAM_FLOPS(hosts)['TOTAL'] if attr['CPU'] > req_cpu and attr['RAM'] > req_ram: chosen_slot = slot break del hosts[:] if chosen_slot is None: return None, None resources_needed = {} resources_available = chosen_slot[2] logger.debug('resources available' + pformat(resources_available)) iter_clusters = cycle(clusters) while req_ram > 0 or req_cpu > 0: cluster = iter_clusters.next() if resources_available[cluster] == 0: clusters.remove(cluster) iter_clusters = cycle(clusters) else: host = cluster + '-1' attr = get_CPU_RAM_FLOPS([host]) resources_available[cluster] -= 1 req_ram -= float(attr[host]['RAM'] / mem) * mem req_cpu -= float(attr[host]['CPU'] / cpu) * cpu if cluster not in resources_needed: resources_needed[cluster] = 0 resources_needed[cluster] += 1 if 'kavlan' in elements: resources_needed['kavlan'] = 1 logger.debug('resources needed' + pformat(resources_needed)) return chosen_slot[0], distribute_hosts(chosen_slot[2], resources_needed, excluded_elements)
def get_vms_slot(vms, elements, slots, excluded_elements=None): """Return a slot with enough RAM and CPU """ chosen_slot = None mem = vms[0]['mem'] cpu = vms[0]['n_cpu'] req_ram = sum([vm['mem'] for vm in vms]) req_cpu = sum([vm['n_cpu'] for vm in vms]) / 3 logger.debug('RAM %s CPU %s', req_ram, req_cpu) for element in excluded_elements: if element in get_g5k_sites(): excluded_elements += [cluster for cluster in get_site_clusters(element) if cluster not in excluded_elements] if 'grid5000' in elements: clusters = [cluster for cluster in get_g5k_clusters() if cluster not in excluded_elements and get_cluster_site not in excluded_elements] else: clusters = [element for element in elements if element in get_g5k_clusters() and element not in excluded_elements] for element in elements: if element in get_g5k_sites(): clusters += [cluster for cluster in get_site_clusters(element) if cluster not in clusters and cluster not in excluded_elements] for slot in slots: hosts = [] for element in slot[2]: if str(element) in clusters: n_hosts = slot[2][element] for i in range(n_hosts): hosts.append(Host(str(element + '-1.' + \ get_cluster_site(element) + '.grid5000.fr'))) attr = get_CPU_RAM_FLOPS(hosts)['TOTAL'] if attr['CPU'] > req_cpu and attr['RAM'] > req_ram: chosen_slot = slot break del hosts[:] if chosen_slot is None: return None, None resources_needed = {} resources_available = chosen_slot[2] logger.debug('resources available' + pformat(resources_available)) iter_clusters = cycle(clusters) while req_ram > 0 or req_cpu > 0: cluster = iter_clusters.next() if resources_available[cluster] == 0: clusters.remove(cluster) iter_clusters = cycle(clusters) else: host = cluster + '-1' attr = get_CPU_RAM_FLOPS([host]) resources_available[cluster] -= 1 req_ram -= float(attr[host]['RAM'] / mem) * mem req_cpu -= float(attr[host]['CPU'] / cpu) * cpu if cluster not in resources_needed: resources_needed[cluster] = 0 resources_needed[cluster] += 1 if 'kavlan' in elements: resources_needed['kavlan'] = 1 logger.debug('resources needed' + pformat(resources_needed)) return chosen_slot[0], distribute_hosts(chosen_slot[2], resources_needed, excluded_elements)
def get_planning(elements=['grid5000'], vlan=False, subnet=False, storage=False, out_of_chart=False, starttime=None, endtime=None, ignore_besteffort=True, queues='default'): """Retrieve the planning of the elements (site, cluster) and others resources. Element planning structure is ``{'busy': [(123456,123457), ... ], 'free': [(123457,123460), ... ]}.`` :param elements: a list of Grid'5000 elements ('grid5000', <site>, <cluster>) :param vlan: a boolean to ask for KaVLAN computation :param subnet: a boolean to ask for subnets computation :param storage: a boolean to ask for sorage computation :param out_of_chart: if True, consider that days outside weekends are busy :param starttime: start of time period for which to compute the planning, defaults to now + 1 minute :param endtime: end of time period for which to compute the planning, defaults to 4 weeks from now :param ignore_besteffort: True by default, to consider the resources with besteffort jobs as available :param queues: list of oar queues for which to get the planning Return a dict whose keys are sites, whose values are dict whose keys are cluster, subnets, kavlan or storage, whose values are planning dicts, whose keys are hosts, subnet address range, vlan number or chunk id planning respectively. """ if not starttime: starttime = int(time() + timedelta_to_seconds(timedelta(minutes=1))) starttime = int(get_unixts(starttime)) if not endtime: endtime = int(starttime + timedelta_to_seconds(timedelta(weeks=4, minutes=1))) endtime = int(get_unixts(endtime)) if 'grid5000' in elements: sites = elements = get_g5k_sites() else: sites = list( set([site for site in elements if site in get_g5k_sites()] + [ get_cluster_site(cluster) for cluster in elements if cluster in get_g5k_clusters(queues=queues) ] + [ get_host_site(host) for host in elements if host in get_g5k_hosts() or get_host_shortname(host) in get_g5k_hosts() ])) if len(sites) == 0: logger.error('Wrong elements given: %s' % (elements, )) return None planning = {} for site in sites: planning[site] = {} for cluster in get_site_clusters(site, queues=queues): planning[site][cluster] = {} for site in sites: if vlan: planning[site].update({'vlans': {}}) if subnet: planning[site].update({'subnets': {}}) if storage: planning[site].update({'storage': {}}) if _retrieve_method == 'API': _get_planning_API(planning, ignore_besteffort) elif _retrieve_method == 'PostgreSQL': _get_planning_PGSQL(planning, ignore_besteffort) if out_of_chart: _add_charter_to_planning(planning, starttime, endtime) for site_pl in planning.values(): for res_pl in site_pl.values(): for el_planning in res_pl.values(): el_planning['busy'].sort() _merge_el_planning(el_planning['busy']) _trunc_el_planning(el_planning['busy'], starttime, endtime) _fill_el_planning_free(el_planning, starttime, endtime) # cleaning real_planning = deepcopy(planning) for site, site_pl in planning.items(): for cl, cl_pl in site_pl.items(): if cl in ['vlans']: continue keep_cluster = False for h in cl_pl: if not (get_host_site(h) in elements or get_host_cluster(h) in elements or get_host_shortname(h) in elements or h in elements): del real_planning[site][cl][h] else: keep_cluster = True if not keep_cluster: del real_planning[site][cl] return real_planning
def distribute_hosts(resources_available, resources_wanted, excluded_elements=None, ratio=None): """ Distribute the resources on the different sites and cluster :param resources_available: a dict defining the resources available :param resources_wanted: a dict defining the resources available you really want :param excluded_elements: a list of elements that won't be used :param ratio: if not None (the default), a float between 0 and 1, to actually only use a fraction of the resources.""" if excluded_elements == None: excluded_elements = [] resources = {} #Defining the cluster you want clusters_wanted = {} for element, n_nodes in resources_wanted.items(): if element in get_g5k_clusters(queues=None): clusters_wanted[element] = n_nodes for cluster, n_nodes in clusters_wanted.items(): nodes = n_nodes if n_nodes > 0 else resources_available[cluster] resources_available[get_cluster_site(cluster)] -= nodes resources[cluster] = nodes # Blacklisting clusters for element in excluded_elements: if element in get_g5k_clusters( queues=None) and element in resources_available: resources_available['grid5000'] -= resources_available[element] resources_available[get_cluster_site( element)] -= resources_available[element] resources_available[element] = 0 #Defining the sites you want sites_wanted = {} for element, n_nodes in resources_wanted.items(): if element in get_g5k_sites() and element not in excluded_elements: sites_wanted[element] = n_nodes for site, n_nodes in sites_wanted.items(): resources[site] = n_nodes if n_nodes > 0 else resources_available[site] # Blacklisting sites for element in excluded_elements: if element in get_g5k_sites() and element in resources_available: resources_available['grid5000'] -= resources_available[element] resources_available[element] = 0 #Distributing hosts on grid5000 elements logger.debug(pformat(resources_wanted)) if 'grid5000' in resources_wanted: g5k_nodes = resources_wanted['grid5000'] if resources_wanted[ 'grid5000'] > 0 else resources_available['grid5000'] total_nodes = 0 sites = [ element for element in resources_available if element in get_g5k_sites() ] iter_sites = cycle(sites) while total_nodes < g5k_nodes: site = next(iter_sites) if resources_available[site] == 0: sites.remove(site) iter_sites = cycle(sites) else: resources_available[site] -= 1 if site in resources: resources[site] += 1 else: resources[site] = 1 total_nodes += 1 logger.debug(pformat(resources)) if 'kavlan' in resources_wanted: resources['kavlan'] = resources_available['kavlan'] # apply optional ratio if ratio != None: resources.update( (x, int(floor(y * ratio))) for x, y in resources.items()) return resources
def get_jobs_specs(resources, excluded_elements=None, name=None): """ Generate the several job specifications from the dict of resources and the blacklisted elements :param resources: a dict, whose keys are Grid'5000 element and values the corresponding number of n_nodes :param excluded_elements: a list of elements that won't be used :param name: the name of the jobs that will be given """ jobs_specs = [] if excluded_elements == None: excluded_elements = [] # Creating the list of sites used sites = [] real_resources = resources.copy() for resource in resources: if resource in get_g5k_sites() and resource not in sites: sites.append(resource) if resource in get_g5k_clusters(queues=None): if resource not in excluded_elements: site = get_cluster_site(resource) if site not in sites: sites.append(site) if site not in real_resources: real_resources[site] = 0 # Checking if we need a Kavlan, a KaVLAN global or none get_kavlan = 'kavlan' in resources if get_kavlan: kavlan = 'kavlan' n_sites = 0 for resource in real_resources: if resource in sites: n_sites += 1 if n_sites > 1: kavlan += '-global' break blacklisted_hosts = {} for element in excluded_elements: if element not in get_g5k_clusters(queues=None) + get_g5k_sites(): site = get_host_site(element) if not 'site' in blacklisted_hosts: blacklisted_hosts[site] = [element] else: blacklisted_hosts[site].append(element) for site in sites: sub_resources = '' # Adding a KaVLAN if needed if get_kavlan: if not 'global' in kavlan: sub_resources = "{type='" + kavlan + "'}/vlan=1+" get_kavlan = False elif site in resources['kavlan']: sub_resources = "{type='" + kavlan + "'}/vlan=1+" get_kavlan = False base_sql = '{' end_sql = '}/' # Creating blacklist SQL string for hosts host_blacklist = False str_hosts = '' if site in blacklisted_hosts and len(blacklisted_hosts[site]) > 0: str_hosts = ''.join([ "host not in ('" + get_host_longname(host) + "') and " for host in blacklisted_hosts[site] ]) host_blacklist = True #Adding the clusters blacklist str_clusters = str_hosts if host_blacklist else '' cl_blacklist = False clusters_nodes = 0 for cluster in get_site_clusters(site, queues=None): if cluster in resources and resources[cluster] > 0: if str_hosts == '': sub_resources += "{cluster='" + cluster + "'}" else: sub_resources += base_sql + str_hosts + "cluster='" + \ cluster + "'" + end_sql sub_resources += "/nodes=" + str(resources[cluster]) + '+' clusters_nodes += resources[cluster] if cluster in excluded_elements: str_clusters += "cluster not in ('" + cluster + "') and " cl_blacklist = True # Generating the site blacklist string from host and cluster blacklist str_site = '' if host_blacklist or cl_blacklist: str_site += base_sql if not cl_blacklist: str_site += str_hosts[:-4] else: str_site += str_clusters[:-4] str_site = str_site + end_sql if real_resources[site] > 0: sub_resources += str_site + "nodes=" + str(real_resources[site]) +\ '+' if sub_resources != '': jobs_specs.append((OarSubmission(resources=sub_resources[:-1], name=name), site)) return jobs_specs