Exemple #1
0
    def __getitem__(self, idx):
        id = self.idxs[idx]

        if self.model_type == 'moco':
            arch = self.total_archs[self.total_keys[id]][0][0]
            ops = self.total_archs[self.total_keys[id]][0][1]
            path_encoding = self.total_archs[self.total_keys[id]][-1]
            return arch, ops, path_encoding
        elif self.model_type == 'SS_CCL':
            arch = self.total_archs[self.total_keys[id]][0][0]
            ops = self.total_archs[self.total_keys[id]][0][1]
            path_encoding = self.total_archs[self.total_keys[id]][-1]

            edge_index, node_f = nasbench2graph_201((arch, ops), is_idx=True)
            g_d = Data(edge_index=edge_index.long(), x=node_f.float())
            return g_d, path_encoding
        else:
            raise NotImplementedError(
                f'The model type {self.model_type} does not support!')
Exemple #2
0
def gin_predictor_nasbench_201(search_space,
                               num_init=10,
                               k=10,
                               total_queries=150,
                               acq_opt_type='mutation',
                               allow_isomorphisms=False,
                               verbose=1,
                               agent=None,
                               logger=None,
                               gpu='0',
                               lr=0.01,
                               candidate_nums=100,
                               epochs=1000,
                               rate=10,
                               algo_name=None):
    """
    Bayesian optimization with a neural network model
    """
    device = torch.device('cuda:%d' % gpu)
    data = search_space.generate_random_dataset(
        num=num_init,
        allow_isomorphisms=allow_isomorphisms,
        deterministic_loss=True)
    query = num_init + k
    search_agent = agent
    if len(data) <= 10:
        batch_size = 10
    else:
        batch_size = 16
    while query <= total_queries:
        arch_data = [d[0] for d in data]
        agent = NasBenchGinPredictorTrainer(search_agent,
                                            lr=lr,
                                            device=device,
                                            epochs=epochs,
                                            train_images=len(data),
                                            batch_size=batch_size,
                                            input_dim=8,
                                            rate=rate,
                                            algo_name=algo_name)
        val_accuracy = np.array([d[4] for d in data])
        arch_data_edge_idx_list = []
        arch_data_node_f_list = []
        for arch in arch_data:
            edge_index, node_f = nasbench2graph_201(arch)
            arch_data_edge_idx_list.append(edge_index)
            arch_data_node_f_list.append(node_f)
        candidates = search_space.get_candidates(
            data, num=candidate_nums, allow_isomorphisms=allow_isomorphisms)
        candiate_edge_list = []
        candiate_node_list = []
        for cand in candidates:
            edge_index, node_f = nasbench2graph_201(cand[0])
            candiate_edge_list.append(edge_index)
            candiate_node_list.append(node_f)
        agent.fit(arch_data_edge_idx_list,
                  arch_data_node_f_list,
                  val_accuracy,
                  logger=None)
        acc_train = agent.pred(arch_data_edge_idx_list, arch_data_node_f_list)
        acc_pred = agent.pred(candiate_edge_list, candiate_node_list)
        candidate_np = acc_pred.cpu().numpy()
        sorted_indices = np.argsort(candidate_np)
        for i in sorted_indices[:k]:
            archtuple = candidates[i]
            data.append(archtuple)
        if verbose:
            top_5_loss = sorted([d[4] for d in data])[:min(5, len(data))]
            logger.info('Query {}, training mean loss is  {}'.format(
                query,
                np.mean(np.abs(acc_train.cpu().numpy() - val_accuracy))))
            logger.info('Query {}, top 5 val losses {}'.format(
                query, top_5_loss))
        query += k
    return data
Exemple #3
0
def gin_unsupervised_predictor(search_space,
                               model_dir=None,
                               num_init=10,
                               k=10,
                               total_queries=150,
                               acq_opt_type='mutation',
                               allow_isomorphisms=False,
                               verbose=1,
                               agent=None,
                               logger=None,
                               gpu='0',
                               lr=0.01,
                               candidate_nums=100,
                               epochs=1000,
                               predictor_type=None,
                               algo_name=None,
                               benchmark=None,
                               rate=10):
    assert benchmark is not None, 'The benchmark have to be nasbench_101 or nasbench_201.'
    device = torch.device('cuda:%d' % gpu)
    data = search_space.generate_random_dataset(
        num=num_init,
        allow_isomorphisms=allow_isomorphisms,
        deterministic_loss=True)
    query = num_init + k
    search_agent = agent
    input_dim = 6 if benchmark == 'nasbench_101' else 8
    if len(data) <= 10:
        batch_size = 10
    else:
        batch_size = 16
    while query <= total_queries:
        arch_data = [d[0] for d in data]
        agent = NasBenchGinPredictorTrainer(search_agent,
                                            lr=lr,
                                            device=device,
                                            epochs=epochs,
                                            train_images=len(data),
                                            batch_size=batch_size,
                                            input_dim=input_dim,
                                            model_dir=model_dir,
                                            predictor_type=predictor_type,
                                            logger=logger,
                                            algo_name=algo_name,
                                            rate=rate)
        val_accuracy = np.array([d[4] for d in data])
        arch_data_edge_idx_list = []
        arch_data_node_f_list = []
        for arch in arch_data:
            edge_index, node_f = nasbench2graph_101(
                arch) if benchmark == 'nasbench_101' else nasbench2graph_201(
                    arch)
            arch_data_edge_idx_list.append(edge_index)
            arch_data_node_f_list.append(node_f)
        candidates = search_space.get_candidates(
            data, num=candidate_nums, allow_isomorphisms=allow_isomorphisms)
        candiate_edge_list = []
        candiate_node_list = []
        for cand in candidates:
            edge_index, node_f = nasbench2graph_101(
                cand[0]
            ) if benchmark == 'nasbench_101' else nasbench2graph_201(cand[0])
            candiate_edge_list.append(edge_index)
            candiate_node_list.append(node_f)
        agent.fit(arch_data_edge_idx_list,
                  arch_data_node_f_list,
                  val_accuracy,
                  logger=None)
        acc_train = agent.pred(arch_data_edge_idx_list, arch_data_node_f_list)
        acc_pred = agent.pred(candiate_edge_list, candiate_node_list)
        candidate_np = acc_pred.cpu().numpy()
        sorted_indices = np.argsort(candidate_np)
        for i in sorted_indices[:k]:
            if benchmark == 'nasbench_101':
                archtuple = search_space.query_arch(matrix=candidates[i][1],
                                                    ops=candidates[i][2])
            elif benchmark == 'nasbench_201':
                archtuple = candidates[i]
            else:
                raise NotImplementedError()
            data.append(archtuple)
        if verbose:
            top_5_loss = sorted([d[4] for d in data])[:min(5, len(data))]
            logger.info('Query {}, training mean loss is  {}'.format(
                query,
                np.mean(np.abs(acc_train.cpu().numpy() - val_accuracy))))
            logger.info('Query {}, top 5 val losses {}'.format(
                query, top_5_loss))
        query += k
    # data = [list(dd) for dd in data]
    # for i, d in enumerate(data):
    #     data[i][4] = (1 - d[4]) * 100
    #     data[i][5] = (1 - d[5]) * 100
    return data