def run(n_epochs, lr, train_batch_size, val_batch_size, base_model, clustering_loss_weight, embedding_extractor, annealing_alphas, dataset, val_dataset, result_dir, early_stopping, early_stopping_tol, device, random_state): # Set random states np.random.seed(random_state) torch.manual_seed(random_state) torch.cuda.manual_seed_all(random_state) # load data train_df = pd.read_csv(dataset) train_texts = train_df['texts'].to_numpy() train_labels = train_df['labels'].to_numpy() train_data = TextDataset(train_texts, train_labels) train_data_loader = DataLoader(dataset=train_data, batch_size=train_batch_size, shuffle=False) val_df = pd.read_csv(val_dataset) val_texts = val_df['texts'].to_numpy() val_labels = val_df['labels'].to_numpy() val_data = TextDataset(val_texts, val_labels) val_data_loader = DataLoader(dataset=val_data, batch_size=val_batch_size, shuffle=False) # init lm model & tokenizer lm_model = AutoModelForMaskedLM.from_pretrained(base_model, return_dict=True, output_hidden_states=True) tokenizer = AutoTokenizer.from_pretrained(base_model, return_dict=True, output_hidden_states=True) lm_model.to(device) # init clustering model model, initial_centroids, initial_embeddings = init_model( lm_model=lm_model, tokenizer=tokenizer, data_loader=train_data_loader, embedding_extractor=embedding_extractor, n_clusters=np.unique(train_labels).shape[0], device=device) # init optimizer & scheduler opt = torch.optim.RMSprop( params=model.parameters(), lr=lr, # 2e-5, 5e-7, eps=1e-8) total_steps = len(train_data_loader) * n_epochs scheduler = get_linear_schedule_with_warmup( optimizer=opt, num_warmup_steps=int(len(train_data_loader) * 0.5), num_training_steps=total_steps) # train the model hist = train(n_epochs=n_epochs, model=model, optimizer=opt, scheduler=scheduler, annealing_alphas=annealing_alphas, train_data_loader=train_data_loader, eval_data_loader=val_data_loader, clustering_loss_weight=clustering_loss_weight, early_stopping=early_stopping, early_stopping_tol=early_stopping_tol, verbose=True) # do eval run_results = {} predicted_labels, true_labels = evaluate(model=model, eval_data_loader=val_data_loader, verbose=True) best_matching, accuracy = cluster_accuracy(true_labels, predicted_labels) ari = adjusted_rand_score(true_labels, predicted_labels) nmi = normalized_mutual_info_score(true_labels, predicted_labels) purity = purity_score(y_true=true_labels, y_pred=predicted_labels) run_results['best_matching'] = best_matching run_results['accuracy'] = accuracy run_results['ari'] = ari run_results['nmi'] = nmi run_results[ 'purity'] = purity # use purity to compare with microsoft paper # save train hist os.makedirs(result_dir, exist_ok=True) result_df = pd.DataFrame.from_records([run_results]) result_df.to_csv(os.path.join(result_dir, '20_newsgroups-distilbert.csv'), index=False) # save results & model os.makedirs(result_dir, exist_ok=True) with open(os.path.join(result_dir, 'train_hist.h'), 'wb') as f: pickle.dump(hist, file=f) torch.save(model, os.path.join(result_dir, 'model.bin'))
def run(n_epochs, lr, batch_size, base_model, clustering_loss_weight, embedding_extractor, annealing_alphas, dataset, train_idx_file, result_dir, early_stopping, early_stopping_tol, device, random_state ): # Set random states np.random.seed(random_state) torch.manual_seed(random_state) torch.cuda.manual_seed_all(random_state) # load data df = pd.read_csv(dataset) with open(train_idx_file, 'r') as f: train_idx = np.array(list(map(int, f.readlines()))) df = df.iloc[train_idx].copy() texts = df['texts'].to_numpy() labels = df['labels'].to_numpy() data = TextDataset(texts, labels) data_loader = DataLoader(dataset=data, batch_size=batch_size, shuffle=False) # init lm model & tokenizer lm_model = AutoModelForMaskedLM.from_pretrained(base_model, return_dict=True, output_hidden_states=True) tokenizer = AutoTokenizer.from_pretrained(base_model, return_dict=True, output_hidden_states=True) lm_model.to(device) # init clustering model model, initial_centroids, initial_embeddings = init_model( lm_model=lm_model, tokenizer=tokenizer, data_loader=data_loader, embedding_extractor=embedding_extractor, n_clusters=np.unique(labels).shape[0], device=device ) # init optimizer & scheduler opt = torch.optim.RMSprop( params=model.parameters(), lr=lr, # 2e-5, 5e-7, eps=1e-8 ) total_steps = len(data_loader) * n_epochs scheduler = get_linear_schedule_with_warmup( optimizer=opt, num_warmup_steps=int(len(data_loader) * 0.5), num_training_steps=total_steps ) # train the model hist = train( n_epochs=n_epochs, model=model, optimizer=opt, scheduler=scheduler, annealing_alphas=annealing_alphas, train_data_loader=data_loader, clustering_loss_weight=clustering_loss_weight, early_stopping=early_stopping, early_stopping_tol=early_stopping_tol, verbose=True ) # save results & model os.makedirs(result_dir) with open(os.path.join(result_dir, 'train_hist.h'), 'wb') as f: pickle.dump(hist, file=f) torch.save(model, os.path.join(result_dir, 'model.bin'))
def run(n_epochs, hyperparam_grid, lr, batch_size, val_batch_size, base_model, dataset, train_idx_file, val_idx_file, result_dir, early_stopping, early_stopping_tol, device, random_state): # Set random states np.random.seed(random_state) torch.manual_seed(random_state) torch.cuda.manual_seed_all(random_state) # load data df = pd.read_csv(dataset) with open(train_idx_file, 'r') as f: train_idx = np.array(list(map(int, f.readlines()))) with open(val_idx_file, 'r') as f: val_idx = np.array(list(map(int, f.readlines()))) all_idx = np.append(train_idx, val_idx) df_train = df.iloc[all_idx] train_texts = df_train['texts'].to_numpy() train_labels = df_train['labels'].to_numpy() train_data = TextDataset(train_texts, train_labels) train_data_loader = DataLoader(dataset=train_data, batch_size=batch_size, shuffle=False) df_val = df.iloc[val_idx] val_texts = df_val['texts'].to_numpy() val_labels = df_val['labels'].to_numpy() val_data = TextDataset(val_texts, val_labels) val_data_loader = DataLoader(dataset=val_data, batch_size=val_batch_size, shuffle=False) # insert code here! results = [] param_grid = ParameterGrid(hyperparam_grid) for run_idx, params in enumerate(param_grid): print(f'Run: {run_idx+1}/{len(list(param_grid))}') print("Running with params:") pprint(params) # init lm model & tokenizer lm_model = AutoModelForMaskedLM.from_pretrained( base_model, return_dict=True, output_hidden_states=True) tokenizer = AutoTokenizer.from_pretrained(base_model, return_dict=True, output_hidden_states=True) lm_model.to(device) # init clustering model model, initial_centroids, initial_embeddings = init_model( lm_model=lm_model, tokenizer=tokenizer, data_loader=train_data_loader, embedding_extractor=params['embedding_extractor'], n_clusters=np.unique(train_labels).shape[0], device=device) # init optimizer & scheduler opt = torch.optim.RMSprop( params=model.parameters(), lr=lr, # hier weitermachen eps=1e-8) total_steps = len(train_data_loader) * n_epochs scheduler = get_linear_schedule_with_warmup( optimizer=opt, num_warmup_steps=int(len(train_data_loader) * 0.5), num_training_steps=total_steps) # train the model hist = train(n_epochs=n_epochs, model=model, optimizer=opt, scheduler=scheduler, annealing_alphas=params['annealing_alphas'], train_data_loader=train_data_loader, eval_data_loader=val_data_loader, clustering_loss_weight=params['clustering_loss_weight'], early_stopping=early_stopping, early_stopping_tol=early_stopping_tol, verbose=True) # do eval run_results = { **{f'param_{key}': value for key, value in params.items()} } predicted_labels, true_labels = evaluate( model=model, eval_data_loader=val_data_loader, verbose=True) best_matching, accuracy = cluster_accuracy(true_labels, predicted_labels) ari = adjusted_rand_score(true_labels, predicted_labels) nmi = normalized_mutual_info_score(true_labels, predicted_labels) run_results['best_matching'] = best_matching run_results['accuracy'] = accuracy run_results['ari'] = ari run_results['nmi'] = nmi # save train hist os.makedirs(result_dir, exist_ok=True) results.append(run_results) result_df = pd.DataFrame.from_records(results) result_df.to_csv(os.path.join(result_dir, 'opt_results_ag_news_subset5.csv'), index=False) with open(os.path.join(result_dir, f'train_hist_run{run_idx}.h'), 'wb') as f: pickle.dump(hist, file=f)