def random_search(data_and_labels: tuple, model: TransformerAutoEncoder, params: Dict) -> Dict: x_train, x_test, _, y_test = data_and_labels scores = [] for conf in zip(*params.values()): kwargs = {k: val for k, val in zip(params.keys(), conf)} model.set_params(**kwargs) print(f'Model current hyperparameters are: {kwargs}.') model.fit(x_train) y_pred = model.predict(x_test) # return reconstruction errors theta, f1 = find_optimal_threshold(y_test, y_pred) y_pred = classify(y_pred, theta) metrics_report(y_test, y_pred) scores.append( create_experiment_report(get_metrics(y_test, y_pred), kwargs)) # visualize_distribution_with_labels(y_pred, y_test, to_file=False) from sklearn.metrics import confusion_matrix print(confusion_matrix(y_test, y_pred)) create_checkpoint({'experiments': scores}, EXPERIMENT_PATH) return {'experiments': scores}
def random_search_unsupervised(data_and_labels: tuple, model: Union[LocalOutlierFactor, IsolationForest], params: Dict) -> Dict: x_train, x_test, _, y_test = data_and_labels scores = [] for experiment in params['experiments']: model.set_params(**experiment['hyperparameters']) print(f'Model current hyperparameters are: {experiment["hyperparameters"]}.') if isinstance(model, LocalOutlierFactor): y_pred = model.fit_predict(x_test) # return labels else: model.fit(x_train) y_pred = model.predict(x_test) # return labels y_pred = convert_predictions(y_pred) metrics_report(y_test, y_pred) model_path = create_model_path(DIR_TO_EXPERIMENTS, str(uuid.uuid4())) torch.save(model, model_path) res = create_experiment_report(get_metrics(y_test, y_pred), experiment['hyperparameters'], file_path=model_path) scores.append(res) create_checkpoint({'experiments': scores}, EXPERIMENT_PATH) return { 'experiments': scores }
def random_search(data_and_labels: tuple, model: Union[AutoEncoder, VanillaTCN, AETCN, AECNN1D, CNN1D, CNN2D, TCNCNN1D, SACNN1D, SACNN2D], params: Dict) -> Dict: x_train, x_test, _, y_test = data_and_labels scores = [] for experiment in params['experiments']: model.set_params(**experiment['hyperparameters']) print(f'Model current hyperparameters are: {experiment["hyperparameters"]}.') model.fit(x_train) y_pred = model.predict(x_test) # return reconstruction errors theta, f1 = find_optimal_threshold(y_test, y_pred) y_pred = classify(y_pred, theta) metrics_report(y_test, y_pred) model_path = create_model_path(DIR_TO_EXPERIMENTS, str(uuid.uuid4())) torch.save(model, model_path) res = create_experiment_report(get_metrics(y_test, y_pred), experiment['hyperparameters'], theta, model_path) scores.append(res) create_checkpoint({'experiments': scores}, EXPERIMENT_PATH) return { 'experiments': scores }
def random_search(data_and_labels: tuple, model: Union[AutoEncoder, IsolationForest], params: Dict) -> Dict: x_train, x_test, _, y_test = data_and_labels scores = [] for conf in zip(*params.values()): kwargs = {k: val for k, val in zip(params.keys(), conf)} model.set_params(**kwargs) print(f'Model current hyperparameters are: {kwargs}.') model.fit(x_train) y_pred = model.predict(x_test) # return reconstruction errors theta, f1 = find_optimal_threshold(y_test, y_pred) y_pred = classify(y_pred, theta) metrics_report(y_test, y_pred) scores.append( create_experiment_report(get_metrics(y_test, y_pred), kwargs)) create_checkpoint({'experiments': scores}, EXPERIMENT_PATH) return {'experiments': scores}
def train_window(x_train: List, x_test: List, y_train: np.array, y_test: np.array) -> Dict: sc = CustomMinMaxScaler() x_train = sc.fit_transform(x_train) x_test = sc.transform(x_test) scores = [] for w in range(1, 50, 2): print('Window:', w) model = VanillaTCN(epochs=1, window=w) model.fit(x_train[y_train == 0]) y_pred = model.predict(x_test) # return reconstruction errors theta, f1 = find_optimal_threshold(y_test, y_pred) y_pred = classify(y_pred, theta) metrics_report(y_test, y_pred) scores.append( create_experiment_report(get_metrics(y_test, y_pred), {'window': w})) create_checkpoint( {'experiments': scores}, '../../models/TCN-cropped-window-embeddings-HDFS1.json') return {'experiments': scores}