def train( main_config, model_config, model_name, experiment_name, dataset_name, ): main_cfg = MainConfig(main_config) model = MODELS[model_name] dataset = dataset_type.get_dataset(dataset_name) train_data = dataset.train_set_pairs() vectorizer = DatasetVectorizer(main_cfg.model_dir, raw_sentence_pairs=train_data) dataset_helper = Dataset(vectorizer, dataset, main_cfg.batch_size) max_sentence_len = vectorizer.max_sentence_len vocabulary_size = vectorizer.vocabulary_size train_mini_sen1, train_mini_sen2, train_mini_labels = dataset_helper.pick_train_mini_batch( ) train_mini_labels = train_mini_labels.reshape(-1, 1) test_sentence1, test_sentence2 = dataset_helper.test_instances() test_labels = dataset_helper.test_labels() test_labels = test_labels.reshape(-1, 1) num_batches = dataset_helper.num_batches model = model( max_sentence_len, vocabulary_size, main_config, model_config, ) model_saver = ModelSaver( main_cfg.model_dir, experiment_name, main_cfg.checkpoints_to_keep, ) config = tf.ConfigProto( allow_soft_placement=True, log_device_placement=main_cfg.log_device_placement, ) with tf.Session(config=config) as session: global_step = 0 init = tf.global_variables_initializer() session.run(init) log_saver = LogSaver( main_cfg.logs_path, experiment_name, dataset_name, session.graph, ) model_evaluator = ModelEvaluator(model, session) metrics = {'acc': 0.0} time_per_epoch = [] log('Training model for {} epochs'.format(main_cfg.num_epochs)) for epoch in tqdm(range(main_cfg.num_epochs), desc='Epochs'): start_time = time.time() train_sentence1, train_sentence2 = dataset_helper.train_instances( shuffle=True) train_labels = dataset_helper.train_labels() train_batch_helper = BatchHelper( train_sentence1, train_sentence2, train_labels, main_cfg.batch_size, ) # small eval set for measuring dev accuracy dev_sentence1, dev_sentence2, dev_labels = dataset_helper.dev_instances( ) dev_labels = dev_labels.reshape(-1, 1) tqdm_iter = tqdm(range(num_batches), total=num_batches, desc="Batches", leave=False, postfix=metrics) for batch in tqdm_iter: global_step += 1 sentence1_batch, sentence2_batch, labels_batch = train_batch_helper.next( batch) feed_dict_train = { model.x1: sentence1_batch, model.x2: sentence2_batch, model.is_training: True, model.labels: labels_batch, } loss, _ = session.run([model.loss, model.opt], feed_dict=feed_dict_train) if batch % main_cfg.eval_every == 0: feed_dict_train = { model.x1: train_mini_sen1, model.x2: train_mini_sen2, model.is_training: False, model.labels: train_mini_labels, } train_accuracy, train_summary = session.run( [model.accuracy, model.summary_op], feed_dict=feed_dict_train, ) log_saver.log_train(train_summary, global_step) feed_dict_dev = { model.x1: dev_sentence1, model.x2: dev_sentence2, model.is_training: False, model.labels: dev_labels } dev_accuracy, dev_summary = session.run( [model.accuracy, model.summary_op], feed_dict=feed_dict_dev, ) log_saver.log_dev(dev_summary, global_step) tqdm_iter.set_postfix( dev_acc='{:.2f}'.format(float(dev_accuracy)), train_acc='{:.2f}'.format(float(train_accuracy)), loss='{:.2f}'.format(float(loss)), epoch=epoch) if global_step % main_cfg.save_every == 0: model_saver.save(session, global_step=global_step) model_evaluator.evaluate_dev(dev_sentence1, dev_sentence2, dev_labels) end_time = time.time() total_time = timer(start_time, end_time) time_per_epoch.append(total_time) model_saver.save(session, global_step=global_step) model_evaluator.evaluate_test(test_sentence1, test_sentence2, test_labels) model_evaluator.save_evaluation( '{}/{}'.format(main_cfg.model_dir, experiment_name), time_per_epoch[-1], dataset)
def train(main_config, model_config, model_name, dataset_name): main_cfg = MainConfig(main_config) model = MODELS[model_name] dataset = DATASETS[dataset_name]() model_name = '{}_{}'.format(model_name, main_config['PARAMS']['embedding_size']) train_data = dataset.train_set_pairs() vectorizer = DatasetVectorizer(train_data, main_cfg.model_dir) dataset_helper = Dataset(vectorizer, dataset, main_cfg.batch_size) max_sentence_len = vectorizer.max_sentence_len vocabulary_size = vectorizer.vocabulary_size train_mini_sen1, train_mini_sen2, train_mini_labels = dataset_helper.pick_train_mini_batch( ) train_mini_labels = train_mini_labels.reshape(-1, 1) test_sentence1, test_sentence2 = dataset_helper.test_instances() test_labels = dataset_helper.test_labels() test_labels = test_labels.reshape(-1, 1) num_batches = dataset_helper.num_batches model = model(max_sentence_len, vocabulary_size, main_config, model_config) model_saver = ModelSaver(main_cfg.model_dir, model_name, main_cfg.checkpoints_to_keep) config = tf.ConfigProto(allow_soft_placement=True, log_device_placement=main_cfg.log_device_placement) with tf.Session(config=config) as session: global_step = 0 init = tf.global_variables_initializer() session.run(init) log_saver = LogSaver(main_cfg.logs_path, model_name, dataset_name, session.graph) model_evaluator = ModelEvaluator(model, session) metrics = {'acc': 0.0} time_per_epoch = [] for epoch in tqdm(range(main_cfg.num_epochs), desc='Epochs'): start_time = time.time() train_sentence1, train_sentence2 = dataset_helper.train_instances( shuffle=True) train_labels = dataset_helper.train_labels() train_batch_helper = BatchHelper(train_sentence1, train_sentence2, train_labels, main_cfg.batch_size) # small eval set for measuring dev accuracy dev_sentence1, dev_sentence2, dev_labels = dataset_helper.dev_instances( ) dev_labels = dev_labels.reshape(-1, 1) tqdm_iter = tqdm(range(num_batches), total=num_batches, desc="Batches", leave=False, postfix=metrics) for batch in tqdm_iter: global_step += 1 sentence1_batch, sentence2_batch, labels_batch = train_batch_helper.next( batch) feed_dict_train = { model.x1: sentence1_batch, model.x2: sentence2_batch, model.is_training: True, model.labels: labels_batch } loss, _ = session.run([model.loss, model.opt], feed_dict=feed_dict_train) if batch % main_cfg.eval_every == 0: feed_dict_train = { model.x1: train_mini_sen1, model.x2: train_mini_sen2, model.is_training: False, model.labels: train_mini_labels } train_accuracy, train_summary = session.run( [model.accuracy, model.summary_op], feed_dict=feed_dict_train) log_saver.log_train(train_summary, global_step) feed_dict_dev = { model.x1: dev_sentence1, model.x2: dev_sentence2, model.is_training: False, model.labels: dev_labels } dev_accuracy, dev_summary = session.run( [model.accuracy, model.summary_op], feed_dict=feed_dict_dev) log_saver.log_dev(dev_summary, global_step) tqdm_iter.set_postfix( dev_acc='{:.2f}'.format(float(dev_accuracy)), train_acc='{:.2f}'.format(float(train_accuracy)), loss='{:.2f}'.format(float(loss)), epoch=epoch) if global_step % main_cfg.save_every == 0: model_saver.save(session, global_step=global_step) model_evaluator.evaluate_dev(dev_sentence1, dev_sentence2, dev_labels) end_time = time.time() total_time = timer(start_time, end_time) time_per_epoch.append(total_time) model_saver.save(session, global_step=global_step) feed_dict_train = { model.x1: test_sentence1, model.x2: test_sentence2, model.is_training: False, model.labels: test_labels } #train_accuracy, train_summary, train_e = session.run([model.accuracy, model.summary_op, model.e], # feed_dict=feed_dict_train) train_e = session.run([model.e], feed_dict=feed_dict_train) plt.clf() f = plt.figure(figsize=(8, 8.5)) ax = f.add_subplot(1, 1, 1) i = ax.imshow(train_e[0][0], interpolation='nearest', cmap='gray') cbaxes = f.add_axes([0.2, 0, 0.6, 0.03]) cbar = f.colorbar(i, cax=cbaxes, orientation='horizontal') cbar.ax.set_xlabel('Probability', labelpad=2) f.savefig('attention_maps.pdf', bbox_inches='tight') f.show() plt.show() feed_dict_test = { model.x1: test_sentence1, model.x2: test_sentence2, model.is_training: False, model.labels: test_labels } test_accuracy, test_summary = session.run( [model.accuracy, model.summary_op], feed_dict=feed_dict_test) print('tst_acc:%.2f loss:%.2f', test_accuracy, loss) model_evaluator.evaluate_test(test_sentence1, test_sentence2, test_labels) model_evaluator.save_evaluation( '{}/{}'.format(main_cfg.model_dir, model_name), time_per_epoch[-1], dataset)