示例#1
0
def main(args):
    if args.model == 'LSTM':
        x_eval, y_eval = load_from_folder(args.input)
        params = read_params_json(args.model_path)
        ignored_label = "IGNORE"
        label_map = {
            label: i
            for i, label in enumerate(params['label_list'], 1)
        }
        label_map[ignored_label] = 0
        device = 'cuda:3' if (torch.cuda.is_available()
                              and not args.no_cuda) else 'cpu'
        biLSTM = LSTM(n_labels=params['num_labels'] - 1,
                      embedding_path=args.embedding,
                      hidden_size=1024,
                      dropout=params['dropout'],
                      input_size=args.batch_size * args.max_seq_length)
        state_dict = torch.load(
            open(os.path.join(args.model_path, 'model.pt'), 'rb'))
        biLSTM.load_state_dict(state_dict)
        biLSTM.eval()
        biLSTM.to(device)
        trainer = Trainer()
        f1, report = trainer.evaluate_model(biLSTM, x_eval, y_eval, label_map,
                                            args.batch_size, device,
                                            args.max_seq_length)
        print(" I AM SUPREME ")
        print(report)
        print(f1)
    else:
        params = read_params_json(args.model_path)
        device = 'cuda:3' if (torch.cuda.is_available()
                              and not args.no_cuda) else 'cpu'
        transformers = Transformers()
        transformers.evaluate(pretrained_path=args.pretrained,
                              dropout=params['dropout'],
                              num_labels=params['num_labels'],
                              label_list=params['label_list'],
                              path_model=args.model_path,
                              device=device,
                              eval_batch_size=args.batch_size,
                              max_seq_length=args.max_seq_length,
                              data_path=args.input,
                              model_name=args.model)
        print(" I AM SUPREME ")
示例#2
0
    y_val = torch.from_numpy(y_val).contiguous()

    targets_train = y_train[:, :, :, [0]]
    features_train = y_train[:, :, :, 1:]

    targets_val = y_val[:, :, :, [0]]
    features_val = y_val[:, :, :, 1:]

    targets_test = y_test[:, :, :, [0]]
    features_test = y_test[:, :, :, 1:]

    lstm = LSTM(input_size, hidden_size, output_size, n_layers, dropout)

    if os.path.isfile(checkpoint_file):
        print("Loading checkpoint...")
        lstm.load_state_dict(torch.load(checkpoint_file))

    if use_cuda:
        lstm.cuda()

    # optimizer = optim.Adam(lstm.parameters(), lr=lr)
    #
    # best_val_loss = 1000
    # train_loss = 0
    # for epoch in range(n_epochs):
    #     n_batches = x_train.shape[0]
    #     for i in range(n_batches):
    #         lstm.hidden = None
    #         input_batches = x_train[i]
    #         target_batches = targets_train[i]
    #         train_loss = train(input_batches, target_batches, lstm, optimizer, use_cuda)
示例#3
0
class Model(object):
    def __init__(self, args, device, rel2id, word_emb=None):
        lr = args.lr
        lr_decay = args.lr_decay
        self.cpu = torch.device('cpu')
        self.device = device
        self.args = args
        self.max_grad_norm = args.max_grad_norm
        if args.model == 'pa_lstm':
            self.model = PositionAwareLSTM(args, rel2id, word_emb)
        elif args.model == 'bgru':
            self.model = BGRU(args, rel2id, word_emb)
        elif args.model == 'cnn':
            self.model = CNN(args, rel2id, word_emb)
        elif args.model == 'pcnn':
            self.model = PCNN(args, rel2id, word_emb)
        elif args.model == 'lstm':
            self.model = LSTM(args, rel2id, word_emb)
        else:
            raise ValueError
        self.model.to(device)
        self.criterion = nn.CrossEntropyLoss()
        self.parameters = [
            p for p in self.model.parameters() if p.requires_grad
        ]
        # self.parameters = self.model.parameters()
        self.optimizer = torch.optim.SGD(self.parameters, lr)

    def update(self, batch):
        inputs = [p.to(self.device) for p in batch[:-1]]
        labels = batch[-1].to(self.device)
        self.model.train()
        logits = self.model(inputs)
        loss = self.criterion(logits, labels)
        self.optimizer.zero_grad()
        loss.backward()
        torch.nn.utils.clip_grad_norm_(self.parameters, self.max_grad_norm)
        self.optimizer.step()
        return loss.item()

    def predict(self, batch):
        inputs = [p.to(self.device) for p in batch[:-1]]
        labels = batch[-1].to(self.device)
        logits = self.model(inputs)
        loss = self.criterion(logits, labels)
        pred = torch.argmax(logits, dim=1).to(self.cpu)
        # corrects = torch.eq(pred, labels)
        # acc_cnt = torch.sum(corrects, dim=-1)
        return pred, batch[-1], loss.item()

    def eval(self, dset, vocab=None, output_false_file=None):
        rel_labels = [''] * len(dset.rel2id)
        for label, id in dset.rel2id.items():
            rel_labels[id] = label
        self.model.eval()
        pred = []
        labels = []
        loss = 0.0
        for idx, batch in enumerate(tqdm(dset.batched_data)):
            pred_b, labels_b, loss_b = self.predict(batch)
            pred += pred_b.tolist()
            labels += labels_b.tolist()
            loss += loss_b
            if output_false_file is not None and vocab is not None:
                all_words, pos, ner, subj_pos, obj_pos, labels_ = batch
                all_words = all_words.tolist()
                labels_ = labels_.tolist()
                for i, word_ids in enumerate(all_words):
                    if labels[i] != pred[i]:
                        length = 0
                        for wid in word_ids:
                            if wid != utils.PAD_ID:
                                length += 1
                        words = [vocab[wid] for wid in word_ids[:length]]
                        sentence = ' '.join(words)

                        subj_words = []
                        for sidx in range(length):
                            if subj_pos[i][sidx] == 0:
                                subj_words.append(words[sidx])
                        subj = '_'.join(subj_words)

                        obj_words = []
                        for oidx in range(length):
                            if obj_pos[i][oidx] == 0:
                                obj_words.append(words[oidx])
                        obj = '_'.join(obj_words)

                        output_false_file.write(
                            '%s\t%s\t%s\t%s\t%s\n' %
                            (sentence, subj, obj, rel_labels[pred[i]],
                             rel_labels[labels[i]]))

        loss /= len(dset.batched_data)
        return loss, utils.eval(pred, labels)

    def save(self, filename, epoch):
        params = {
            'model': self.model.state_dict(),
            'config': self.args,
            'epoch': epoch
        }
        try:
            torch.save(params, filename)
            print("model saved to {}".format(filename))
        except BaseException:
            print("[Warning: Saving failed... continuing anyway.]")

    def load(self, filename):
        params = torch.load(filename, map_location=self.device.type)
        self.model.load_state_dict(params['model'])
示例#4
0
class Model(object):
    def __init__(self, args, device, rel2id, word_emb=None):
        lr = args.lr
        lr_decay = args.lr_decay
        self.cpu = torch.device('cpu')
        self.device = device
        self.args = args
        self.rel2id = rel2id
        self.max_grad_norm = args.max_grad_norm
        if args.model == 'pa_lstm':
            self.model = PositionAwareRNN(args, rel2id, word_emb)
        elif args.model == 'bgru':
            self.model = BGRU(args, rel2id, word_emb)
        elif args.model == 'cnn':
            self.model = CNN(args, rel2id, word_emb)
        elif args.model == 'pcnn':
            self.model = PCNN(args, rel2id, word_emb)
        elif args.model == 'lstm':
            self.model = LSTM(args, rel2id, word_emb)
        else:
            raise ValueError
        self.model.to(device)
        self.criterion = nn.CrossEntropyLoss()
        if args.fix_bias:
            self.model.flinear.bias.requires_grad = False
        self.parameters = [
            p for p in self.model.parameters() if p.requires_grad
        ]
        # self.parameters = self.model.parameters()
        self.optimizer = torch.optim.SGD(self.parameters, lr)
        self.scheduler = lr_scheduler.ReduceLROnPlateau(self.optimizer,
                                                        'min',
                                                        patience=3,
                                                        factor=lr_decay)

    def update_lr(self, valid_loss):
        self.scheduler.step(valid_loss)

    def update(self, batch, penalty=False, weight=1.0):
        inputs = [p.to(self.device) for p in batch[:5]]
        labels = batch[5].to(self.device)
        self.model.train()
        logits = self.model(inputs)
        loss = self.criterion(logits, labels)
        # batch_ent = utils.calcEntropy(logits)
        # ent = torch.sum(batch_ent) / len(batch_ent)
        # if penalty:
        # 	loss = loss - ent*weight
        self.optimizer.zero_grad()
        loss.backward()
        torch.nn.utils.clip_grad_norm_(self.parameters, self.max_grad_norm)
        self.optimizer.step()
        return loss.item()

    def get_bias(self):
        return self.model.flinear.bias.data

    def set_bias(self, bias):
        self.model.flinear.bias.data = bias

    def predict(self, batch):
        inputs = [p.to(self.device) for p in batch[:5]]
        labels = batch[5].to(self.cpu)
        orig_idx = batch[6]
        logits = self.model(inputs).to(self.cpu)
        loss = self.criterion(logits, labels)
        pred = torch.argmax(logits, dim=1).to(self.cpu)
        # corrects = torch.eq(pred, labels)
        # acc_cnt = torch.sum(corrects, dim=-1)
        recover_idx = utils.recover_idx(orig_idx)
        logits = [logits[idx].tolist() for idx in recover_idx]
        pred = [pred[idx].item() for idx in recover_idx]
        labels = [labels[idx].item() for idx in recover_idx]
        return logits, pred, labels, loss.item()

    def eval(self,
             dset,
             vocab=None,
             output_false_file=None,
             output_label_file=None,
             weights=None):
        if weights is None:
            weights = [1.0] * len(dset.rel2id)

        rel_labels = [''] * len(dset.rel2id)
        for label, id in dset.rel2id.items():
            rel_labels[id] = label
        self.model.eval()
        pred = []
        labels = []
        loss = 0.0

        for idx, batch in enumerate(dset.batched_data):
            scores_b, pred_b, labels_b, loss_b = self.predict(batch)
            pred += pred_b
            labels += labels_b
            loss += loss_b

            if output_false_file is not None and vocab is not None:
                all_words, pos, ner, subj_pos, obj_pos, labels_, _ = batch
                all_words = all_words.tolist()
                output_false_file.write('\n')
                for i, word_ids in enumerate(all_words):
                    if labels[i] != pred[i]:
                        length = 0
                        for wid in word_ids:
                            if wid != utils.PAD_ID:
                                length += 1
                        words = [vocab[wid] for wid in word_ids[:length]]
                        sentence = ' '.join(words)

                        subj_words = []
                        for sidx in range(length):
                            if subj_pos[i][sidx] == 0:
                                subj_words.append(words[sidx])
                        subj = '_'.join(subj_words)

                        obj_words = []
                        for oidx in range(length):
                            if obj_pos[i][oidx] == 0:
                                obj_words.append(words[oidx])
                        obj = '_'.join(obj_words)

                        output_false_file.write(
                            '%s\t%s\t%s\t%s\t%s\n' %
                            (sentence, subj, obj, rel_labels[pred[i]],
                             rel_labels[labels[i]]))

        if output_label_file is not None and vocab is not None:
            output_label_file.write(json.dumps(pred) + '\n')
            output_label_file.write(json.dumps(labels) + '\n')

        loss /= len(dset.batched_data)
        return loss, utils.eval(pred, labels, weights)

    def TuneEntropyThres(self,
                         test_dset,
                         noneInd=utils.NO_RELATION,
                         ratio=0.2,
                         cvnum=100):
        '''
		Tune threshold on test set
		'''
        rel_labels = [''] * len(test_dset.rel2id)
        for label, id in test_dset.rel2id.items():
            rel_labels[id] = label
        self.model.eval()
        pred = []
        labels = []
        scores = []
        loss = 0.0
        for idx, batch in enumerate(test_dset.batched_data):
            scores_b, pred_b, labels_b, loss_b = self.predict(batch)
            pred += pred_b
            labels += labels_b
            scores += scores_b
            loss += loss_b
        loss /= len(test_dset.batched_data)

        # start tuning
        scores = torch.tensor(scores)
        f1score = 0.0
        recall = 0.0
        precision = 0.0

        pre_ind = utils.calcInd(scores)
        pre_entropy = utils.calcEntropy(scores)
        valSize = int(np.floor(ratio * len(pre_ind)))
        data = [[pre_ind[ind], pre_entropy[ind], labels[ind]]
                for ind in range(0, len(pre_ind))]

        for cvind in tqdm(range(cvnum)):
            random.shuffle(data)
            val = data[0:valSize]
            eva = data[valSize:]

            # find best threshold
            max_ent = max(val, key=lambda t: t[1])[1]
            min_ent = min(val, key=lambda t: t[1])[1]
            stepSize = (max_ent - min_ent) / 100
            thresholdList = [min_ent + ind * stepSize for ind in range(0, 100)]
            ofInterest = 0
            for ins in val:
                if ins[2] != noneInd:
                    ofInterest += 1
            bestThreshold = float('nan')
            bestF1 = float('-inf')
            for threshold in thresholdList:
                corrected = 0
                predicted = 0
                for ins in val:
                    if ins[1] < threshold and ins[0] != noneInd:
                        predicted += 1
                        if ins[0] == ins[2]:
                            corrected += 1
                curF1 = 2.0 * corrected / (ofInterest + predicted)
                if curF1 > bestF1:
                    bestF1 = curF1
                    bestThreshold = threshold
            ofInterest = 0
            corrected = 0
            predicted = 0
            for ins in eva:
                if ins[2] != noneInd:
                    ofInterest += 1
                if ins[1] < bestThreshold and ins[0] != noneInd:
                    predicted += 1
                    if ins[0] == ins[2]:
                        corrected += 1

            f1score += (2.0 * corrected / (ofInterest + predicted))
            recall += (1.0 * corrected / ofInterest)
            precision += (1.0 * corrected / (predicted + 0.00001))

        f1score /= cvnum
        recall /= cvnum
        precision /= cvnum

        return loss, f1score, recall, precision

    def TuneMaxThres(self,
                     test_dset,
                     noneInd=utils.NO_RELATION,
                     ratio=0.2,
                     cvnum=100):
        '''
		Tune threshold on test set
		'''
        rel_labels = [''] * len(test_dset.rel2id)
        for label, id in test_dset.rel2id.items():
            rel_labels[id] = label
        self.model.eval()
        pred = []
        labels = []
        scores = []
        loss = 0.0
        for idx, batch in enumerate(test_dset.batched_data):
            scores_b, pred_b, labels_b, loss_b = self.predict(batch)
            pred += pred_b
            labels += labels_b
            scores += scores_b
            loss += loss_b
        loss /= len(test_dset.batched_data)

        # start tuning
        scores = torch.tensor(scores)
        f1score = 0.0
        recall = 0.0
        precision = 0.0

        pre_prob, pre_ind = torch.max(scores, 1)
        valSize = int(np.floor(ratio * len(pre_ind)))
        data = [[pre_ind[ind], pre_prob[ind], labels[ind]]
                for ind in range(0, len(pre_ind))]
        for cvind in tqdm(range(cvnum)):
            random.shuffle(data)
            val = data[0:valSize]
            eva = data[valSize:]

            # find best threshold
            max_ent = max(val, key=lambda t: t[1])[1]
            min_ent = min(val, key=lambda t: t[1])[1]
            stepSize = (max_ent - min_ent) / 100
            thresholdList = [min_ent + ind * stepSize for ind in range(0, 100)]
            ofInterest = 0
            for ins in val:
                if ins[2] != noneInd:
                    ofInterest += 1
            bestThreshold = float('nan')
            bestF1 = float('-inf')
            for threshold in thresholdList:
                corrected = 0
                predicted = 0
                for ins in val:
                    if ins[1] > threshold and ins[0] != noneInd:
                        predicted += 1
                        if ins[0] == ins[2]:
                            corrected += 1
                curF1 = 2.0 * corrected / (ofInterest + predicted)
                if curF1 > bestF1:
                    bestF1 = curF1
                    bestThreshold = threshold

            ofInterest = 0
            corrected = 0
            predicted = 0
            for ins in eva:
                if ins[2] != noneInd:
                    ofInterest += 1
                if ins[1] > bestThreshold and ins[0] != noneInd:
                    predicted += 1
                    if ins[0] == ins[2]:
                        corrected += 1
            f1score += (2.0 * corrected / (ofInterest + predicted))
            recall += (1.0 * corrected / ofInterest)
            precision += (1.0 * corrected / (predicted + 0.00001))

        f1score /= cvnum
        recall /= cvnum
        precision /= cvnum

        return loss, f1score, recall, precision

    def save(self, filename, epoch):
        params = {
            'model': self.model.state_dict(),
            'config': self.args,
            'epoch': epoch
        }
        try:
            torch.save(params, filename)
            print("Epoch {}, model saved to {}".format(epoch, filename))
        except BaseException:
            print("[Warning: Saving failed... continuing anyway.]")
        # json.dump(vars(self.args), open('%s.json' % filename, 'w'))

    def count_parameters(self):
        return sum(p.numel() for p in self.model.parameters()
                   if p.requires_grad)

    def load(self, filename):
        params = torch.load(filename)
        if type(params).__name__ == 'dict' and 'model' in params:
            self.model.load_state_dict(params['model'])
        else:
            self.model.load_state_dict(params)