コード例 #1
0
def train(args, labeled, resume_from, ckpt_file):
    batch_size = args["batch_size"]
    lr = 4.0
    momentum = 0.9
    epochs = args["train_epochs"]

    if not os.path.isdir('./.data'):
        os.mkdir('./.data')

    global train_dataset, test_dataset
    train_dataset, test_dataset = text_classification.DATASETS['AG_NEWS'](
        root='./.data', ngrams=args["N_GRAMS"], vocab=None)

    global VOCAB_SIZE, EMBED_DIM, NUN_CLASS
    VOCAB_SIZE = len(train_dataset.get_vocab())
    EMBED_DIM = args["EMBED_DIM"]
    NUN_CLASS = len(train_dataset.get_labels())

    trainloader = DataLoader(train_dataset,
                             batch_size=batch_size,
                             shuffle=False,
                             collate_fn=generate_batch)
    net = TextSentiment(VOCAB_SIZE, EMBED_DIM, NUN_CLASS).to(device)
    criterion = nn.CrossEntropyLoss().to(device)
    optimizer = optim.SGD(net.parameters(), lr=lr)
    scheduler = torch.optim.lr_scheduler.StepLR(optimizer, 1, gamma=0.9)

    if resume_from is not None:
        ckpt = torch.load(os.path.join(args["EXPT_DIR"], resume_from))
        net.load_state_dict(ckpt["model"])
        optimizer.load_state_dict(ckpt["optimizer"])
    else:
        getdatasetstate()

    net.train()
    for epoch in tqdm(range(epochs), desc="Training"):
        running_loss = 0.0
        train_acc = 0
        for i, data in enumerate(trainloader):
            text, offsets, cls = data
            text, offsets, cls = text.to(device), offsets.to(device), cls.to(
                device)
            outputs = net(text, offsets)
            loss = criterion(outputs, cls)

            optimizer.zero_grad()
            loss.backward()
            optimizer.step()

            train_acc += (outputs.argmax(1) == cls).sum().item()
            running_loss += loss.item()
        scheduler.step()

    print("Finished Training. Saving the model as {}".format(ckpt_file))
    print("Training accuracy: {}".format(
        (train_acc / len(train_dataset) * 100)))
    ckpt = {"model": net.state_dict(), "optimizer": optimizer.state_dict()}
    torch.save(ckpt, os.path.join(args["EXPT_DIR"], ckpt_file))

    return
コード例 #2
0
def infer(sample):
    train_dataset, test_dataset, mytrainloader, mytestloader = get_loaders()
    classes = ("World", "Sports", "Business", "Sci/Tec")

    VOCAB_SIZE = len(train_dataset.get_vocab())
    EMBED_DIM = 32
    NUM_CLASS = len(train_dataset.get_labels())
    mynet = TextSentiment(VOCAB_SIZE, EMBED_DIM, NUM_CLASS).to(device)
    mycriterion = nn.CrossEntropyLoss().to(device)
    myoptimizer = optim.SGD(mynet.parameters(), lr=4.0)
    myscheduler = torch.optim.lr_scheduler.StepLR(myoptimizer, 1, gamma=0.9)

    sampler = SubsetSampler(sample)
    dataloader = torch.utils.data.DataLoader(
        train_dataset,
        batch_size=1,
        num_workers=4,
        sampler=sampler,
        collate_fn=generate_batch,
    )
    soft = torch.nn.Softmax(dim=0)
    results = []
    infer_outs = {}
    with torch.no_grad():
        with tqdm(total=len(dataloader),
                  desc="Inferring on unlabeled ...") as tq:
            for r, (text, offsets, cls) in enumerate(dataloader):
                text, offsets, cls = text.to(device), offsets.to(
                    device), cls.to(device)
                outputs = mynet(text, offsets)
                _, predicted = torch.max(outputs.data, 1)
                ground_truth = cls.item()
                prediction = predicted.item()
                infer_outs[r] = soft(outputs[0]).numpy().tolist()
                tq.update(1)
            # results.append([sample[r], classes[ground_truth], classes[prediction], probability[prediction],classwiseprobs])

    return infer_outs
コード例 #3
0
def main():

    device = "gpu" if torch.cuda.is_available() else "cpu"
    train_dataset, test_dataset = get_dataset()
    VOCAB_SIZE = len(train_dataset.get_vocab())
    EMBED_DIM = 32
    NUN_CLASS = len(train_dataset.get_labels())
    model = TextSentiment(VOCAB_SIZE, EMBED_DIM, NUN_CLASS).to(device)
    BATCH_SIZE = 16
    N_EPOCHS = 5
    min_valid_loss = float('inf')

    criterion = torch.nn.CrossEntropyLoss().to(
        device)  # mutil-class use the CrossEntropy
    optimizer = torch.optim.SGD(model.parameters(), lr=4.0)
    scheduler = torch.optim.lr_scheduler.StepLR(optimizer, 1, gamma=0.9)

    train_len = int(len(train_dataset) * 0.95)
    sub_train_, sub_valid_ = \
        random_split(train_dataset, [train_len, len(train_dataset) - train_len])
    train_loader = DataLoader(sub_train_,
                              batch_size=BATCH_SIZE,
                              shuffle=True,
                              collate_fn=generate_batch)
    valid_loader = DataLoader(sub_valid_,
                              batch_size=BATCH_SIZE,
                              collate_fn=generate_batch)
    test_loader = DataLoader(test_dataset,
                             batch_size=BATCH_SIZE,
                             collate_fn=generate_batch)

    for epoch in tqdm(range(N_EPOCHS)):

        start_time = time.time()
        train_loss, train_acc = train_fn(dataLoader=train_loader,
                                         model=model,
                                         optimizer=optimizer,
                                         scheduler=scheduler,
                                         criterion=criterion,
                                         device=device)
        valid_loss, valid_acc = evaluate_fn(dataLoader=valid_loader,
                                            model=model,
                                            criterion=criterion,
                                            device=device)

        secs = int(time.time() - start_time)
        mins = secs / 60
        secs = secs % 60

        print('Epoch: %d' % (epoch + 1),
              " | time in %d minutes, %d seconds" % (mins, secs))
        print(
            f'\tLoss: {train_loss:.4f}(train)\t|\tAcc: {train_acc * 100:.1f}%(train)'
        )
        print(
            f'\tLoss: {valid_loss:.4f}(valid)\t|\tAcc: {valid_acc * 100:.1f}%(valid)'
        )
        if valid_loss < min_valid_loss:
            torch.save(model.state_dict(),
                       "../weights/text_news{}.pth".format(valid_loss))
            print(min_valid_loss, "--------->>>>>>>>", valid_loss)
            min_valid_loss = valid_loss

    print('Checking the results of test dataset...')
    test_loss, test_acc = evaluate_fn(dataLoader=test_loader,
                                      model=model,
                                      criterion=criterion,
                                      device=device)
    print(
        f'\tLoss: {test_loss:.4f}(test)\t|\tAcc: {test_acc * 100:.1f}%(test)')
コード例 #4
0
    vocab = build_vocab_from_iterator(yield_tokens(train_iter, ngrams),
                                      specials=["<unk>"])
    vocab.set_default_index(vocab["<unk>"])

    def text_pipeline(x):
        return vocab(list(ngrams_iterator(tokenizer(x), ngrams)))

    def label_pipeline(x):
        return int(x) - 1

    train_iter = DATASETS[args.dataset](root='.data', split='train')
    num_class = len(set([label for (label, _) in train_iter]))
    model = TextSentiment(len(vocab), embed_dim, num_class).to(device)

    criterion = torch.nn.CrossEntropyLoss().to(device)
    optimizer = torch.optim.SGD(model.parameters(), lr=lr)
    scheduler = torch.optim.lr_scheduler.StepLR(optimizer, 1.0, gamma=0.1)
    train_iter, test_iter = DATASETS[args.dataset]()
    train_dataset = to_map_style_dataset(train_iter)
    test_dataset = to_map_style_dataset(test_iter)
    num_train = int(len(train_dataset) * 0.95)
    split_train_, split_valid_ = random_split(
        train_dataset, [num_train, len(train_dataset) - num_train])
    train_dataloader = DataLoader(split_train_,
                                  batch_size=batch_size,
                                  shuffle=True,
                                  collate_fn=collate_batch)
    valid_dataloader = DataLoader(split_valid_,
                                  batch_size=batch_size,
                                  shuffle=True,
                                  collate_fn=collate_batch)
コード例 #5
0
def train_rating_model(
    YELP_TRAIN,
    fields,
    criterion,
    N_EPOCHS=20,
    split_ratio=0.9,
    num_hidden=30,
    embed_dim=50,
    actual_embed_dim=50,
):
    SEED = 0
    BATCH_SIZE = 16

    # Load and process data
    train_data = data.TabularDataset(path=YELP_TRAIN,
                                     format="json",
                                     fields=fields)
    print(YELP_TRAIN)
    print("NUM TRAIN", len(train_data.examples))
    assert len(train_data.examples) > 2
    TEXT = fields["text"][1]
    TEXT.build_vocab(train_data, vectors="glove.6B.%dd" % embed_dim)

    # Load model
    model = TextSentiment(
        vocab_size=len(TEXT.vocab),
        vocab=TEXT.vocab,
        embed_dim=actual_embed_dim,
        num_class=1,
        num_hidden=num_hidden,
    )

    # define optimizer and loss
    optimizer = optim.Adam(model.parameters())
    # criterion = nn.CrossEntropyLoss()

    # Train the model
    random.seed(0)
    train_data, valid_data = train_data.split(split_ratio=split_ratio,
                                              random_state=random.getstate())
    train_iterator, valid_iterator = data.Iterator.splits(
        (train_data, valid_data),
        batch_size=BATCH_SIZE,
        sort_key=lambda x: len(x.text),
        sort_within_batch=True,
        shuffle=True,
    )
    # iterator = data.Iterator(
    #    train_data,
    #    batch_size = BATCH_SIZE,
    #    sort_key = lambda x: len(x.text),
    #    sort_within_batch=True,
    #    shuffle=True)
    for epoch in range(N_EPOCHS):
        train_loss = train(model, train_iterator, optimizer, criterion)
        if epoch % 5 == 0:
            print(f"\tTrain Loss {epoch}: {train_loss:.3f}")
            evaluate(model, valid_iterator, criterion)

    evaluate(model, valid_iterator, criterion)
    return model