Esempio n. 1
0
def re_train():
    iter = 5
    iter_size = 10000
    train, dev = read_corpus(random_state=1234,
                             separator='\t',
                             iter=iter,
                             iter_size=iter_size)

    mp = "tf_cnn"
    model_path = os.path.join(MODEL_PATH, mp, 'checkpoints')
    ckpt_file = tf.train.latest_checkpoint(model_path)
    logger.info("load pre-train model from {}".format(ckpt_file))
    textCNN = TextCNN(
        model_path=ckpt_file,
        vocab=word2int,
        tag2label=tag2label,
        sequence_length=FLAGS.sequence_length,
        eopches=FLAGS.epoches,
    )

    saver = tf.compat.v1.train.Saver()

    with tf.compat.v1.Session(config=cfg()) as sess:
        saver.restore(sess, ckpt_file)
        textCNN.set_model_path(model_path=os.path.join(MODEL_PATH, mp))
        textCNN.train(sess, train, dev, shuffle=True, re_train=True)
Esempio n. 2
0
def re_train():
    train, dev = read_corpus(filename='emergency_train.tsv', test_size=0.2)

    model_path = os.path.join(MODEL_PATH, FLAGS.DEMO, 'checkpoints')
    ckpt_file = tf.train.latest_checkpoint(model_path)

    logging.info("load pre-train model from {}".format(ckpt_file))
    textAttRNN = TextAttRNN(
        config=cfg(),
        model_path=ckpt_file,
        vocab=word2int,
        tag2label=tag2label,
        batch_size=FLAGS.batch_size,
        embed_size=FLAGS.embed_size,
        sequence_length=FLAGS.sequence_length,
        eopches=FLAGS.epoches,
    )

    saver = tf.compat.v1.train.Saver()

    with tf.compat.v1.Session(config=cfg()) as sess:
        saver.restore(sess, ckpt_file)
        textAttRNN.set_model_path(
            model_path=os.path.join(MODEL_PATH, FLAGS.DEMO))
        textAttRNN.train(sess, train, dev, shuffle=True, re_train=True)
Esempio n. 3
0
def re_train():
    tag2label = {'0': 0, '1': 1}
    iter = 4
    iter_size = 10000
    train, dev = read_corpus(random_state=1234,
                             separator='\t',
                             iter=iter,
                             iter_size=iter_size)

    mp = 'torch_rnn'
    model_path = os.path.join(MODEL_PATH, mp, 'model.pth')

    logger.info("load pre-train model from {}".format(model_path))
    textRNN = TextAttBiRNN(
        model_path=model_path,
        vocab=word2int,
        tag2label=tag2label,
        bidirectional=True,
        sequence_length=FLAGS.sequence_length,
        epoches=FLAGS.epoches,
        batch_size=FLAGS.batch_size,
        layer_size=2,
    )

    textRNN.load_state_dict(torch.load(model_path), strict=False)
    textRNN.set_model_path(model_path=os.path.join(MODEL_PATH, mp))

    textRNN.train(train, dev, shuffle=True, re_train=True)
Esempio n. 4
0
def train():
    train, dev = read_corpus(filename='emergency_train.tsv')
    textAttRNN = TextAttRNN(config=cfg(),
                            model_path=os.path.join(MODEL_PATH, FLAGS.DEMO),
                            vocab=word2int,
                            tag2label=tag2label,
                            batch_size=FLAGS.batch_size,
                            embed_size=FLAGS.embed_size,
                            sequence_length=FLAGS.sequence_length,
                            eopches=FLAGS.epoches)

    with tf.compat.v1.Session(config=cfg()) as sess:
        textAttRNN.train(sess, train, dev, shuffle=True)
Esempio n. 5
0
def train():
    iter = -1
    iter_size = 20000
    train, dev = read_corpus(random_state=1234, separator='\t', iter=iter, iter_size=iter_size)
    textCNN = TextAttRNN(config=cfg(),
                         model_path=os.path.join(MODEL_PATH, FLAGS.DEMO),
                         vocab=word2int,
                         tag2label=tag2label,
                         batch_size=FLAGS.batch_size,
                         embed_size=FLAGS.embed_size,
                         eopches=FLAGS.epoches)

    with tf.compat.v1.Session(config=cfg()) as sess:
        textCNN.train(sess, train, dev, shuffle=True)
Esempio n. 6
0
def train():
    iter = 0
    iter_size = 10000
    train, dev = read_corpus(random_state=1234,
                             separator='\t',
                             iter=iter,
                             iter_size=iter_size)

    mp = "iter_{}_size_{}_epochs_{}".format(str(iter + 1), iter_size,
                                            FLAGS.epoches)
    textCNN = TextCNN(
        # model_path=os.path.join(MODEL_PATH, str(int(time.time()))),
        model_path=os.path.join(MODEL_PATH, "tf_cnn", mp),
        vocab=word2int,
        tag2label=tag2label,
        batch_size=FLAGS.batch_size,
        eopches=FLAGS.epoches)
    with tf.compat.v1.Session(config=cfg()) as sess:
        textCNN.train(sess, train, dev, shuffle=True)
Esempio n. 7
0
def train():
    iter = 0
    iter_size = 10000
    train, dev = read_corpus(random_state=1234,
                             separator='\t',
                             iter=iter,
                             iter_size=iter_size)

    mp = "torch_rnn".format(str(iter + 1), iter_size, FLAGS.epoches)

    model = TextAttBiRNN(
        model_path=os.path.join(MODEL_PATH, "", mp),
        vocab=word2int,
        tag2label=tag2label,
        bidirectional=True,
        sequence_length=FLAGS.sequence_length,
        epoches=FLAGS.epoches,
        batch_size=FLAGS.batch_size,
        layer_size=2,
    )
    model.train(train, dev, shuffle=True)