def main():
    model_path = '/home/warvisionary/parlai_transfer/from_pretrained/model'
    print(f"Training model, writing output to {model_path}")
    TrainModel.main(
        # use our task on a pre-trained model
        task='empathetic_dialogues_ru', 
        model='transformer/generator',
        model_file=model_path,
        
        # initialize with a pretrained model
        init_model='zoo:tutorial_transformer_generator/model',
        
        # arguments we get from the pretrained model.
        # Unfortunately, these must be looked up separately for each model.
        n_heads=16, n_layers=8, n_positions=512, text_truncate=512,
        label_truncate=128, ffn_size=2048, embedding_size=512,
        activation='gelu', variant='xlm',
        dict_lower=True, dict_tokenizer='bpe',
        dict_file='zoo:tutorial_transformer_generator/model.dict',
        learn_positional_embeddings=True,
        
        # some training arguments, specific to this fine-tuning
        # use a small learning rate with ADAM optimizer
        lr=1e-5, optimizer='adam',
        warmup_updates=100,
        # early stopping on perplexity
        validation_metric='ppl',
        # train at most 10 minutes, and validate every 0.25 epochs
        max_train_time=600, validation_every_n_epochs=0.25,
        
        # gpu-based params
        batchsize=12, fp16=True, fp16_impl='mem_efficient',
        
        # speeds up validation
        skip_generation=True,
        
        # helps us cram more examples into our gpu at a time
        dynamic_batching='full',
    )
#!/usr/bin/env python3

# Copyright (c) Facebook, Inc. and its affiliates.
# This source code is licensed under the MIT license found in the
# LICENSE file in the root directory of this source tree.
"""
Train a model using parlai's standard training loop.

For documentation, see parlai.scripts.train_model.
"""

from parlai.scripts.train_model import TrainModel

if __name__ == '__main__':
    TrainModel.main()
示例#3
0
文件: train.py 项目: simplecoka/cortx
                if len(splits) == 3:
                    yield (splits[0].replace('text:', ''),
                           splits[1].replace('labels:', '')), new_episode
                    new_episode = True
                else:
                    yield (splits[0].replace('text:', ''),
                           splits[1].replace('labels:', '')), new_episode
                    new_episode = False


DisplayModel.main(task='poly_teacher', model='poly')
TrainModel.main(
    model='seq2seq',
    model_file='poly-encoder/model',
    dict_file='zoo:dodecadialogue/empathetic_dialogues_ft/model.dicts',
    task='poly_teacher',
    batchsize=3,
    validation_every_n_secs=10,
    max_train_time=60,
)

DisplayModel.main(
    task='poly_teacher',
    model_file='poly-encoder/model',
    num_examples=6,
)
for source, dirs, files in walks:
    print('Directory: ' + source)
    for filename in files:
        # construct the full local path
        local_file = os.path.join(source, filename)
TrainModel.main(
    # similar to before
    task='bankQA',
    model='projects.wizard_of_wikipedia.generator.agents:EndToEndAgent',
    model_file='/from_pretrained_wiki/model',
    # initialize with a pretrained model
    init_model='zoo:wizard_of_wikipedia/end2end_generator/model',

    # arguments we get from the pretrained model.
    # Unfortunately, these must be looked up separately for each model.
    # eps
    dict_file='zoo:wizard_of_wikipedia/end2end_generator/model.dict',
    num_epochs=2,
    dict_lower=True,
    dict_tokenizer='bpe',
    n_layers=5,
    n_heads=2,
    dropout=0.20,
    ffn_size=512,
    embedding_size=256,
    log_every_n_secs=10,
    validation_patience=12,
    validation_metric='ppl',
    validation_metric_mode='min',
    validation_every_n_epochs=0.5,
    n_positions=128,
    truncate=128,
    max_knowledge=32,
    knowledge_alpha=0.95,
    knowledge_truncate=32,
    learningrate=5e-4,
    warmup_updates=5000,
    clip=0.1,
    lr_scheduler='invsqrt',
    embedding_type='fasttext',
    beam_size=1,
    skip_generation=False,
    batchsize=64,
    # fp16=True
)
示例#5
0
TrainModel.main(
    model_file='model/kvmemnn/self_persona_short',
    task='personachat:self',
    
    # model_file='model/kvmemnn/self_persona_rv',
    # task='personachat:self_revised',

    # model_file='model/kvmemnn/no_persona',
    # task='personachat:none',
    
    # model_file='model/kvmemnn/no_persona_rv',
    # task='personachat:none_revised', 

    # model_file='model/kvmemnn/other_persona',
    # task='personachat:other',
    
    # model_file='model/kvmemnn/other_persona_rv',
    # task='personachat:other_revised',

    # model_file='model/kvmemnn/both_persona',
    # task='personachat:both',

    # model_file='model/kvmemnn/both_persona_rv',
    # task='personachat:both_revised',

    model='projects.personachat.kvmemnn.kvmemnn:KvmemnnAgent',
    log_every_n_secs=60,
    validation_max_exs=10000,
    validation_every_n_secs=120,
    hops=1,
    lins=0,
    validation_patience=-1,
    validation_metric='hits@1',
    max_train_time=3600,
    share_embeddings=True,
    batchsize=1,
    learningrate=0.1,
    embeddingsize=500,
    margin=0.1,
    tfidf=False,
    numthreads=10,

    #### FROM : https://github.com/facebookresearch/ParlAI/pull/662#issuecomment-475629794

    # model_file='model/seq2seq/self_persona_2',
    # task='personachat:self',

    # model='seq2seq',

    # log_every_n_secs=300,
    # # validation_every_n_secs=1800,
    # # validation_patience=-1,
    # # validation_metric='ppl',
    # max_train_time=11 * 3600,
    # batchsize=128,
    # learningrate=0.001,
    # embeddingsize=300,
    # numlayers=2,
    # hiddensize=512,
    # dropout=0.5,
    # optimizer='adam',
    # lookuptable='enc_dec',
    # # margin=0.1,
    # # tfidf=False,
    # #numthreads=8,

    ## GPU : https://dailylime.kr/2020/06/wsl2%EC%97%90%EC%84%9C-ubuntu%EC%99%80-cuda-%EC%82%AC%EC%9A%A9%ED%95%98%EA%B8%B0/
    # fp16=True, fp16_impl='mem_efficient',
)