Exemple #1
0
def train(arguments):

    # Parse input arguments
    json_filename = arguments.config
    network_debug = arguments.debug

    # Load options
    json_opts = json_file_to_pyobj(json_filename)
    train_opts = json_opts.training

    # Architecture type
    arch_type = train_opts.arch_type

    # Setup Dataset and Augmentation
    ds_class = get_dataset(arch_type)
    ds_path = get_dataset_path(arch_type, json_opts.data_path)
    ds_transform = get_dataset_transformation(arch_type,
                                              opts=json_opts.augmentation)

    # Setup the NN Model
    model = get_model(json_opts.model)
    network_debug = False
    if network_debug:
        print('# of pars: ', model.get_number_parameters())
        print('fp time: {0:.3f} sec\tbp time: {1:.3f} sec per sample'.format(
            *model.get_fp_bp_time()))
        exit()

    # Setup Data Loader
    train_dataset = ds_class(ds_path,
                             split='train',
                             transform=ds_transform['train'],
                             preload_data=train_opts.preloadData)
    valid_dataset = ds_class(ds_path,
                             split='validation',
                             transform=ds_transform['valid'],
                             preload_data=train_opts.preloadData)
    test_dataset = ds_class(ds_path,
                            split='test',
                            transform=ds_transform['valid'],
                            preload_data=train_opts.preloadData)

    train_loader = DataLoader(dataset=train_dataset,
                              num_workers=0,
                              batch_size=train_opts.batchSize,
                              shuffle=True)
    valid_loader = DataLoader(dataset=valid_dataset,
                              num_workers=0,
                              batch_size=train_opts.batchSize,
                              shuffle=False)

    test_loader = DataLoader(dataset=test_dataset,
                             num_workers=0,
                             batch_size=train_opts.batchSize,
                             shuffle=False)

    # Visualisation Parameters
    visualizer = Visualiser(json_opts.visualisation, save_dir=model.save_dir)
    error_logger = ErrorLogger()

    # Training Function

    model.set_scheduler(
        train_opts,
        len_train=len(train_loader),
        max_lr=json_opts.model.max_lr,
        division_factor=json_opts.model.division_factor,
        last_epoch=-1 if not json_opts.model.continue_train else
        (json_opts.model.which_epoch * len(train_loader)))

    for epoch in range(model.which_epoch, train_opts.n_epochs):
        print('(epoch: %d, total # iters: %d)' % (epoch, len(train_loader)))
        if epoch % 2 == 0:
            print("freezing model")
            model.freeze()
        else:
            model.unfreeze()

        # Training Iterations

        for epoch_iter, (images, labels) in tqdm(enumerate(train_loader, 1),
                                                 total=len(train_loader)):
            # Make a training update
            model.set_input(images, labels)
            # model.optimize_parameters()
            model.optimize_parameters_accumulate_grd(epoch_iter)
            lr = model.update_learning_rate()

            lr = {"lr": lr}
            # Error visualisation
            errors = model.get_current_errors()
            stats = model.get_segmentation_stats()
            error_logger.update({**errors, **stats, **lr}, split='train')
            visualizer.plot_current_errors(epoch,
                                           error_logger.get_errors('train'),
                                           split_name='train')

        # Validation and Testing Iterations
        for loader, split in zip([valid_loader], ['validation']):
            for epoch_iter, (images, labels) in tqdm(enumerate(loader, 1),
                                                     total=len(loader)):

                # Make a forward pass with the model
                model.set_input(images, labels)
                model.validate()

                # Error visualisation
                errors = model.get_current_errors()
                stats = model.get_segmentation_stats()
                error_logger.update({**errors, **stats}, split=split)

                # Visualise predictions
                visuals = model.get_current_visuals()
                visualizer.display_current_results(visuals,
                                                   epoch=epoch,
                                                   save_result=False)

        # Update the plots
        for split in ['validation']:
            visualizer.plot_current_errors(epoch,
                                           error_logger.get_errors(split),
                                           split_name=split)
            visualizer.print_current_errors(epoch,
                                            error_logger.get_errors(split),
                                            split_name=split)
        error_logger.reset()

        # Save the model parameters
        if epoch % train_opts.save_epoch_freq == 0:
            model.save(epoch)
            visualizer.save_model(epoch)

        # Update the model learning rate
        # model.update_learning_rate()

    visualizer.finish()
Exemple #2
0
            errors = model.get_current_errors()
            stats = model.get_segmentation_stats()
            error_logger.update({**errors, **stats}, split=split)

            # Visualise predictions
            visuals = model.get_current_visuals()
            visualizer.display_current_results(visuals,
                                               epoch=epoch,
                                               save_result=False)

    # Update the plots
    for split in ['validation']:
        visualizer.plot_current_errors(epoch,
                                       error_logger.get_errors(split),
                                       split_name=split)
        visualizer.print_current_errors(epoch,
                                        error_logger.get_errors(split),
                                        split_name=split)
    error_logger.reset()
    visualizer.upload_limit = 45

    # Save the model parameters
    if epoch % train_opts.save_epoch_freq == 0:
        model.save(epoch)
        visualizer.save_model(epoch)

    # Update the model learning rate
    # model.update_learning_rate()

visualizer.finish()