Beispiel #1
0
import time
from math import ceil
import utils.imgs
import utils.training as train_utils
from torch.autograd import Variable
from unet.unet_model import UNet4

#################################################data
PAP_PATH = Path('/home/zhaojie/zhaojie/PG/Pdata/')
WEIGHTS_PATH = Path('./weights/train_1024/')

WEIGHTS_PATH.mkdir(exist_ok=True)
batch_size = 60
normalize = transforms.Normalize(mean=data_PG.mean, std=data_PG.std)

train_joint_transformer = transforms.Compose([joint_transforms.JointRandomHorizontalFlip()])
train_dset = data_PG.CamVid(PAP_PATH, 'train', joint_transform=train_joint_transformer, transform=transforms.Compose([transforms.ToTensor(), normalize]))
train_loader = torch.utils.data.DataLoader(train_dset, batch_size=batch_size, shuffle=True)
	
val_dset = data_PG.CamVid(PAP_PATH, 'val', joint_transform=None, transform=transforms.Compose([ transforms.ToTensor(), normalize]))
val_loader = torch.utils.data.DataLoader(val_dset, batch_size=batch_size, shuffle=False)
	
test_dset = data_PG.CamVid(PAP_PATH, 'test', joint_transform=None, transform=transforms.Compose([ transforms.ToTensor(), normalize]))
test_loader = torch.utils.data.DataLoader(test_dset, batch_size=batch_size, shuffle=False)

print("Train: %d" %len(train_loader.dataset.imgs))
print("Val: %d" %len(val_loader.dataset.imgs))
print("Test: %d" %len(test_loader.dataset.imgs))
print("Classes: %d" % len(train_loader.dataset.classes))

inputs, targets = next(iter(train_loader))
Beispiel #2
0
from math import ceil
import utils.imgs
import utils.training as train_utils
from torch.autograd import Variable
from unet.unet_model import UNet1, UNet2, UNet3, UNet4

###############################################data

PAP_PATH = Path('/home/zhaojie/zhaojie/PG/Pdata/')
WEIGHTS_PATH = Path('./weights/train_PG_1024/')
WEIGHTS_PATH.mkdir(exist_ok=True)
batch_size = 30
normalize = transforms.Normalize(mean=data_PG.mean, std=data_PG.std)

train_joint_transformer = transforms.Compose(
    [joint_transforms.JointRandomHorizontalFlip()])
# train_joint_transform = None
train_dset = data_PG.CamVid(PAP_PATH,
                            'train',
                            joint_transform=train_joint_transformer,
                            transform=transforms.Compose(
                                [transforms.ToTensor(), normalize]))
train_loader = torch.utils.data.DataLoader(train_dset,
                                           batch_size=batch_size,
                                           shuffle=True)

val_dset = data_PG.CamVid(PAP_PATH,
                          'val',
                          joint_transform=None,
                          transform=transforms.Compose(
                              [transforms.ToTensor(), normalize]))
Beispiel #3
0
def main_worker(gpu, ngpus_per_node, args):

    dist.init_process_group(backend='nccl')
    torch.cuda.set_device(gpu)

    #################################################data
    PAP_PATH = Path('/home/zhaojie/zhaojie/PG/Pdata/')
    WEIGHTS_PATH = Path('./weights/train_1024/')

    WEIGHTS_PATH.mkdir(exist_ok=True)
    batch_size = 60
    normalize = transforms.Normalize(mean=data_PG.mean, std=data_PG.std)

    train_joint_transformer = transforms.Compose(
        [joint_transforms.JointRandomHorizontalFlip()])
    train_dset = data_PG.CamVid(PAP_PATH,
                                'train',
                                joint_transform=train_joint_transformer,
                                transform=transforms.Compose(
                                    [transforms.ToTensor(), normalize]))
    val_dset = data_PG.CamVid(PAP_PATH,
                              'val',
                              joint_transform=None,
                              transform=transforms.Compose(
                                  [transforms.ToTensor(), normalize]))
    test_dset = data_PG.CamVid(PAP_PATH,
                               'test',
                               joint_transform=None,
                               transform=transforms.Compose(
                                   [transforms.ToTensor(), normalize]))
    #######################################
    train_sampler = torch.utils.data.distributed.DistributedSampler(train_dset)
    train_loader = torch.utils.data.DataLoader(train_dset,
                                               batch_size=batch_size,
                                               num_workers=2,
                                               pin_memory=True,
                                               sampler=train_sampler)
    #######################################
    val_sampler = torch.utils.data.distributed.DistributedSampler(val_dset)
    val_loader = torch.utils.data.DataLoader(val_dset,
                                             batch_size=batch_size,
                                             num_workers=2,
                                             pin_memory=True,
                                             sampler=val_sampler)
    #######################################
    test_sampler = torch.utils.data.distributed.DistributedSampler(test_dset)
    test_loader = torch.utils.data.DataLoader(test_dset,
                                              batch_size=batch_size,
                                              num_workers=2,
                                              pin_memory=True,
                                              sampler=test_sampler)
    #######################################

    print("Train: %d" % len(train_loader.dataset.imgs))
    print("Val: %d" % len(val_loader.dataset.imgs))
    print("Test: %d" % len(test_loader.dataset.imgs))
    print("Classes: %d" % len(train_loader.dataset.classes))

    inputs, targets = next(iter(train_loader))
    print("Inputs: ", inputs.size())
    print("Targets: ", targets.size())
    # utils.imgs.view_image(inputs[0])
    # utils.imgs.view_annotated(targets[0])
    EE = 4
    device = 'cuda'
    EE_size = 256
    LR = 1e-3
    model = UNet4(n_channels=3, n_classes=4)
    # gpu = args.local_rank
    torch.cuda.set_device(gpu)
    model.cuda(gpu)
    # model = model.to(device)
    # model = torch.nn.DataParallel(model).cuda()
    ########################################
    model = torch.nn.parallel.DistributedDataParallel(
        model, find_unused_parameters=True, device_ids=[gpu])
    ###################################################
    optimizer = torch.optim.RMSprop(model.parameters(),
                                    lr=LR,
                                    weight_decay=1e-4)
    # print('EE, model', EE, model)
    pred_dir = './train_PG_pred/'
    FILE_test_imgs_original = '/home/zhaojie/zhaojie/PG/Pdata/test'

    #############################

    EE0 = 1  #EE =1-8#2-16#3-32#4-64#5-128#6-256#7-512#8-1024#

    epoch_num = 10000
    best_loss = 1.
    best_dice = 0.
    LR_DECAY = 0.95
    DECAY_EVERY_N_EPOCHS = 10
    criterion = nn.NLLLoss(weight=data_PG.class_weight.cuda()).cuda(gpu)
    cudnn.benchmark = True
    for epoch in range(1, epoch_num):
        start_time = datetime.datetime.now()
        print('start_time', start_time)
        model = model.cuda()

        ##################################################
        ### Train ###
        trn_loss, trn_err, train_DICE = train_net(model, train_loader,
                                                  criterion, optimizer,
                                                  EE_size)
        print('Epoch {:d}\nTrain - Loss: {:.4f}, Acc: {:.4f}, Dice: {:.4f}'.
              format(epoch, trn_loss, 1 - trn_err, train_DICE))
        ## Test ###
        val_loss, val_err, val_DICE = eval_net(model, val_loader, criterion,
                                               EE_size)
        print('Val - Loss: {:.4f} | Acc: {:.4f}, Dice: {:.4f}'.format(
            val_loss, 1 - val_err, val_DICE))
        ### Checkpoint ###
        DICE1 = view_sample_predictions(model, test_loader,
                                        FILE_test_imgs_original, pred_dir,
                                        EE_size)
        print('-----------test_dice', DICE1)
        if best_dice < DICE1:
            # save_weights_dice(WEIGHTS_PATH, model, epoch, train_DICE, val_DICE, DICE1, EE)
            best_dice = DICE1
        ### Adjust Lr ###
        adjust_learning_rate(LR, LR_DECAY, optimizer, epoch,
                             DECAY_EVERY_N_EPOCHS)
        end_time = datetime.datetime.now()
        print('end_time', end_time)
        print('time', (end_time - start_time).seconds)