示例#1
0
    def __init__(self, model=None, optimizer=None):
        if model is None:
            self.model = CNN().cuda()
        else:
            self.model = model

        if optimizer is None:
            self.optimizer = torch.optim.Adam(self.model.parameters(), lr=1e-4, weight_decay=1e-5)
        else:
            self.optimizer = optimizer

        self.bestModel = self.model
        self.epochLosses = []
示例#2
0
class Training:
    def __init__(self, model=None, optimizer=None):
        if model is None:
            self.model = CNN().cuda()
        else:
            self.model = model

        if optimizer is None:
            self.optimizer = torch.optim.Adam(self.model.parameters(), lr=1e-4, weight_decay=1e-5)
        else:
            self.optimizer = optimizer

        self.bestModel = self.model
        self.epochLosses = []

    #Training procedure
    def training(self, epoch, train_dataloader, dataset):
        self.model.train()
        mse_list = []
        rmse_list = []

        for batch_id, (data, target) in enumerate(train_dataloader):
            target = target.view(-1, 1)
            self.optimizer.zero_grad()
            data = data.float().cuda()
            target = target.float().cuda()
            out = self.model(data)
            criterion = torch.nn.MSELoss()
            loss = criterion(out, target)
            mse_list.append(loss.data.item())
            rmse_list.append(np.sqrt(loss.data.item()))
            loss.backward()
            self.optimizer.step()
            # Print is adapted from https://github.com/pytorch/examples/blob/master/mnist_hogwild/train.py
            print('Train Epoch: {} [{}/{} ({:.0f}%)]\tLoss: {:.6f} MSE-Mean: {:.6f} RMSE-Mean:  {:.6f}'.format(
                epoch, batch_id * len(data), dataset.__len__(),
                       100. * (batch_id * len(data)) / dataset.__len__(), loss.data.item(), np.mean(mse_list),
                np.mean(rmse_list)))

        print('Train Epoch: {} MSE (loss): {:.4f}, RMSE: {:.4f} Dataset length {}'.format(epoch, np.mean(mse_list),
                                                                                          np.mean(rmse_list),
                                                                                          dataset.__len__()))
        return np.mean(mse_list), np.mean(rmse_list)

    # Tool-Class to calculate the prediction, if ensemble-models were used. -> Only used in validation not in training
    def calcPred(self, prediction, remember, batchnum):
        preds = []
        preds.append(prediction)
        lene = len(self.epochLosses)
        if lene < remember:
            lene2 = -1
        else:
            lene2 = lene - remember
        for i in range(lene - 1, lene2, -1):
            preds.append(self.epochLosses[i][batchnum])
        print(preds)
        preds = torch.tensor(torch.mean(torch.tensor(preds)))
        print(preds)
        return preds

    # Testing (validation)
    def testing(self, epoch, test_dataloader, dataset, ensemble=False, remember=10, rotation=True):
        self.model.eval()
        mse_list = []
        rmse_list = []
        batchnum = 0
        batch_losses = []
        criterion = torch.nn.MSELoss()
        if rotation:
            rot = Rotations()

        for batch_id, (data, target) in enumerate(test_dataloader):
            target = target.view(-1, 1)
            data = data.float().cuda()
            target = target.float()
            if rotation:
                outs = []
                datatemp = data.cpu().numpy()[0]
                for j in range(24):
                    newdat = rot.rotation(datatemp, j).copy()
                    newdat = torch.from_numpy(newdat).float().cuda().view((1, 16, 24, 24, 24))
                    outs.append(self.model(newdat))
                out = torch.tensor([[torch.mean(torch.tensor(outs))]])
            else:
                out = self.model(data)
            if ensemble:
                pred = out
                if epoch == 0:
                    prediction = pred
                else:
                    prediction = self.calcPred(pred, remember=remember, batchnum=batchnum)
                batch_losses.append(out)
                prediction = out
            batchnum += 1

            loss = criterion(prediction, target)
            mse_list.append(loss.data.item())
            rmse_list.append(np.sqrt(loss.data.item()))

            # Print is adapted from https://github.com/pytorch/examples/blob/master/mnist_hogwild/train.py
            print('Test Epoch: {} [{}/{} ({:.0f}%)]\tLoss: {:.6f} MSE-Mean: {:.6f} RMSE-Mean:  {:.6f}'.format(
                epoch, batch_id * len(data), dataset.__len__(),
                       100. * (batch_id * len(data)) / dataset.__len__(), loss.data.item(), np.mean(mse_list),
                np.mean(rmse_list)))

        if ensemble:
            self.epochLosses.append(batch_losses)
        print('Test Epoch: {} MSE (loss): {:.4f}, RMSE: {:.4f} Dataset length {}'.format(epoch, np.mean(mse_list),
                                                                                         np.mean(rmse_list),
                                                                                         dataset.__len__()))
        return np.mean(mse_list), np.mean(rmse_list)

    # Benchmark very similar to validation
    def benchmark(self, n_datapoints, datapath, rotations=True, model=None, ensemble=False, version = 2):
        if ensemble:
            best_models = []
            for i in os.listdir(model):
                if 'bestModel' in i:
                    self.model = torch.load(model+i)
                    self.model.eval()
                    best_models.append(self.model)

        if rotations:
            rot = Rotations()
            datafile = datapath
            labels = []
            outs = []

            if model is not None and not ensemble:
                self.bestModel.load_state_dict(torch.load(model + 'bestModel.pt'))
                self.bestModel.eval()

            for i in range(n_datapoints):
                outs1 = []
                target1 = []

                for j in range(24):
                    with h5py.File(datafile, 'r') as file:
                        data = rot.rotation(data=file[str(i) + '/data'][()][0], k=j)
                        # In the old version, wrong log was used -> recalc
                        if version == 1:
                            label = -np.log10(np.exp(-(file[str(i) + '/label'][()])))
                        else:
                            label = file[str(i) + '/label'][()]
                    data = torch.from_numpy(data.reshape(1, 16, 24, 24, 24).copy()).float().cuda()
                    if ensemble:
                        outall = []
                        for m in best_models:
                            outall.append(m(data))
                        out = torch.mean(torch.tensor(outall))
                    else:
                        out = self.bestModel(data)
                    outs1.append(out.cpu().data.numpy())
                    target1.append(label)

                labels.append(np.mean(target1))
                outs.append(np.mean(outs1))

            error = []
            for i in range(290):
                error.append((outs[i] - labels[i]) ** 2)
            print("testmean: ", np.mean(error))

            return error, labels, outs
        else:
            kwargs = {'num_workers': 4}
            indices = np.arange(n_datapoints)
            test_set = OwnDataset(indices, datapath, rotations=False)
            test_dataloader = DataLoader(dataset=test_set, batch_size=1, shuffle=False, **kwargs)

            if model is not None and not ensemble:
                self.bestModel.load_state_dict(torch.load(model + 'bestModel.pt'))
            self.bestModel.eval()

            outs1 = []
            target1 = []

            for batch_id, (data, target) in enumerate(test_dataloader):
                target = target.view(-1, 1)
                target1.append(target.cpu().data.numpy())
                data = data.float().cuda()
                if ensemble:
                    outall = []
                    for m in best_models:
                        out = m(data)
                        outall.append(out)
                    out = torch.mean(torch.tensor(outall))
                else:
                    out = self.bestModel(data)
                outs1.append(out.cpu().data.numpy())

            error = []
            for i in range(290):
                error.append((outs1[i] - target1[i]) ** 2)
            print(outs1, target1)
            print("testmean: ", np.mean(error))

            return error, target1, outs1

    # overall model training -> combines training, validation and testing
    def fit(self, epochs, train_path, result_datapath, kwargs=None, n_datapoints=3767, prct_train=0.8, test_path=None, n_test=290,
            batch_size_train=64, batch_size_test=32, ensemble=False, remember=10, augmentation=True):
        print(self.model)
        print(Training.count_parameters(self.model))
        lowest_loss = np.inf
        train_mse = []
        test_mse = []
        train_rmse = []
        test_rmse = []
        if kwargs is None:
            kwargs = {'num_workers': 4}

        # split sets and create datasets
        indices = np.arange(n_datapoints)
        train_size = int(prct_train * n_datapoints)
        test_size = n_datapoints - train_size
        train, test = torch.utils.data.random_split(indices, [train_size, test_size])
        train_set = OwnDataset(train, train_path, rotations=augmentation)

        if test_path is None:
            test_set = OwnDataset(test, train_path, rotations=False)
        else:
            test_set = OwnDataset(np.arange(n_test), test_path, rotations=False)

        train_dataloader = DataLoader(dataset=train_set, batch_size=batch_size_train, shuffle=True, **kwargs)
        if ensemble:
            test_dataloader = DataLoader(dataset=test_set, batch_size=1, shuffle=False, **kwargs)
        else:
            test_dataloader = DataLoader(dataset=test_set, batch_size=1, shuffle=False, **kwargs)

        best_losses = []

        # start training
        for epoch in range(epochs):
            mse, rmse = self.training(epoch, train_dataloader, train_set)
            train_mse.append(mse)
            train_rmse.append(rmse)

            mse, rmse = self.testing(epoch, test_dataloader, test_set, ensemble, remember)
            test_mse.append(mse)
            test_rmse.append(rmse)

            if ensemble:
                if len(best_losses) < remember:
                    best_losses.append(test_mse[-1])
                    torch.save(self.model, result_datapath + 'bestmodels/' + 'bestModel' + str(len(best_losses)) + '.pt')
                else:
                    bad_loss = np.argmax(best_losses)
                    if best_losses[bad_loss] > test_mse[-1]:
                        best_losses[bad_loss] = test_mse[-1]
                        torch.save(self.model, result_datapath + 'bestmodels/' + 'bestModel' + str(bad_loss) + '.pt')
            else:
                if test_mse[-1] < lowest_loss:
                    lowest_loss = test_mse[-1]
                    torch.save(self.model.state_dict(), result_datapath + 'bestmodels/' + 'bestModel.pt')
                    self.bestModel = self.model
            torch.save(self.model, result_datapath + 'lastModel.pt')
            torch.save({'epoch': epoch, 'model_state_dict': self.model.state_dict(),
                        'optimizer_state_dict': self.optimizer.state_dict()}, result_datapath + 'lastModel.tar')

        hist = h5py.File(result_datapath + 'history.hdf5', 'w')
        hist.create_dataset('train_mse', data=train_mse)
        hist.create_dataset('test_mse', data=test_mse)
        hist.create_dataset('train_rmse', data=train_rmse)
        hist.create_dataset('test_rmse', data=test_rmse)
        hist.close()

    # This function counts the number of trainable parameters of the network
    # Taken from https://stackoverflow.com/questions/48393608/pytorch-network-parameter-calculation (Wasi Ahmad https://stackoverflow.com/users/5352399/wasi-ahmad)
    def count_parameters(model):
        paramCount = 0
        for parameterName, parameterValue in model.named_parameters():
            if parameterValue.requires_grad:
                n = np.prod(parameterValue.size())
                if parameterValue.dim() <= 1:
		            print(parameterName, ':', n)
                else:
                    layerSize = list(parameterValue.size())
                    if (len(layerSize) > 0):
                        layer = str(layerSize[0])
                        for i in range(1, len(layerSize)):
                            layer += "x"+  str(layerSize[i])
                            print(parameterName, ':', layer , '=', n)                    
                paramCount += n
        return paramCount
示例#3
0
class Training:
    def __init__(self, model=None, optimizer=None):
        if model is None:
            self.model = CNN().cuda()
        else:
            self.model = model

        if optimizer is None:
            self.optimizer = torch.optim.Adam(self.model.parameters(),
                                              lr=1e-4,
                                              weight_decay=1e-5)
        else:
            self.optimizer = optimizer

        self.bestModel = self.model
        self.epochLosses = []

    def training(self, epoch, train_dataloader, dataset):
        self.model.train()
        mse_list = []
        rmse_list = []

        for batch_id, (data, target) in enumerate(train_dataloader):
            target = target.view(-1, 1)
            self.optimizer.zero_grad()
            data = data.float().cuda()
            target = target.float().cuda()
            out = self.model(data)
            criterion = torch.nn.MSELoss()
            loss = criterion(out, target)
            mse_list.append(loss.data.item())
            rmse_list.append(np.sqrt(loss.data.item()))
            loss.backward()
            self.optimizer.step()

            print(
                'Train Epoch: {} [{}/{} ({:.0f}%)]\tLoss: {:.6f} MSE-Mean: {:.6f} RMSE-Mean:  {:.6f}'
                .format(epoch, batch_id * len(data), dataset.__len__(),
                        100. * (batch_id * len(data)) / dataset.__len__(),
                        loss.data.item(), np.mean(mse_list),
                        np.mean(rmse_list)))

        print(
            'Train Epoch: {} MSE (loss): {:.4f}, RMSE: {:.4f} Dataset length {}'
            .format(epoch, np.mean(mse_list), np.mean(rmse_list),
                    dataset.__len__()))
        return np.mean(mse_list), np.mean(rmse_list)

    def calcPred(self, prediction, remember, batchnum):
        preds = []
        preds.append(prediction)
        lene = len(self.epochLosses)
        if lene < remember:
            lene2 = -1
        else:
            lene2 = lene - remember
        for i in range(lene - 1, lene2, -1):
            preds.append(self.epochLosses[i][batchnum])
        print(preds)
        preds = torch.tensor(torch.mean(torch.tensor(preds)))
        print(preds)
        return preds

    def testing(self,
                epoch,
                test_dataloader,
                dataset,
                ensemble=False,
                remember=10,
                rotation=True):
        self.model.eval()
        mse_list = []
        rmse_list = []
        batchnum = 0
        batch_losses = []
        criterion = torch.nn.MSELoss()
        if rotation:
            rot = Rotations()

        for batch_id, (data, target) in enumerate(test_dataloader):
            target = target.view(-1, 1)
            data = data.float().cuda()
            target = target.float()
            if rotation:
                outs = []
                datatemp = data.cpu().numpy()[0]
                for j in range(24):
                    newdat = rot.rotation(datatemp, j).copy()
                    newdat = torch.from_numpy(newdat).float().cuda().view(
                        (1, 16, 24, 24, 24))
                    outs.append(self.model(newdat))
                out = torch.tensor([[torch.mean(torch.tensor(outs))]])
            else:
                out = self.model(data)
            if ensemble:
                pred = out
                if epoch == 0:
                    prediction = pred
                else:
                    prediction = self.calcPred(pred,
                                               remember=remember,
                                               batchnum=batchnum)
                batch_losses.append(out)
            else:
                prediction = out
            batchnum += 1

            loss = criterion(prediction, target)
            mse_list.append(loss.data.item())
            rmse_list.append(np.sqrt(loss.data.item()))

            print(
                'Test Epoch: {} [{}/{} ({:.0f}%)]\tLoss: {:.6f} MSE-Mean: {:.6f} RMSE-Mean:  {:.6f}'
                .format(epoch, batch_id * len(data), dataset.__len__(),
                        100. * (batch_id * len(data)) / dataset.__len__(),
                        loss.data.item(), np.mean(mse_list),
                        np.mean(rmse_list)))

        if ensemble:
            self.epochLosses.append(batch_losses)
        print(
            'Test Epoch: {} MSE (loss): {:.4f}, RMSE: {:.4f} Dataset length {}'
            .format(epoch, np.mean(mse_list), np.mean(rmse_list),
                    dataset.__len__()))
        return np.mean(mse_list), np.mean(rmse_list)

    def benchmark(self,
                  n_datapoints,
                  datapath,
                  rotations=True,
                  model=None,
                  ensemble=False):
        if ensemble:
            best_models = []
            for i in os.listdir(model):
                if 'bestModel' in i:
                    self.model = torch.load(model + i)
                    self.model.eval()
                    best_models.append(self.model)

        if rotations:
            rot = Rotations()
            datafile = datapath
            labels = []
            outs = []

            if model is not None and not ensemble:
                self.bestModel.load_state_dict(
                    torch.load(model + 'bestModel.pt'))
                self.bestModel.eval()

            for i in range(n_datapoints):
                outs1 = []
                target1 = []

                for j in range(24):
                    with h5py.File(datafile, 'r') as file:
                        data = rot.rotation(data=file[str(i) + '/data'][()][0],
                                            k=j)
                        label = file[str(i) + '/label'][()]
                    data = torch.from_numpy(
                        data.reshape(1, 16, 24, 24, 24).copy()).float().cuda()
                    if ensemble:
                        outall = []
                        for m in best_models:
                            outall.append(m(data))
                        out = torch.mean(torch.tensor(outall))
                    else:
                        out = self.bestModel(data)
                    outs1.append(out.cpu().data.numpy())
                    target1.append(label)

                labels.append(np.mean(target1))
                outs.append(np.mean(outs1))

            error = []
            for i in range(290):
                error.append((outs[i] - labels[i])**2)
            print("testmean: ", np.mean(error))

            return error, labels, outs
        else:
            kwargs = {'num_workers': 8}
            indices = np.arange(n_datapoints)
            test_set = OwnDataset(indices, datapath, rotations=False)
            test_dataloader = DataLoader(dataset=test_set,
                                         batch_size=1,
                                         shuffle=False,
                                         **kwargs)

            if model is not None and not ensemble:
                self.bestModel.load_state_dict(
                    torch.load(model + 'bestModel.pt'))
            self.bestModel.eval()

            outs1 = []
            target1 = []

            for batch_id, (data, target) in enumerate(test_dataloader):
                target = target.view(-1, 1)
                target1.append(target.cpu().data.numpy())
                data = data.float().cuda()
                if ensemble:
                    outall = []
                    for m in best_models:
                        out = m(data)
                        outall.append(out)
                    out = torch.mean(torch.tensor(outall))
                else:
                    out = self.bestModel(data)
                outs1.append(out.cpu().data.numpy())

            error = []
            for i in range(290):
                error.append((outs1[i] - target1[i])**2)
            print(outs1, target1)
            print("testmean: ", np.mean(error))

            return error, target1, outs1

    def fit(self,
            epochs,
            train_path,
            result_datapath,
            kwargs=None,
            n_datapoints=3767,
            prct_train=0.8,
            test_path=None,
            n_test=290,
            batch_size_train=64,
            batch_size_test=32,
            ensemble=False,
            remember=10,
            augmentation=True):
        print(self.model)
        print(Training.count_parameters(self.model))
        lowest_loss = np.inf
        train_mse = []
        test_mse = []
        train_rmse = []
        test_rmse = []
        if kwargs is None:
            kwargs = {'num_workers': 8}

#         indices = np.arange(n_datapoints)
#         train_size = int(prct_train * n_datapoints)
#         test_size = n_datapoints - train_size
#         train, test = torch.utils.data.random_split(indices, [train_size, test_size])
# Use the entire refined set as train set and core set as test/validation set
        train_set = OwnDataset(np.arange(n_datapoints),
                               train_path,
                               rotations=augmentation)
        test_set = OwnDataset(
            np.arange(n_test), test_path, rotations=False
        )  # Data is augmented in self.testing(rotation=True)

        train_dataloader = DataLoader(dataset=train_set,
                                      batch_size=batch_size_train,
                                      shuffle=True,
                                      **kwargs)
        test_dataloader = DataLoader(dataset=test_set,
                                     batch_size=1,
                                     shuffle=False,
                                     **kwargs)

        best_losses = []

        for epoch in range(epochs):
            mse, rmse = self.training(epoch, train_dataloader, train_set)
            train_mse.append(mse)
            train_rmse.append(rmse)

            mse, rmse = self.testing(epoch, test_dataloader, test_set,
                                     ensemble, remember)
            test_mse.append(mse)
            test_rmse.append(rmse)

            if ensemble:
                if len(best_losses) < remember:
                    best_losses.append(test_mse[-1])
                    torch.save(
                        self.model, result_datapath + 'bestmodels/' +
                        'bestModel' + str(len(best_losses)) + '.pt')
                else:
                    bad_loss = np.argmax(best_losses)
                    if best_losses[bad_loss] > test_mse[-1]:
                        best_losses[bad_loss] = test_mse[-1]
                        torch.save(
                            self.model, result_datapath + 'bestmodels/' +
                            'bestModel' + str(bad_loss) + '.pt')
            else:
                if test_mse[-1] < lowest_loss:
                    lowest_loss = test_mse[-1]
                    torch.save(
                        self.model.state_dict(),
                        result_datapath + 'bestmodels/' + 'bestModel.pt')
                    self.bestModel = self.model
            torch.save(self.model, result_datapath + 'lastModel.pt')
            torch.save(
                {
                    'epoch': epoch,
                    'model_state_dict': self.model.state_dict(),
                    'optimizer_state_dict': self.optimizer.state_dict()
                }, result_datapath + 'lastModel.tar')

        hist = h5py.File(result_datapath + 'history.hdf5', 'w')
        hist.create_dataset('train_mse', data=train_mse)
        hist.create_dataset('test_mse', data=test_mse)
        hist.create_dataset('train_rmse', data=train_rmse)
        hist.create_dataset('test_rmse', data=test_rmse)
        hist.close()

    @staticmethod
    def count_parameters(model):
        total_param = 0
        for name, param in model.named_parameters():
            if param.requires_grad:
                num_param = np.prod(param.size())
                if param.dim() > 1:
                    print(name, ':',
                          'x'.join(str(x) for x in list(param.size())), '=',
                          num_param)
                else:
                    print(name, ':', num_param)
                total_param += num_param
        return total_param