class CalibrationTest(unittest.TestCase): def setUp(self): self.model = DummyModel() self.criterion = nn.CrossEntropyLoss() self.wrapper = ModelWrapper(self.model, self.criterion) self.optim = torch.optim.SGD(self.wrapper.get_params(), 0.01) self.dataset = DummyDataset() self.calibrator = DirichletCalibrator(self.wrapper, 2, lr=0.001, reg_factor=0.001) def test_calibrated_model(self): assert len(list(self.calibrator.init_model.modules())) < len( list(self.calibrator.calibrated_model.modules())) def test_calibration(self): before_calib_param_init = list( map(lambda x: x.clone(), self.calibrator.init_model.parameters())) before_calib_param = list( map(lambda x: x.clone(), self.calibrator.calibrated_model.parameters())) self.calibrator.calibrate(self.dataset, self.dataset, batch_size=10, epoch=5, use_cuda=False, double_fit=False, workers=0) after_calib_param_init = list( map(lambda x: x.clone(), self.calibrator.init_model.parameters())) after_calib_param = list( map(lambda x: x.clone(), self.calibrator.calibrated_model.parameters())) assert all([ np.allclose(i.detach(), j.detach()) for i, j in zip(before_calib_param_init, after_calib_param_init) ]) assert not all([ np.allclose(i.detach(), j.detach()) for i, j in zip(before_calib_param, after_calib_param) ]) def test_reg_l2_called(self): self.calibrator.l2_reg = Mock(return_value=torch.Tensor([0])) self.calibrator.calibrate(self.dataset, self.dataset, batch_size=10, epoch=5, use_cuda=False, double_fit=False, workers=0) self.calibrator.l2_reg.assert_called()
class CalibrationTest(unittest.TestCase): def setUp(self): self.model = DummyModel() self.criterion = nn.CrossEntropyLoss() self.wrapper = ModelWrapper(self.model, self.criterion) self.optim = torch.optim.SGD(self.wrapper.get_params(), 0.01) self.dataset = DummyDataset() self.calibrator = DirichletCalibrator(self.wrapper, 2, lr=0.001, reg_factor=0.001) def test_calibrated_model(self): # Check that a layer was added. assert len(list(self.wrapper.model.modules())) < len( list(self.calibrator.calibrated_model.modules())) def test_calibration(self): before_calib_param_init = list( map(lambda x: x.clone(), _get_first_module(self.calibrator.wrapper.model).parameters())) before_calib_param = list( map(lambda x: x.clone(), self.calibrator.calibrated_model.parameters())) self.calibrator.calibrate(self.dataset, self.dataset, batch_size=10, epoch=5, use_cuda=False, double_fit=False, workers=0) after_calib_param_init = list( map(lambda x: x.clone(), _get_first_module(self.calibrator.wrapper.model).parameters())) after_calib_param = list( map(lambda x: x.clone(), self.calibrator.calibrated_model.parameters())) assert all([np.allclose(i.detach(), j.detach()) for i, j in zip(before_calib_param_init, after_calib_param_init)]) assert not all([np.allclose(i.detach(), j.detach()) for i, j in zip(before_calib_param, after_calib_param)]) def test_reg_l2_called(self): self.calibrator.l2_reg = Mock(return_value=torch.Tensor([0])) self.calibrator.calibrate(self.dataset, self.dataset, batch_size=10, epoch=5, use_cuda=False, double_fit=False, workers=0) self.calibrator.l2_reg.assert_called() def test_weight_assignment(self): params = list(self.wrapper.model.parameters()) self.wrapper.train_on_dataset(self.dataset, self.optim, 32, 1, False) assert all([k is v for k, v in zip(params, self.optim.param_groups[0]['params'])]) self.calibrator.calibrate(self.dataset, self.dataset, 32, 1, False, True) assert all( [k is v for k, v in zip(self.wrapper.model.parameters(), self.optim.param_groups[0]['params'])]) # Check that we can train the original model before_params = list( map(lambda x: x.clone(), self.wrapper.model.parameters())) self.wrapper.train_on_dataset(self.dataset, self.optim, 10, 2, False) after_params = list( map(lambda x: x.clone(), self.wrapper.model.parameters())) assert not all([np.allclose(i.detach(), j.detach()) for i, j in zip(before_params, after_params)]) # Check that the parameters are still tied. calib_params = list( map(lambda x: x.clone(), _get_first_module(self.calibrator.wrapper.model).parameters())) assert all([np.allclose(i.detach(), j.detach()) for i, j in zip(calib_params, after_params)])
class ModelWrapperMultiOutMultiInTest(unittest.TestCase): def setUp(self): class MultiOutModel(nn.Module): def __init__(self): super().__init__() self.model = DummyModel() def forward(self, x): return [self.model(x)] * 2 self._crit = nn.MSELoss() self.criterion = lambda x, y: self._crit(x[0], y) + self._crit(x[1], y) self.model = MultiOutModel() self.wrapper = ModelWrapper(self.model, self.criterion) self.optim = torch.optim.SGD(self.wrapper.get_params(), 0.01) self.dataset = DummyDataset() def test_train_on_batch(self): self.wrapper.train() old_param = list(map(lambda x: x.clone(), self.model.parameters())) input, target = [ torch.stack(v) for v in zip(*(self.dataset[0], self.dataset[1])) ] self.wrapper.train_on_batch(input, target, self.optim) new_param = list(map(lambda x: x.clone(), self.model.parameters())) assert any( [not torch.allclose(i, j) for i, j in zip(old_param, new_param)]) def test_test_on_batch(self): self.wrapper.eval() input, target = [ torch.stack(v) for v in zip(*(self.dataset[0], self.dataset[1])) ] preds = torch.stack([ self.wrapper.test_on_batch(input, target, cuda=False) for _ in range(10) ]).view(10, -1) # Same loss assert torch.allclose(torch.mean(preds, 0), preds[0]) preds = torch.stack([ self.wrapper.test_on_batch(input, target, cuda=False, average_predictions=10) for _ in range(10) ]).view(10, -1) assert torch.allclose(torch.mean(preds, 0), preds[0]) def test_predict_on_batch(self): self.wrapper.eval() input = torch.stack((self.dataset[0][0], self.dataset[1][0])) # iteration == 1 pred = self.wrapper.predict_on_batch(input, 1, False) assert pred[0].size() == (2, 1, 1) # iterations > 1 pred = self.wrapper.predict_on_batch(input, 10, False) assert pred[0].size() == (2, 1, 10) # iteration == 1 self.wrapper = ModelWrapper(self.model, self.criterion, replicate_in_memory=False) pred = self.wrapper.predict_on_batch(input, 1, False) assert pred[0].size() == (2, 1, 1) # iterations > 1 pred = self.wrapper.predict_on_batch(input, 10, False) assert pred[0].size() == (2, 1, 10) def test_out_of_mem_raises_error(self): self.wrapper.eval() input = torch.stack((self.dataset[0][0], self.dataset[1][0])) with pytest.raises(RuntimeError) as e_info: self.wrapper.predict_on_batch(input, 0, False) assert 'CUDA ran out of memory while BaaL tried to replicate data' in str( e_info.value) def test_raising_type_errors(self): iterations = math.inf self.wrapper.eval() input = torch.stack((self.dataset[0][0], self.dataset[1][0])) with pytest.raises(TypeError): self.wrapper.predict_on_batch(input, iterations, False) def test_using_cuda_raises_error_while_testing(self): '''CUDA is not available on test environment''' self.wrapper.eval() input = torch.stack((self.dataset[0][0], self.dataset[1][0])) with pytest.raises(Exception): self.wrapper.predict_on_batch(input, 1, True) def test_train(self): history = self.wrapper.train_on_dataset(self.dataset, self.optim, 10, 2, use_cuda=False, workers=0) assert len(history) == 2 def test_test(self): l = self.wrapper.test_on_dataset(self.dataset, 10, use_cuda=False, workers=0) assert np.isfinite(l) l = self.wrapper.test_on_dataset(self.dataset, 10, use_cuda=False, workers=0, average_predictions=10) assert np.isfinite(l) def test_predict(self): l = self.wrapper.predict_on_dataset(self.dataset, 10, 20, use_cuda=False, workers=0) self.wrapper.eval() assert np.allclose( self.wrapper.predict_on_batch(self.dataset[0][0].unsqueeze(0), 20)[0].detach().numpy(), l[0][0]) assert np.allclose( self.wrapper.predict_on_batch(self.dataset[19][0].unsqueeze(0), 20)[0][0].detach().numpy(), l[0][19]) assert l[0].shape == (len(self.dataset), 1, 20) # Test generators l_gen = self.wrapper.predict_on_dataset_generator(self.dataset, 10, 20, use_cuda=False, workers=0) assert np.allclose(next(l_gen)[0][0], l[0][0]) for last in l_gen: pass # Get last item assert np.allclose(last[0][-1], l[0][-1]) # Test Half l_gen = self.wrapper.predict_on_dataset_generator(self.dataset, 10, 20, use_cuda=False, workers=0, half=True) l = self.wrapper.predict_on_dataset(self.dataset, 10, 20, use_cuda=False, workers=0, half=True) assert next(l_gen)[0].dtype == np.float16 assert l[0].dtype == np.float16 data_s = [] l_gen = self.wrapper.predict_on_dataset_generator(data_s, 10, 20, use_cuda=False, workers=0, half=True) assert len(list(l_gen)) == 0
class ModelWrapperTest(unittest.TestCase): def setUp(self): # self.model = nn.Sequential( # nn.Linear(10, 8), nn.ReLU(), nn.Dropout(), nn.Linear(8, 1), nn.Sigmoid() # ) self.model = DummyModel() self.criterion = nn.BCEWithLogitsLoss() self.wrapper = ModelWrapper(self.model, self.criterion) self.optim = torch.optim.SGD(self.wrapper.get_params(), 0.01) self.dataset = DummyDataset() def test_train_on_batch(self): self.wrapper.train() old_param = list(map(lambda x: x.clone(), self.model.parameters())) input, target = torch.randn([1, 3, 10, 10]), torch.randn(1, 1) self.wrapper.train_on_batch(input, target, self.optim) new_param = list(map(lambda x: x.clone(), self.model.parameters())) assert any( [not torch.allclose(i, j) for i, j in zip(old_param, new_param)]) # test reset weights properties linear_weights = list( self.wrapper.model.named_children())[3][1].weight.clone() conv_weights = list( self.wrapper.model.named_children())[0][1].weight.clone() self.wrapper.reset_fcs() linear_new_weights = list( self.wrapper.model.named_children())[3][1].weight.clone() conv_new_weights = list( self.wrapper.model.named_children())[0][1].weight.clone() assert all([ not torch.allclose(i, j) for i, j in zip(linear_new_weights, linear_weights) ]) assert all([ torch.allclose(i, j) for i, j in zip(conv_new_weights, conv_weights) ]) self.wrapper.reset_all() conv_next_new_weights = list( self.wrapper.model.named_children())[0][1].weight.clone() assert all([ not torch.allclose(i, j) for i, j in zip(conv_new_weights, conv_next_new_weights) ]) def test_test_on_batch(self): self.wrapper.eval() input, target = torch.randn([1, 3, 10, 10]), torch.randn(1, 1) preds = torch.stack([ self.wrapper.test_on_batch(input, target, cuda=False) for _ in range(10) ]).view(10, -1) # Same loss assert torch.allclose(torch.mean(preds, 0), preds[0]) preds = torch.stack([ self.wrapper.test_on_batch(input, target, cuda=False, average_predictions=10) for _ in range(10) ]).view(10, -1) assert torch.allclose(torch.mean(preds, 0), preds[0]) def test_predict_on_batch(self): self.wrapper.eval() input = torch.randn([2, 3, 10, 10]) # iteration == 1 pred = self.wrapper.predict_on_batch(input, 1, False) assert pred.size() == (2, 1, 1) # iterations > 1 pred = self.wrapper.predict_on_batch(input, 10, False) assert pred.size() == (2, 1, 10) # iteration == 1 self.wrapper = ModelWrapper(self.model, self.criterion, replicate_in_memory=False) pred = self.wrapper.predict_on_batch(input, 1, False) assert pred.size() == (2, 1, 1) # iterations > 1 pred = self.wrapper.predict_on_batch(input, 10, False) assert pred.size() == (2, 1, 10) def test_train(self): history = self.wrapper.train_on_dataset(self.dataset, self.optim, 10, 2, use_cuda=False, workers=0) assert len(history) == 2 def test_test(self): l = self.wrapper.test_on_dataset(self.dataset, 10, use_cuda=False, workers=0) assert np.isfinite(l) l = self.wrapper.test_on_dataset(self.dataset, 10, use_cuda=False, workers=0, average_predictions=10) assert np.isfinite(l) def test_predict(self): l = self.wrapper.predict_on_dataset(self.dataset, 10, 20, use_cuda=False, workers=0) self.wrapper.eval() assert np.allclose( self.wrapper.predict_on_batch(self.dataset[0][0].unsqueeze(0), 20)[0].detach().numpy(), l[0]) assert np.allclose( self.wrapper.predict_on_batch(self.dataset[19][0].unsqueeze(0), 20)[0].detach().numpy(), l[19]) assert l.shape == (len(self.dataset), 1, 20) # Test generators l_gen = self.wrapper.predict_on_dataset_generator(self.dataset, 10, 20, use_cuda=False, workers=0) assert np.allclose(next(l_gen)[0], l[0]) for last in l_gen: pass # Get last item assert np.allclose(last[-1], l[-1]) # Test Half l_gen = self.wrapper.predict_on_dataset_generator(self.dataset, 10, 20, use_cuda=False, workers=0, half=True) l = self.wrapper.predict_on_dataset(self.dataset, 10, 20, use_cuda=False, workers=0, half=True) assert next(l_gen).dtype == np.float16 assert l.dtype == np.float16 def test_states(self): input = torch.randn([1, 3, 10, 10]) def pred_with_dropout(replicate_in_memory): self.wrapper = ModelWrapper( self.model, self.criterion, replicate_in_memory=replicate_in_memory) self.wrapper.train() # Dropout make the pred changes preds = torch.stack([ self.wrapper.predict_on_batch(input, iterations=1, cuda=False) for _ in range(10) ]).view(10, -1) assert not torch.allclose(torch.mean(preds, 0), preds[0]) pred_with_dropout(replicate_in_memory=True) pred_with_dropout(replicate_in_memory=False) def pred_without_dropout(replicate_in_memory): self.wrapper = ModelWrapper( self.model, self.criterion, replicate_in_memory=replicate_in_memory) # Dropout is not active in eval self.wrapper.eval() preds = torch.stack([ self.wrapper.predict_on_batch(input, iterations=1, cuda=False) for _ in range(10) ]).view(10, -1) assert torch.allclose(torch.mean(preds, 0), preds[0]) pred_without_dropout(replicate_in_memory=True) pred_without_dropout(replicate_in_memory=False) def test_add_metric(self): self.wrapper.add_metric('cls_report', lambda: ClassificationReport(2)) assert 'test_cls_report' in self.wrapper.metrics assert 'train_cls_report' in self.wrapper.metrics self.wrapper.train_on_dataset(self.dataset, self.optim, 32, 2, False) self.wrapper.test_on_dataset(self.dataset, 32, False) assert (self.wrapper.metrics['train_cls_report'].value['accuracy'] != 0).any() assert (self.wrapper.metrics['test_cls_report'].value['accuracy'] != 0).any() def test_train_and_test(self): res = self.wrapper.train_and_test_on_datasets( self.dataset, self.dataset, self.optim, 32, 5, False, return_best_weights=False) assert len(res) == 5 res = self.wrapper.train_and_test_on_datasets(self.dataset, self.dataset, self.optim, 32, 5, False, return_best_weights=True) assert len(res) == 2 assert len(res[0]) == 5 assert isinstance(res[1], dict) mock = Mock() mock.side_effect = (((np.linspace(0, 50) - 10) / 10)**2).tolist() self.wrapper.test_on_dataset = mock res = self.wrapper.train_and_test_on_datasets(self.dataset, self.dataset, self.optim, 32, 50, False, return_best_weights=True, patience=1) assert len(res) == 2 assert len(res[0]) < 50 mock = Mock() mock.side_effect = (((np.linspace(0, 50) - 10) / 10)**2).tolist() self.wrapper.test_on_dataset = mock res = self.wrapper.train_and_test_on_datasets(self.dataset, self.dataset, self.optim, 32, 50, False, return_best_weights=True, patience=1, min_epoch_for_es=20) assert len(res) == 2 assert len(res[0]) < 50 and len(res[0]) > 20
class ModelWrapperMultiOutTest(unittest.TestCase): def setUp(self): class MultiOutModel(nn.Module): def __init__(self): super().__init__() self.model = DummyModel() def forward(self, x): return [self.model(x)] * 2 self._crit = nn.MSELoss() self.criterion = lambda x, y: self._crit(x[0], y) + self._crit(x[1], y) self.model = MultiOutModel() self.wrapper = ModelWrapper(self.model, self.criterion) self.optim = torch.optim.SGD(self.wrapper.get_params(), 0.01) self.dataset = DummyDataset() def test_train_on_batch(self): self.wrapper.train() old_param = list(map(lambda x: x.clone(), self.model.parameters())) input, target = [torch.stack(v) for v in zip(*(self.dataset[0], self.dataset[1]))] self.wrapper.train_on_batch(input, target, self.optim) new_param = list(map(lambda x: x.clone(), self.model.parameters())) assert any([not torch.allclose(i, j) for i, j in zip(old_param, new_param)]) def test_test_on_batch(self): self.wrapper.eval() input, target = [torch.stack(v) for v in zip(*(self.dataset[0], self.dataset[1]))] preds = torch.stack( [self.wrapper.test_on_batch(input, target, cuda=False) for _ in range(10)] ).view(10, -1) # Same loss assert torch.allclose(torch.mean(preds, 0), preds[0]) preds = torch.stack( [ self.wrapper.test_on_batch( input, target, cuda=False, average_predictions=10 ) for _ in range(10) ] ).view(10, -1) assert torch.allclose(torch.mean(preds, 0), preds[0]) def test_predict_on_batch(self): self.wrapper.eval() input = torch.stack((self.dataset[0][0], self.dataset[1][0])) # iteration == 1 pred = self.wrapper.predict_on_batch(input, 1, False) assert pred[0].size() == (2, 1, 1) # iterations > 1 pred = self.wrapper.predict_on_batch(input, 10, False) assert pred[0].size() == (2, 1, 10) def test_train(self): history = self.wrapper.train_on_dataset(self.dataset, self.optim, 10, 2, use_cuda=False, workers=0) assert len(history) == 2 def test_test(self): l = self.wrapper.test_on_dataset(self.dataset, 10, use_cuda=False, workers=0) assert np.isfinite(l) l = self.wrapper.test_on_dataset( self.dataset, 10, use_cuda=False, workers=0, average_predictions=10 ) assert np.isfinite(l) def test_predict(self): l = self.wrapper.predict_on_dataset(self.dataset, 10, 20, use_cuda=False, workers=0) self.wrapper.eval() assert np.allclose( self.wrapper.predict_on_batch(self.dataset[0][0].unsqueeze(0), 20)[0].detach().numpy(), l[0][0]) assert np.allclose( self.wrapper.predict_on_batch(self.dataset[19][0].unsqueeze(0), 20)[0][ 0].detach().numpy(), l[0][19]) assert l[0].shape == (len(self.dataset), 1, 20) # Test generators l_gen = self.wrapper.predict_on_dataset_generator(self.dataset, 10, 20, use_cuda=False, workers=0) assert np.allclose(next(l_gen)[0][0], l[0][0]) for last in l_gen: pass # Get last item assert np.allclose(last[0][-1], l[0][-1]) # Test Half l_gen = self.wrapper.predict_on_dataset_generator(self.dataset, 10, 20, use_cuda=False, workers=0, half=True) l = self.wrapper.predict_on_dataset(self.dataset, 10, 20, use_cuda=False, workers=0, half=True) assert next(l_gen)[0].dtype == np.float16 assert l[0].dtype == np.float16