示例#1
0
    def __init__(self, *args, **kwargs):

        super(TestSddrDataset, self).__init__(*args, **kwargs)

        #define distribution
        self.current_distribution = 'Poisson'
        self.family = Family(self.current_distribution)

        #define formulas and network shape
        formulas = {
            'rate':
            '~1 + x1 + x2 + spline(x1, bs="bs",df=9) + spline(x2, bs="bs",df=9)+d1(x1)+d2(x2)'
        }

        self.deep_models_dict = {
            'd1': {
                'model': nn.Sequential(nn.Linear(1, 15)),
                'output_shape': 15
            },
            'd2': {
                'model': nn.Sequential(nn.Linear(1, 3), nn.ReLU(),
                                       nn.Linear(3, 8)),
                'output_shape': 8
            }
        }

        self.train_parameters = {
            'batch_size': 1000,
            'epochs': 2500,
            'degrees_of_freedom': {
                'rate': 4
            }
        }

        # load data
        self.data_path = '../data/test_data/x.csv'
        self.ground_truth_path = '../data/test_data/y.csv'

        self.data = pd.read_csv(self.data_path, sep=None, engine='python')
        self.target = pd.read_csv(self.ground_truth_path)

        self.true_feature_names = ["x1", "x2", "x3", "x4"]
        self.true_x2_11 = np.float32(self.data.x2[11])
        self.true_target_11 = self.target.values[11]

        # perform checks on given distribution name, parameter names and number of formulas given
        self.formulas = checkups(self.family.get_params(), formulas)

        self.prepare_data = PrepareData(
            self.formulas, self.deep_models_dict,
            self.train_parameters['degrees_of_freedom'])
示例#2
0
    def test_patsyfreedummytest_parse_formulas(self):
        """
        Test dummy formulas with only intercept. Pasty is not used here to compute the ground truth.
        """

        # define distribution
        cur_distribution = 'Normal'
        family = Family(cur_distribution)

        # define formulas and network shape
        formulas = dict()
        formulas['loc'] = '~1'
        formulas['scale'] = '~1'

        degrees_of_freedom = {'loc': 4, 'scale': 4}

        deep_models_dict = dict()

        prepare_data = PrepareData(formulas, deep_models_dict,
                                   degrees_of_freedom)
        prepare_data.fit(self.x)
        datadict = prepare_data.transform(self.x)
        dm_info_dict = prepare_data.dm_info_dict
        network_info_dict = prepare_data.network_info_dict
        P = prepare_data.P
        #call parse_formulas
        ground_truth = np.ones([len(self.x), 1])
        ground_truth = torch.from_numpy(ground_truth).float()

        #test if shapes of design matrices and P are as correct
        self.assertTrue((datadict['loc']['structured'] == ground_truth).all())
        self.assertTrue(
            (datadict['loc']['structured'].shape == ground_truth.shape),
            'shape missmatch')
        self.assertEqual(network_info_dict['loc']['struct_shapes'], 1)
        self.assertEqual(P['loc'].shape, (1, 1))
        self.assertEqual(P['loc'], 0)

        self.assertTrue(
            (datadict['scale']['structured'].shape == ground_truth.shape),
            'shape missmatch')
        self.assertTrue(
            (datadict['scale']['structured'] == ground_truth).all())
        self.assertEqual(network_info_dict['scale']['struct_shapes'], 1)
        self.assertEqual(P['scale'].shape, (1, 1))
        self.assertEqual(P['scale'], 0)

        # test if dm_info_dict is correct
        self.assertTrue(
            dm_info_dict['loc']['spline_info']['list_of_spline_slices'] == [])
        self.assertTrue(dm_info_dict['scale']['spline_info']
                        ['list_of_spline_slices'] == [])
        self.assertTrue(dm_info_dict['loc']['spline_info']
                        ['list_of_spline_input_features'] == [])
        self.assertTrue(dm_info_dict['scale']['spline_info']
                        ['list_of_spline_input_features'] == [])
示例#3
0
    def test_spline_penaltymatrix_lambda(self):
        """
        Test if P-matrix is correctly computed and test if regularization parameter lambda is correctly computed from degrees of freedom.
        We test here explicitly if for a smoothing spline
            - the penalty matrix is computed correctly
            - the penalty matrix is regularized correctly
            - the regularization parameter lambda is computed correctly from degrees of freedom
        """

        # define distributions
        cur_distribution = 'Poisson'
        family = Family(cur_distribution)

        # define formulas and network shape
        formulas = dict()
        formulas['rate'] = "~ -1 + spline(x1, bs='bs', df=9, degree=3)"

        degrees_of_freedom = {'rate': 4}

        deep_models_dict = dict()

        # call parse_formulas
        prepare_data = PrepareData(formulas, deep_models_dict,
                                   degrees_of_freedom)
        prepare_data.fit(self.x)
        datadict = prepare_data.transform(self.x)
        dm_info_dict = prepare_data.dm_info_dict
        network_info_dict = prepare_data.network_info_dict

        # get original P and get penalized P (by lambda)
        sp = Spline()
        sp.memorize_chunk(self.x.x1,
                          bs="bs",
                          df=9,
                          degree=3,
                          return_penalty=True)
        P_original = sp.penalty_matrices[0]
        P_penalized = prepare_data.P['rate']

        # calculate regularization parameter lambda
        dm_spline = dmatrix(formulas['rate'], self.x, return_type='dataframe')
        df_lam = df2lambda(dm_spline, P_original, degrees_of_freedom['rate'])

        # calculate lambda from original and penalized P matrix
        lambdas = np.divide(P_penalized,
                            P_original,
                            out=np.zeros_like(P_penalized),
                            where=P_original != 0).flatten()
        lam = np.unique(lambdas[lambdas != 0])

        # test if lambda value, degrees of freedom and penalty matrix are correct
        self.assertAlmostEqual(df_lam[1], 1.00276544, places=4)
        self.assertTrue(lam == df_lam[1])
        self.assertTrue(df_lam[0] == degrees_of_freedom['rate'])
        self.assertTrue((P_original == (P_penalized / df_lam[1])).all())
示例#4
0
    def test_orthogonalization_of_unstructured_part_in_parse_formulas(self):
        """
        Test if parse_formulas is correctly computing the orthogonalization pattern of the unstructured part.
        """

        # define distributions
        cur_distribution = 'Normal'
        family = Family(cur_distribution)

        # define formulas and network shape
        formulas = dict()
        formulas[
            'loc'] = "~1 + x1 + x2 + spline(x3, bs='bs', df=9, degree=3) + d1(x1) + d2(x1,x3)"
        formulas['scale'] = '~1 + x1 + d1(x1)'

        degrees_of_freedom = {'loc': 4, 'scale': 4}

        deep_models_dict = dict()
        deep_models_dict['d1'] = {
            'model': nn.Sequential(nn.Linear(1, 15)),
            'output_shape': 42
        }
        deep_models_dict['d2'] = {
            'model': nn.Sequential(nn.Linear(2, 15)),
            'output_shape': 42
        }

        #call parse_formulas
        prepare_data = PrepareData(formulas, deep_models_dict,
                                   degrees_of_freedom)
        prepare_data.fit(self.x)
        datadict = prepare_data.transform(self.x)
        dm_info_dict = prepare_data.dm_info_dict
        network_info_dict = prepare_data.network_info_dict

        X = dmatrix("~1 + x1 + x2 + spline(x3, bs='bs', df=9, degree=3)",
                    self.x,
                    return_type='dataframe')

        orthogonalization_pattern = network_info_dict['loc'][
            'orthogonalization_pattern']['d1']
        true_column_names = []
        true_column_names.append('Intercept')
        true_column_names.append('x1')
        column_names = set([
            list(X.iloc[:, sl].columns)[-1] for sl in orthogonalization_pattern
        ])
        self.assertTrue(
            len(column_names.symmetric_difference(set(true_column_names))) ==
            0)  #test if column names and true_column_names are identical

        orthogonalization_pattern = network_info_dict['loc'][
            'orthogonalization_pattern']['d2']
        true_column_names = []
        true_column_names.append('Intercept')
        true_column_names.append('x1')
        true_column_names.append("spline(x3, bs='bs', df=9, degree=3)[8]")
        column_names = set([
            list(X.iloc[:, sl].columns)[-1] for sl in orthogonalization_pattern
        ])
        self.assertTrue(
            len(column_names.symmetric_difference(set(true_column_names))) ==
            0)  #test if column names and true_column_names are identical

        X = dmatrix("~1 + x1", self.x, return_type='dataframe')

        orthogonalization_pattern = network_info_dict['loc'][
            'orthogonalization_pattern']['d1']
        true_column_names = []
        true_column_names.append('Intercept')
        true_column_names.append('x1')
        column_names = set([
            list(X.iloc[:, sl].columns)[-1] for sl in orthogonalization_pattern
        ])
        self.assertTrue(
            len(column_names.symmetric_difference(set(true_column_names))) ==
            0)  #test if column names and true_column_names are identical
示例#5
0
    def test_smoothingspline_parse_formulas(self):
        """
        Test if parse_formulas is correctly dealing with smoothingsplines.
        We test here explicitly if parse_formulas deals correctly with
            - a missing intercept
            - reordering of the arguments in the spline functions
            - explicitly adding the return_penalty= False statement
            - having explicit interactions between splines and linear terms
        """

        # define distributions
        cur_distribution = 'Normal'
        family = Family(cur_distribution)

        # define formulas and network shape
        formulas = dict()
        formulas[
            'loc'] = '~-1 + spline(x1,bs="bs",df=4, degree=3):x2 + x1:spline(x2,bs="bs",df=5, degree=3)'
        formulas[
            'scale'] = '~1 + x1 + spline(x1,df=10,return_penalty=False, degree=3,bs="bs")'

        degrees_of_freedom = {'loc': 4, 'scale': [4]}

        deep_models_dict = dict()

        #call parse_formulas
        prepare_data = PrepareData(formulas, deep_models_dict,
                                   degrees_of_freedom)
        prepare_data.fit(self.x)
        datadict = prepare_data.transform(self.x)
        dm_info_dict = prepare_data.dm_info_dict
        network_info_dict = prepare_data.network_info_dict
        P = prepare_data.P

        ground_truth_loc = dmatrix(
            '~-1 + spline(x1,bs="bs",df=4, degree=3):x2 + spline(x2,bs="bs",df=5, degree=3):x1',
            self.x,
            return_type='dataframe').to_numpy()
        ground_truth_scale = dmatrix(
            '~1 + x1 + spline(x1,bs="bs",df=10, degree=3)',
            self.x,
            return_type='dataframe').to_numpy()
        ground_truth_loc = torch.from_numpy(ground_truth_loc).float()
        ground_truth_scale = torch.from_numpy(ground_truth_scale).float()

        #test if shapes of design matrices and P are as correct
        self.assertTrue(
            (datadict['loc']['structured'] == ground_truth_loc).all())
        self.assertTrue(
            (datadict['loc']['structured'].shape == ground_truth_loc.shape),
            'shape missmatch')
        self.assertEqual(network_info_dict['loc']['struct_shapes'], 9)
        self.assertEqual(P['loc'].shape, (9, 9))
        self.assertTrue((P['loc'] == 0).all())

        self.assertFalse((datadict['scale']['structured'] == ground_truth_scale
                          ).all())  # assertFalse is due to orthogonalization
        self.assertTrue((datadict['scale']['structured'].shape
                         == ground_truth_scale.shape), 'shape missmatch')
        self.assertEqual(network_info_dict["scale"]['struct_shapes'], 12)
        self.assertEqual(P['scale'].shape, (12, 12))

        # test if dm_info_dict is correct
        self.assertTrue(
            dm_info_dict['loc']['spline_info']['list_of_spline_slices'] ==
            [slice(0, 4), slice(4, 9)])
        self.assertTrue(dm_info_dict['scale']['spline_info']
                        ['list_of_spline_slices'] == [slice(2, 12)])
        self.assertTrue(
            dm_info_dict['loc']['spline_info']['list_of_spline_input_features']
            == [list({'x1', 'x2'}), list({'x1', 'x2'})])
        self.assertTrue(dm_info_dict['scale']['spline_info']
                        ['list_of_spline_input_features'] == [list({'x1'})])
示例#6
0
    def test_unstructured_parse_formulas(self):
        """
        Test if parse_formulas is correctly dealing with NNs.
        """

        # define distributions
        cur_distribution = 'Normal'
        family = Family(cur_distribution)

        # define formulas and network shape
        formulas = dict()
        formulas['loc'] = '~1 + d1(x2,x1,x3)'
        formulas['scale'] = '~1 + x1 + d2(x1)'

        degrees_of_freedom = {'loc': 4, 'scale': 4}

        deep_models_dict = dict()
        deep_models_dict['d1'] = {
            'model': nn.Sequential(nn.Linear(1, 15)),
            'output_shape': 42
        }
        deep_models_dict['d2'] = {
            'model': nn.Sequential(nn.Linear(1, 15)),
            'output_shape': 42
        }

        #call parse_formulas
        prepare_data = PrepareData(formulas, deep_models_dict,
                                   degrees_of_freedom)
        prepare_data.fit(self.x)
        datadict = prepare_data.transform(self.x)
        dm_info_dict = prepare_data.dm_info_dict
        network_info_dict = prepare_data.network_info_dict
        P = prepare_data.P

        ground_truth_loc = dmatrix('~1', self.x,
                                   return_type='dataframe').to_numpy()
        ground_truth_scale = dmatrix('~1 + x1',
                                     self.x,
                                     return_type='dataframe').to_numpy()
        ground_truth_loc = torch.from_numpy(ground_truth_loc).float()
        ground_truth_scale = torch.from_numpy(ground_truth_scale).float()

        x2x1x3 = self.x[['x2', 'x1', 'x3']]
        x1 = self.x[['x1']]

        #test if shapes of design matrices and P are as correct
        self.assertTrue(
            (datadict['loc']['structured'] == ground_truth_loc).all())
        self.assertTrue(
            (datadict['loc']['structured'].shape == ground_truth_loc.shape),
            'shape missmatch')
        self.assertTrue(((datadict['loc']['d1'].numpy() - x2x1x3.to_numpy()) <
                         0.0001).all())
        self.assertTrue(
            (datadict['loc']['d1'].shape == self.x[['x2', 'x1', 'x3']].shape),
            'shape missmatch for neural network input')
        self.assertEqual(network_info_dict['loc']['struct_shapes'], 1)
        self.assertEqual(P['loc'].shape, (1, 1))
        self.assertTrue((P['loc'] == 0).all())
        self.assertEqual(
            list(network_info_dict['loc']['deep_models_dict'].keys()), ['d1'])
        self.assertEqual(network_info_dict['loc']['deep_models_dict']['d1'],
                         deep_models_dict['d1']['model'])
        self.assertEqual(network_info_dict['loc']['deep_shapes']['d1'],
                         deep_models_dict['d1']['output_shape'])

        self.assertTrue(
            (datadict['scale']['structured'] == ground_truth_scale).all())
        self.assertTrue((datadict['scale']['structured'].shape
                         == ground_truth_scale.shape), 'shape missmatch')
        self.assertTrue(
            ((datadict['scale']['d2'].numpy() - x1.to_numpy()) < 0.0001).all())
        self.assertTrue(
            (datadict['scale']['d2'].shape == self.x[['x1']].shape),
            'shape missmatch for neural network input')
        self.assertEqual(network_info_dict["scale"]['struct_shapes'], 2)
        self.assertEqual(P['scale'].shape, (2, 2))
        self.assertTrue((P['scale'] == 0).all())
        self.assertEqual(
            list(network_info_dict['scale']['deep_models_dict'].keys()),
            ['d2'])
        self.assertEqual(network_info_dict['scale']['deep_models_dict']['d2'],
                         deep_models_dict['d2']['model'])
        self.assertEqual(network_info_dict['scale']['deep_shapes']['d2'],
                         deep_models_dict['d2']['output_shape'])

        # test if dm_info_dict is correct
        self.assertTrue(
            dm_info_dict['loc']['spline_info']['list_of_spline_slices'] == [])
        self.assertTrue(dm_info_dict['scale']['spline_info']
                        ['list_of_spline_slices'] == [])
        self.assertTrue(dm_info_dict['loc']['spline_info']
                        ['list_of_spline_input_features'] == [])
        self.assertTrue(dm_info_dict['scale']['spline_info']
                        ['list_of_spline_input_features'] == [])
示例#7
0
    def test_structured_parse_formulas(self):
        """
        Test if linear model is correctly processed in parse_formulas.
        """

        # define distribution
        cur_distribution = 'Normal'
        family = Family(cur_distribution)

        # define formulas and network shape
        formulas = dict()
        formulas['loc'] = '~1'
        formulas['scale'] = '~1 + x1'

        degrees_of_freedom = {'loc': 4, 'scale': 4}

        deep_models_dict = dict()

        #call parse_formulas
        prepare_data = PrepareData(formulas, deep_models_dict,
                                   degrees_of_freedom)
        prepare_data.fit(self.x)
        datadict = prepare_data.transform(self.x)
        dm_info_dict = prepare_data.dm_info_dict
        network_info_dict = prepare_data.network_info_dict
        P = prepare_data.P

        ground_truth_loc = dmatrix(formulas['loc'],
                                   self.x,
                                   return_type='dataframe').to_numpy()
        ground_truth_scale = dmatrix(formulas['scale'],
                                     self.x,
                                     return_type='dataframe').to_numpy()
        ground_truth_loc = torch.from_numpy(ground_truth_loc).float()
        ground_truth_scale = torch.from_numpy(ground_truth_scale).float()

        #test if shapes of design matrices and P are as correct
        self.assertTrue(
            (datadict['loc']['structured'] == ground_truth_loc).all())
        self.assertTrue(
            (datadict['loc']['structured'].shape == ground_truth_loc.shape),
            'shape missmatch')
        self.assertEqual(network_info_dict['loc']['struct_shapes'], 1)
        self.assertEqual(P['loc'].shape, (1, 1))
        self.assertTrue((P['loc'] == 0).all())

        self.assertTrue((datadict['scale']['structured'].shape
                         == ground_truth_scale.shape), 'shape missmatch')
        self.assertTrue(
            (datadict['scale']['structured'] == ground_truth_scale).all())
        self.assertEqual(network_info_dict['scale']['struct_shapes'], 2)
        self.assertEqual(P['scale'].shape, (2, 2))
        self.assertTrue((P['scale'] == 0).all())

        # test if dm_info_dict is correct
        self.assertTrue(
            dm_info_dict['loc']['spline_info']['list_of_spline_slices'] == [])
        self.assertTrue(dm_info_dict['scale']['spline_info']
                        ['list_of_spline_slices'] == [])
        self.assertTrue(dm_info_dict['loc']['spline_info']
                        ['list_of_spline_input_features'] == [])
        self.assertTrue(dm_info_dict['scale']['spline_info']
                        ['list_of_spline_input_features'] == [])
示例#8
0
class TestSddrDataset(unittest.TestCase):
    '''
    Test SddrDataset for model with a linear part, splines and deep networks using the iris data set. 
    
    It is tested 
        - if get_list_of_feature_names() returns the correct list of feature names of the features from the input dataset.
        - if get_feature(feature_name) returns the correct features (shape and value)
        - if the structured part and the input to the deep network are in datadict are correct (shape and value)
        - if the correct target (shape and value) are returned)
        
    We do not check network_info_dict and dm_info_dict herem as they are tested by Testparse_formulas.
    '''
    def __init__(self, *args, **kwargs):

        super(TestSddrDataset, self).__init__(*args, **kwargs)

        #define distribution
        self.current_distribution = 'Poisson'
        self.family = Family(self.current_distribution)

        #define formulas and network shape
        formulas = {
            'rate':
            '~1 + x1 + x2 + spline(x1, bs="bs",df=9) + spline(x2, bs="bs",df=9)+d1(x1)+d2(x2)'
        }

        self.deep_models_dict = {
            'd1': {
                'model': nn.Sequential(nn.Linear(1, 15)),
                'output_shape': 15
            },
            'd2': {
                'model': nn.Sequential(nn.Linear(1, 3), nn.ReLU(),
                                       nn.Linear(3, 8)),
                'output_shape': 8
            }
        }

        self.train_parameters = {
            'batch_size': 1000,
            'epochs': 2500,
            'degrees_of_freedom': {
                'rate': 4
            }
        }

        # load data
        self.data_path = '../data/test_data/x.csv'
        self.ground_truth_path = '../data/test_data/y.csv'

        self.data = pd.read_csv(self.data_path, sep=None, engine='python')
        self.target = pd.read_csv(self.ground_truth_path)

        self.true_feature_names = ["x1", "x2", "x3", "x4"]
        self.true_x2_11 = np.float32(self.data.x2[11])
        self.true_target_11 = self.target.values[11]

        # perform checks on given distribution name, parameter names and number of formulas given
        self.formulas = checkups(self.family.get_params(), formulas)

        self.prepare_data = PrepareData(
            self.formulas, self.deep_models_dict,
            self.train_parameters['degrees_of_freedom'])

    def test_pandasinput(self):
        """
        Test if SddrDataset correctly works with a pandas dataframe as input.
        """

        # load data
        data = pd.concat([self.data, self.target], axis=1, sort=False)
        dataset = SddrDataset(data, self.prepare_data, "y")

        feature_names = dataset.get_list_of_feature_names()
        feature_test_value = dataset.get_feature('x2')[11]
        linear_input_test_value = dataset[11]["datadict"]["rate"][
            "structured"].numpy()[2]
        deep_input_test_value = dataset[11]["datadict"]["rate"]["d2"].numpy(
        )[0]
        target_test_value = dataset[11]["target"].numpy()

        #test if outputs are equal to the true values in the iris dataset
        self.assertEqual(feature_names, self.true_feature_names)
        self.assertAlmostEqual(feature_test_value, self.true_x2_11, places=4)
        self.assertAlmostEqual(linear_input_test_value,
                               self.true_x2_11,
                               places=4)
        self.assertAlmostEqual(deep_input_test_value,
                               self.true_x2_11,
                               places=4)
        self.assertAlmostEqual(target_test_value,
                               self.true_target_11,
                               places=4)

        # test shapes of outputs
        self.assertEqual(self.true_target_11.shape, target_test_value.shape)
        self.assertEqual(self.true_x2_11.shape, linear_input_test_value.shape)
        self.assertEqual(self.true_x2_11.shape, deep_input_test_value.shape)
        self.assertEqual(self.true_x2_11.shape, feature_test_value.shape)

    def test_pandasinputpandastarget(self):
        """
        Test if SddrDataset correctly works with a pandas dataframe as input and target also given as dataframe.
        """

        # load data
        dataset = SddrDataset(self.data, self.prepare_data, self.target)

        feature_names = dataset.get_list_of_feature_names()
        feature_test_value = dataset.get_feature('x2')[11]
        linear_input_test_value = dataset[11]["datadict"]["rate"][
            "structured"].numpy()[2]
        deep_input_test_value = dataset[11]["datadict"]["rate"]["d2"].numpy(
        )[0]
        target_test_value = dataset[11]["target"].numpy()

        #test if outputs are equal to the true values in the iris dataset
        self.assertEqual(feature_names, self.true_feature_names)
        self.assertAlmostEqual(feature_test_value, self.true_x2_11, places=4)
        self.assertAlmostEqual(linear_input_test_value,
                               self.true_x2_11,
                               places=4)
        self.assertAlmostEqual(deep_input_test_value,
                               self.true_x2_11,
                               places=4)
        self.assertAlmostEqual(target_test_value,
                               self.true_target_11,
                               places=4)

        # test shapes of outputs
        self.assertEqual(self.true_target_11.shape, target_test_value.shape)
        self.assertEqual(self.true_x2_11.shape, linear_input_test_value.shape)
        self.assertEqual(self.true_x2_11.shape, deep_input_test_value.shape)
        self.assertEqual(self.true_x2_11.shape, feature_test_value.shape)

    def test_filepathinput(self):
        """
        Test if SddrDataset correctly works with file paths as inputs.
        """

        # create dataset
        dataset = SddrDataset(self.data_path, self.prepare_data,
                              self.ground_truth_path)

        feature_names = dataset.get_list_of_feature_names()
        feature_test_value = dataset.get_feature('x2')[11]
        linear_input_test_value = dataset[11]["datadict"]["rate"][
            "structured"].numpy()[2]
        deep_input_test_value = dataset[11]["datadict"]["rate"]["d2"].numpy(
        )[0]
        target_test_value = dataset[11]["target"].numpy()

        #test if outputs are equal to the true values in the iris dataset
        self.assertEqual(feature_names, self.true_feature_names)
        self.assertAlmostEqual(feature_test_value, self.true_x2_11, places=4)
        self.assertAlmostEqual(linear_input_test_value,
                               self.true_x2_11,
                               places=4)
        self.assertAlmostEqual(deep_input_test_value,
                               self.true_x2_11,
                               places=4)
        self.assertAlmostEqual(target_test_value,
                               self.true_target_11,
                               places=4)

        # test shapes of outputs
        self.assertEqual(self.true_target_11.shape, target_test_value.shape)
        self.assertEqual(self.true_x2_11.shape, linear_input_test_value.shape)
        self.assertEqual(self.true_x2_11.shape, deep_input_test_value.shape)
        self.assertEqual(self.true_x2_11.shape, feature_test_value.shape)