Esempio n. 1
0
def test_experiment_duplicated_rename_mlflow(tmpdir_name):
    with Experiment(tmpdir_name, with_mlflow=True) as e:
        e.log_metric('CV', 0.97)
        run_id_old = e.mlflow_run_id

    with Experiment(tmpdir_name, with_mlflow=True, if_exists='rename') as e:
        e.log_metric('LB', 0.95)
        run_id_new = e.mlflow_run_id

    assert run_id_old != run_id_new
Esempio n. 2
0
def test_experiment_duplicated_error(tmpdir_name):
    with Experiment(tmpdir_name) as e:
        e.log_metric('CV', 0.97)

    with pytest.raises(ValueError):
        with Experiment(tmpdir_name):
            pass

    with pytest.raises(ValueError):
        with Experiment(tmpdir_name, if_exists='error'):
            pass
Esempio n. 3
0
def test_experiment_duplicated_rename_mlflow():
    with get_temp_directory() as logging_dir:
        with Experiment(logging_dir, with_mlflow=True) as e:
            e.log_metric('CV', 0.97)
            run_id_old = e.mlflow_run_id

        with Experiment(logging_dir, with_mlflow=True,
                        if_exists='rename') as e:
            e.log_metric('LB', 0.95)
            run_id_new = e.mlflow_run_id

        assert run_id_old != run_id_new
Esempio n. 4
0
def test_experiment_duplicated_error():
    with get_temp_directory() as logging_dir:
        with Experiment(logging_dir) as e:
            e.log_metric('CV', 0.97)

        with pytest.raises(ValueError):
            with Experiment(logging_dir):
                pass

        with pytest.raises(ValueError):
            with Experiment(logging_dir, if_exists='error'):
                pass
Esempio n. 5
0
def test_experiment_duplicated_append(tmpdir_name):
    with Experiment(tmpdir_name) as e:
        e.log_metric('CV', 0.97)

    with Experiment(tmpdir_name, if_exists='append') as e:
        e.log_metric('LB', 0.95)

    with open(os.path.join(tmpdir_name, 'metrics.json')) as f:
        metrics = json.load(f)

        # appended to the existing result
        assert 'LB' in metrics
        assert 'CV' in metrics
Esempio n. 6
0
def test_experiment_duplicated_replace(tmpdir_name):
    with Experiment(tmpdir_name) as e:
        e.log_metric('CV', 0.97)

    with Experiment(tmpdir_name, if_exists='replace') as e:
        e.log_metric('LB', 0.95)

    with open(os.path.join(tmpdir_name, 'metrics.json')) as f:
        metrics = json.load(f)

        # replaced by the new result
        assert 'LB' in metrics
        assert 'CV' not in metrics
Esempio n. 7
0
def test_experiment_duplicated_append():
    with get_temp_directory() as logging_dir:
        with Experiment(logging_dir) as e:
            e.log_metric('CV', 0.97)

        with Experiment(logging_dir, if_exists='append') as e:
            e.log_metric('LB', 0.95)

        with open(os.path.join(logging_dir, 'metrics.json')) as f:
            metrics = json.load(f)

            # appended to the existing result
            assert 'LB' in metrics
            assert 'CV' in metrics
Esempio n. 8
0
def test_experiment_duplicated_replace():
    with get_temp_directory() as logging_dir:
        with Experiment(logging_dir) as e:
            e.log_metric('CV', 0.97)

        with Experiment(logging_dir, if_exists='replace') as e:
            e.log_metric('LB', 0.95)

        with open(os.path.join(logging_dir, 'metrics.json')) as f:
            metrics = json.load(f)

            # replaced by the new result
            assert 'LB' in metrics
            assert 'CV' not in metrics
Esempio n. 9
0
def test_log_metrics_empty(tmpdir_name):
    with Experiment(tmpdir_name):
        pass

    with open(os.path.join(tmpdir_name, 'metrics.json'), 'r') as f:
        params = json.load(f)
        assert params == {}
Esempio n. 10
0
def test_experiment_duplicated_rename(tmpdir_name):
    with Experiment(tmpdir_name) as e:
        e.log_metric('CV', 0.97)

    with Experiment(tmpdir_name, if_exists='rename') as e:
        e.log_metric('LB', 0.95)

    with open(os.path.join(tmpdir_name, 'metrics.json')) as f:
        metrics = json.load(f)
        assert 'LB' not in metrics
        assert 'CV' in metrics

    with open(os.path.join(tmpdir_name + '_1', 'metrics.json')) as f:
        metrics = json.load(f)
        assert 'LB' in metrics
        assert 'CV' not in metrics
Esempio n. 11
0
def test_log_dict(tmpdir_name):
    with Experiment(tmpdir_name) as e:
        e.log_dict(
            'foo', {
                'a': 1,
                'b': 'foo',
                'c': {
                    'd': 'e',
                    'f': {},
                    'g': {
                        'h': 'i'
                    },
                    'j': None
                }
            })

    with open(os.path.join(tmpdir_name, 'params.json'), 'r') as f:
        params = json.load(f)
        assert params == {
            'foo.a': 1,
            'foo.b': 'foo',
            'foo.c.d': 'e',
            'foo.c.f': '{}',
            'foo.c.g.h': 'i',
            'foo.c.j': 'None'
        }
Esempio n. 12
0
def test_log_metrics_empty():
    with get_temp_directory() as logging_dir:
        with Experiment(logging_dir):
            pass

        with open(os.path.join(logging_dir, 'metrics.json'), 'r') as f:
            params = json.load(f)
            assert params == {}
Esempio n. 13
0
def test_experiment_duplicated_replace_mlflow(tmpdir_name):
    import mlflow

    with Experiment(tmpdir_name, with_mlflow=True) as e:
        e.log_metric('CV', 0.97)
        run_id_old = e.mlflow_run_id

    with Experiment(tmpdir_name, with_mlflow=True, if_exists='replace') as e:
        e.log_metric('LB', 0.95)
        run_id_new = e.mlflow_run_id

    assert run_id_old != run_id_new

    client = mlflow.tracking.MlflowClient()
    old_run = client.get_run(run_id_old)
    new_run = client.get_run(run_id_new)
    assert old_run.info.lifecycle_stage == 'deleted'
    assert new_run.info.lifecycle_stage == 'active'
Esempio n. 14
0
def test_experiment_continue():
    with get_temp_directory() as logging_dir:
        with Experiment(logging_dir, with_mlflow=True) as e:
            e.log_metric('CV', 0.97)

        # appending to exising local & mlflow result
        with Experiment.continue_from(logging_dir, with_mlflow=True) as e:
            e.log_metric('LB', 0.95)

            metric_file = os.path.join(logging_dir, 'metrics.json')

            import mlflow

            client = mlflow.tracking.MlflowClient()
            data = client.get_run(mlflow.active_run().info.run_id).data
            assert data.metrics['CV'] == 0.97
            assert data.metrics['LB'] == 0.95

        with open(metric_file, 'r') as f:
            obj = json.load(f)
            assert obj['CV'] == 0.97
            assert obj['LB'] == 0.95

        with Experiment(logging_dir, with_mlflow=True,
                        if_exists='append') as e:
            e.log_metric('X', 1.1)

            import mlflow

            client = mlflow.tracking.MlflowClient()
            data = client.get_run(mlflow.active_run().info.run_id).data
            assert data.metrics['CV'] == 0.97
            assert data.metrics['LB'] == 0.95
            assert data.metrics['X'] == 1.1

        # stop logging to mlflow, still continue logging on local dir
        with Experiment.continue_from(logging_dir, with_mlflow=False) as e:
            e.log_metric('Y', 1.1)
            import mlflow
            assert mlflow.active_run() is None

        with open(metric_file, 'r') as f:
            obj = json.load(f)
            assert 'Y' in obj
Esempio n. 15
0
def test_log_metrics():
    with get_temp_directory() as logging_dir:
        with Experiment(logging_dir) as e:
            e.log_metric('x', 1)
            e.log_metric('x', 2)

        with open(os.path.join(logging_dir, 'metrics.json'), 'r') as f:
            params = json.load(f)

            expected = {'x': 2}
            assert params == expected
Esempio n. 16
0
def test_custom_experiment(tmpdir_name):
    params = {
        'objective': 'binary',
        'max_depth': 8
    }
    X, y = make_classification_df()

    with Experiment(tmpdir_name, with_mlflow=True) as e:
        run_experiment(params, X, y, logging_directory='foobar', inherit_experiment=e)

    # all files are logged into e.logging_directory, instead of 'foobar'
    _check_file_exists(tmpdir_name, with_mlflow=True)
Esempio n. 17
0
def test_experiment_duplicated_append_mlflow(tmpdir_name):
    with Experiment(tmpdir_name, with_mlflow=True) as e:
        e.log_metric('CV', 0.97)
        run_id_old = e.mlflow_run_id

    with Experiment(tmpdir_name, with_mlflow=True, if_exists='append') as e:
        e.log_metric('LB', 0.95)
        run_id_new = e.mlflow_run_id

    with open(os.path.join(tmpdir_name, 'metrics.json')) as f:
        metrics = json.load(f)

        # appended to the existing result
        assert 'LB' in metrics
        assert 'CV' in metrics

    assert run_id_old == run_id_new

    import mlflow
    client = mlflow.tracking.MlflowClient()
    old_run = client.get_run(run_id_old)
    assert old_run.info.lifecycle_stage == 'active'
Esempio n. 18
0
def test_experiment_continue():
    with get_temp_directory() as logging_dir:
        with Experiment(logging_dir, with_mlflow=True) as e:
            e.log_metric('CV', 0.97)

        # appending to exising local & mlflow result
        with Experiment.continue_from(logging_dir) as e:
            e.log_metric('LB', 0.95)

            metric_file = os.path.join(logging_dir, 'metrics.txt')

            with open(metric_file, 'r') as f:
                lines = [line.split(',') for line in f.readlines()]

                assert lines[0][0] == 'CV'
                assert lines[1][0] == 'LB'

            import mlflow

            client = mlflow.tracking.MlflowClient()
            data = client.get_run(mlflow.active_run().info.run_id).data
            assert data.metrics['CV'] == 0.97
            assert data.metrics['LB'] == 0.95
Esempio n. 19
0
def test_log_metrics(tmpdir_name):
    with Experiment(tmpdir_name) as e:
        e.log_metric('x', 1)
        e.log_metric('x', 2)
        e.log_metrics({
            'y': 3,
            'z': 4,
        })

    with open(os.path.join(tmpdir_name, 'metrics.json'), 'r') as f:
        params = json.load(f)

        expected = {
            'x': 2,
            'y': 3,
            'z': 4,
        }
        assert params == expected
Esempio n. 20
0
def test_log_params(tmpdir_name):
    with Experiment(tmpdir_name) as e:
        e.log_param('x', 1)
        e.log_param('x', 2)
        e.log_params({
            'y': 'ABC',
            'z': None,
        })

    with open(os.path.join(tmpdir_name, 'params.json'), 'r') as f:
        params = json.load(f)

        expected = {
            'x': 2,  # if the key is duplicated, the latter one is stored
            'y': 'ABC',
            'z':
            'None'  # all non-numerical values are casted to string before logging
        }
        assert params == expected
Esempio n. 21
0
def test_error_while_experiment(tmpdir_name):
    try:
        with Experiment(tmpdir_name) as e:
            e.log_metric('x', 0.5)
            e.log_param('foo', 'bar')
            e.log_numpy('np', np.zeros(100))
            e.log_dataframe('df', pd.DataFrame({'a': [1, 2, 3]}))

        raise KeyboardInterrupt()
    except KeyboardInterrupt:
        pass

    # all logs are saved even if error raised inside experiment
    with open(os.path.join(tmpdir_name, 'metrics.json'), 'r') as f:
        metrics = json.load(f)
        assert metrics == {'x': 0.5}

    with open(os.path.join(tmpdir_name, 'params.json'), 'r') as f:
        params = json.load(f)
        assert params == {'foo': 'bar'}

    assert os.path.exists(os.path.join(tmpdir_name, 'np.npy'))
    assert os.path.exists(os.path.join(tmpdir_name, 'df.f'))