Example #1
0
    def test_redis_flush_between_instances(self):
        fake_redis = let_fake_redis().evaluate(self)
        fake_redis.set('hello', 'world')

        fake_redis_2 = let_fake_redis().evaluate(self)

        self.assertEqual(0, len(fake_redis_2.keys('*')))
class TestProjectMetrics(Spec):

    mock_redis = let_fake_redis()

    @let
    def consumer(self):
        return ProjectMetrics(self.mock_redis)

    @let
    def job_id(self):
        return self.faker.uuid4()

    @let
    def project_name(self):
        return self.faker.name()

    @let
    def metric_key(self):
        return self.faker.name()

    @let
    def metric_value(self):
        return self.faker.random.randint(1, 230)

    @let
    def random_timestamp(self):
        return self.faker.random.randint(1, 230)

    def test_call_adds_new_metric_to_redis_with_project_name_and_job_id(self):
        from foundations_internal.fast_serializer import deserialize

        message = {
            'key': self.metric_key,
            'value': self.metric_value,
            'job_id': self.job_id,
            'project_name': self.project_name
        }

        self.consumer.call(message, self.random_timestamp, None)
        project_metrics = self.mock_redis.hgetall(
            f'projects:{self.project_name}:metrics')

        serialized_metric = project_metrics[
            f'{self.job_id}:{self.metric_key}'.encode()]
        metric = deserialize(serialized_metric)

        self.assertEqual((self.random_timestamp, self.metric_value), metric)
class TestProjectDescriptionController(Spec):

    mock_redis = let_fake_redis()

    @let
    def project_name(self):
        return self.faker.name()

    @let
    def description(self):
        return self.faker.paragraph()

    @let
    def controller(self):
        return ProjectDescriptionController()

    @set_up
    def set_up(self):
        self.patch('foundations_rest_api.global_state.redis_connection',
                   self.mock_redis)
        self.controller.params = {
            'project_name': self.project_name,
            'project_description': self.description
        }

    def test_get_returns_empty_project_description_when_none_specified(self):
        self.assertEqual({'project_description': ''},
                         self.controller.show().as_json())

    def test_get_returns_project_description_when_specified(self):
        self.mock_redis.set(f'projects:{self.project_name}:description',
                            self.description)
        self.assertEqual({'project_description': self.description},
                         self.controller.show().as_json())

    def test_update_changes_the_description(self):
        self.controller.update().evaluate()
        description = self.mock_redis.get(
            f'projects:{self.project_name}:description').decode()
        self.assertEqual(self.description, description)

    def test_update_returns_a_success_message(self):
        self.controller.update()
        self.assertEqual(f'Updated description for {self.project_name}',
                         self.controller.update().as_json())
Example #4
0
class TestJobTagController(Spec):

    mock_redis = let_fake_redis()
    mock_tag_set_klass = let_patch_mock_with_conditional_return('foundations_events.producers.tag_set.TagSet')
    mock_tag_set = let_mock()
    mock_message_router = let_patch_mock('foundations_contrib.global_state.message_router')

    @let
    def job_id(self):
        return self.faker.uuid4()

    @let
    def key(self):
        return self.faker.word()

    @let
    def value(self):
        return self.faker.sentence()

    @let
    def controller(self):
        return JobTagController()

    @let
    def random_tags(self):
        result = {key: self.faker.sentence() for key in self.faker.words()}
        result[self.key] = self.value
        return result

    @set_up
    def set_up(self):
        self.patch('foundations_rest_api.global_state.redis_connection', self.mock_redis)
        self.controller.params = {'job_id': self.job_id, 'key': self.key}
        self.mock_tag_set_klass.return_when(self.mock_tag_set, self.mock_message_router, self.job_id, self.key, self.value)

    def test_delete_removes_an_existing_tag_from_the_job(self):
        job_annotations_key = f'jobs:{self.job_id}:annotations'
        self.mock_redis.hmset(job_annotations_key, self.random_tags)
        self.controller.delete()
        self.assertNotIn(self.key.encode(), self.mock_redis.hgetall(job_annotations_key))

    def test_delete_returns_a_confirmation_message(self):
        expected_result = f'Tag key: {self.key} deleted from job {self.job_id}'
        self.assertEqual(expected_result, self.controller.delete().as_json())
class TestSyncableDirectory(Spec):

    mock_artifact_file_listing = let_patch_mock_with_conditional_return(
        'foundations_contrib.archiving.upload_artifacts.list_of_files_to_upload_from_artifact_path'
    )
    mock_load_archive = let_patch_mock_with_conditional_return(
        'foundations_contrib.archiving.load_archive')
    mock_mkdtemp = let_patch_mock('tempfile.mkdtemp')
    mock_archive = let_mock()
    mock_redis = let_fake_redis()
    mock_mkdir = let_patch_mock('os.makedirs')
    mock_get_logger = let_patch_mock_with_conditional_return(
        'foundations_contrib.global_state.log_manager.get_logger')
    mock_logger = let_mock()

    @let
    def key(self):
        return self.faker.word()

    @let
    def directory_path(self):
        return self.faker.uri_path()

    @let
    def file_listing_without_directory(self):
        return self.faker.sentences()

    @let
    def file_listing(self):
        return [
            f'{self.directory_path}/{path}'
            for path in self.file_listing_without_directory
        ]

    @let
    def stat_mtime_mapping(self):
        return {
            file_name: self.faker.random.random() * 1000000
            for file_name in self.file_listing_without_directory
        }

    @let
    def stat_mtime_mapping_after_file_change(self):
        import copy

        mapping = copy.deepcopy(self.stat_mtime_mapping)
        key_to_change = list(mapping)[0]
        mapping[key_to_change] += 10

        return mapping

    @let
    def stat_mtime_mapping_after_all_files_change(self):
        import copy

        mapping = copy.deepcopy(self.stat_mtime_mapping)

        for key in mapping:
            mapping[key] += 10

        return mapping

    @let
    def remote_job_id(self):
        return self.faker.uuid4()

    @let
    def local_job_id(self):
        return self.faker.uuid4()

    @let
    def syncable_directory(self):
        self._set_up_for_download()

        from foundations.artifacts.syncable_directory import SyncableDirectory
        syncable_directory = SyncableDirectory(self.key, self.directory_path,
                                               self.local_job_id,
                                               self.remote_job_id)
        self.mock_archive.fetch_file_path_to_target_file_path.reset_mock()
        return syncable_directory

    @let
    def syncable_directory_trailing_slash(self):
        self._set_up_for_download()

        from foundations.artifacts.syncable_directory import SyncableDirectory
        syncable_directory = SyncableDirectory(self.key,
                                               self.directory_path + '/',
                                               self.local_job_id,
                                               self.remote_job_id)
        self.mock_archive.fetch_file_path_to_target_file_path.reset_mock()
        return syncable_directory

    @let
    def syncable_directory_in_uploading_job(self):
        self._set_up_for_download()

        from foundations.artifacts.syncable_directory import SyncableDirectory
        return SyncableDirectory(self.key, self.directory_path,
                                 self.local_job_id, self.local_job_id)

    @set_up
    def set_up(self):
        self.patch('foundations_contrib.global_state.redis_connection',
                   self.mock_redis)
        self.mock_artifact_file_listing.return_when(self.file_listing,
                                                    self.directory_path)
        self.mock_load_archive.return_when(self.mock_archive,
                                           'artifact_archive')
        self.mock_get_logger.return_when(
            self.mock_logger, 'foundations.artifacts.syncable_directory')

        self.mock_os_stat = self.patch('os.stat', ConditionalReturn())
        self.mock_is_file = None

        for file_name, stat_mtime in self.stat_mtime_mapping.items():
            mock_stat = Mock()
            mock_stat.st_mtime = stat_mtime
            path_on_disk = f'{self.directory_path}/{file_name}'
            self.mock_os_stat.return_when(mock_stat, path_on_disk)

    def _set_up_for_download(self):
        if self.mock_is_file is None:
            self.mock_is_file = self.patch('os.path.isfile',
                                           ConditionalReturn())
        else:
            self.mock_is_file.clear()

        for file_name in self.file_listing:
            self.mock_is_file.return_when(
                False, f'{self.directory_path}/{file_name}')

    def test_uploads_all_files(self):
        expected_calls = []
        for file in self.file_listing:
            remote_path = file[len(self.directory_path) + 1:]
            expected_calls.append(
                call(f'synced_directories/{self.key}', file, self.local_job_id,
                     remote_path))
        self.syncable_directory.upload()
        self.mock_archive.append_file.assert_has_calls(expected_calls)

    def test_uploads_all_files_with_trailing_slash(self):
        self.mock_artifact_file_listing.return_when(self.file_listing,
                                                    self.directory_path + '/')

        expected_calls = []
        for file_path in self.file_listing:
            remote_path = file_path[len(self.directory_path) + 1:]
            expected_calls.append(
                call(f'synced_directories/{self.key}', file_path,
                     self.local_job_id, remote_path))
        self.syncable_directory_trailing_slash.upload()
        self.mock_archive.append_file.assert_has_calls(expected_calls)

    def test_uploads_all_files_to_different_package(self):
        from foundations.artifacts.syncable_directory import SyncableDirectory

        expected_calls = []
        for file in self.file_listing:
            remote_path = file[len(self.directory_path) + 1:]
            expected_calls.append(
                call(f'model_package_directories/{self.key}', file,
                     self.local_job_id, remote_path))

        syncable_directory = SyncableDirectory(self.key,
                                               self.directory_path,
                                               self.local_job_id,
                                               self.remote_job_id,
                                               package_name='model_package')
        syncable_directory.upload()
        self.mock_archive.append_file.assert_has_calls(expected_calls)

    def test_upload_should_log_warning_if_no_local_job_set(self):
        from foundations.artifacts.syncable_directory import SyncableDirectory

        self._set_up_for_download()
        directory = SyncableDirectory(self.key, self.directory_path, None,
                                      self.remote_job_id)
        directory.upload()
        self.mock_logger.warning.assert_any_call(
            'local_job_id required for uploading artifacts')

    def test_upload_does_not_upload_if_no_local_job(self):
        from foundations.artifacts.syncable_directory import SyncableDirectory

        self._set_up_for_download()
        directory = SyncableDirectory(self.key, self.directory_path, None,
                                      self.remote_job_id)
        directory.upload()
        self.mock_archive.append_file.assert_not_called()

    def test_tracks_uploaded_files_in_redis(self):
        self.syncable_directory.upload()
        files = self.mock_redis.smembers(
            f'jobs:{self.local_job_id}:synced_artifacts:{self.key}')
        files = [file.decode() for file in files]

        files.sort()
        expected_files = self.file_listing_without_directory.copy()
        expected_files.sort()

        self.assertEqual(expected_files, files)

    def test_tracks_uploaded_file_timestamps_in_redis(self):
        self.syncable_directory.upload()
        timestamp_mapping = self.mock_redis.hgetall(
            f'jobs:{self.local_job_id}:synced_artifacts:{self.key}:timestamps')
        decoded_timestamp_mapping = {
            file.decode(): float(m_time_from_redis)
            for file, m_time_from_redis in timestamp_mapping.items()
        }
        self.assertEqual(self.stat_mtime_mapping, decoded_timestamp_mapping)

    def test_download_all_files(self):
        self._set_up_for_download()

        expected_calls = []
        for file in self.file_listing:
            self.mock_redis.sadd(
                f'jobs:{self.remote_job_id}:synced_artifacts:{self.key}', file)
            download_call = call(f'synced_directories/{self.key}', file,
                                 self.remote_job_id,
                                 f'{self.directory_path}/{file}')
            expected_calls.append(download_call)
        self.syncable_directory.download()
        self.mock_archive.fetch_file_path_to_target_file_path.assert_has_calls(
            expected_calls, any_order=True)

    def test_download_ensures_paths_exist(self):
        self._set_up_for_download()

        import os.path

        expected_calls = []
        for file in self.file_listing:
            self.mock_redis.sadd(
                f'jobs:{self.remote_job_id}:synced_artifacts:{self.key}', file)
            dirname = os.path.dirname(f'{self.directory_path}/{file}')
            mkdir_call = call(dirname, exist_ok=True)
            expected_calls.append(mkdir_call)
        self.syncable_directory.download()
        self.mock_mkdir.assert_has_calls(expected_calls)

    def test_foundations_create_syncable_directory_defaults_to_current_job_for_remote_and_local(
            self):
        from foundations import create_syncable_directory

        self.patch(
            'foundations_internal.pipeline_context.PipelineContext.file_name',
            self.local_job_id)
        instance = self._mock_syncable_directory(self.local_job_id)
        self.assertEqual(
            instance, create_syncable_directory(self.key, self.directory_path))

    def test_foundations_create_syncable_directory_uses_source_job_for_remote(
            self):
        from foundations import create_syncable_directory

        self.patch(
            'foundations_internal.pipeline_context.PipelineContext.file_name',
            self.local_job_id)
        instance = self._mock_syncable_directory(self.remote_job_id)
        self.assertEqual(
            instance,
            create_syncable_directory(self.key, self.directory_path,
                                      self.remote_job_id))

    @skip('Flaky Test')
    def test_foundations_create_syncable_directory_without_any_job_ids(self):
        from foundations import create_syncable_directory

        klass_mock = self.patch(
            'foundations.artifacts.syncable_directory.SyncableDirectory',
            ConditionalReturn())
        instance = Mock()
        klass_mock.return_when(instance, self.key, self.directory_path, None,
                               None)

        self.assertEqual(
            instance, create_syncable_directory(self.key, self.directory_path))

    def test_foundations_create_syncable_directory_with_no_directory_creates_temp_directory(
            self):
        from foundations import create_syncable_directory

        create_syncable_directory(self.key, None, self.remote_job_id)
        self.mock_mkdtemp.assert_called_once()

    def test_path_retuns_correct_synced_directory_path(self):
        self.assertEqual(self.directory_path, str(self.syncable_directory))

    def test_syncable_directory_does_a_download_upon_instantiation(self):
        from foundations import create_syncable_directory

        mock_download = self.patch(
            'foundations.artifacts.syncable_directory.SyncableDirectory.download'
        )
        syncable_directory = create_syncable_directory(self.key, None,
                                                       self.remote_job_id)
        mock_download.assert_called_once()

    def test_syncable_directory_does_not_download_upon_instantiation_if_auto_download_false(
            self):
        mock_download = self.patch(
            'foundations.artifacts.syncable_directory.SyncableDirectory.download'
        )

        from foundations.artifacts.syncable_directory import SyncableDirectory

        SyncableDirectory(self.key,
                          self.directory_path,
                          self.local_job_id,
                          self.remote_job_id,
                          auto_download=False)
        mock_download.assert_not_called()

    def test_upload_only_changed_file(self):
        self.syncable_directory.upload()
        self.mock_os_stat.clear()
        self.mock_archive.append_file.reset_mock()

        for file_name, stat_mtime in self.stat_mtime_mapping_after_file_change.items(
        ):
            mock_stat = Mock()
            mock_stat.st_mtime = stat_mtime
            self.mock_os_stat.return_when(
                mock_stat, f'{self.directory_path}/{file_name}')

        self.syncable_directory.upload()

        remote_path = list(self.stat_mtime_mapping)[0]

        self.mock_archive.append_file.assert_called_once_with(
            f'synced_directories/{self.key}',
            f'{self.directory_path}/{remote_path}', self.local_job_id,
            remote_path)

    def test_upload_file_with_same_name_twice_download_that_file_only_once(
            self):
        self.syncable_directory_in_uploading_job.upload()
        self.syncable_directory_in_uploading_job.upload()

        self.mock_is_file.clear()

        for file in self.file_listing:
            self.mock_is_file.return_when(False, file)

        self.syncable_directory_in_uploading_job.download()
        self.assertEqual(
            len(self.file_listing),
            self.mock_archive.fetch_file_path_to_target_file_path.call_count)

    def test_upload_updated_files_with_same_name_does_not_double_track_those_files(
            self):
        self.syncable_directory_in_uploading_job.upload()

        self.mock_os_stat.clear()

        for file_name, stat_mtime in self.stat_mtime_mapping_after_all_files_change.items(
        ):
            mock_stat = Mock()
            mock_stat.st_mtime = stat_mtime
            self.mock_os_stat.return_when(
                mock_stat, f'{self.directory_path}/{file_name}')

        self.syncable_directory_in_uploading_job.upload()

        self.mock_is_file.clear()

        for file in self.file_listing:
            self.mock_is_file.return_when(False, file)

        self.syncable_directory_in_uploading_job.download()
        self.assertEqual(
            len(self.file_listing),
            self.mock_archive.fetch_file_path_to_target_file_path.call_count)

    def test_download_only_fetches_files_we_do_not_already_have(self):
        self.syncable_directory_in_uploading_job.upload()

        self.mock_is_file.clear()

        file_we_do_not_have = self.file_listing[0]
        self.mock_is_file.return_when(False, file_we_do_not_have)

        for file in self.file_listing[1:]:
            self.mock_is_file.return_when(True, file)

        self.syncable_directory_in_uploading_job.download()

        file_we_do_not_have_in_archive = file_we_do_not_have[
            len(self.directory_path) + 1:]
        self.mock_archive.fetch_file_path_to_target_file_path.assert_called_once_with(
            f'synced_directories/{self.key}', file_we_do_not_have_in_archive,
            self.local_job_id, file_we_do_not_have)

    def test_download_only_fetches_files_that_are_behind_the_ones_in_the_archive(
            self):
        self.syncable_directory_in_uploading_job.upload()

        self.mock_is_file.clear()

        for file in self.file_listing:
            self.mock_is_file.return_when(True, file)

        self.mock_os_stat.clear()

        stat_entries = list(self.stat_mtime_mapping.items())

        old_file, stat_mtime_for_old_file = stat_entries[0]
        old_file_on_disk = f'{self.directory_path}/{old_file}'
        mock_stat = Mock()
        mock_stat.st_mtime = stat_mtime_for_old_file - 100
        self.mock_os_stat.return_when(mock_stat, old_file_on_disk)

        for file_name, stat_mtime in stat_entries[1:]:
            mock_stat = Mock()
            mock_stat.st_mtime = stat_mtime
            path_on_disk = f'{self.directory_path}/{file_name}'
            self.mock_os_stat.return_when(mock_stat, path_on_disk)

        self.syncable_directory_in_uploading_job.download()

        self.mock_archive.fetch_file_path_to_target_file_path.assert_called_once_with(
            f'synced_directories/{self.key}', old_file, self.local_job_id,
            old_file_on_disk)

    def _mock_syncable_directory(self, source_job_id):
        klass_mock = self.patch(
            'foundations.artifacts.syncable_directory.SyncableDirectory',
            ConditionalReturn())
        instance = Mock()
        klass_mock.return_when(instance, self.key, self.directory_path,
                               self.local_job_id, source_job_id)
        return instance
Example #6
0
class TestProjectMetricsController(Spec):

    mock_redis = let_fake_redis()
    mock_tag_set_klass = let_patch_mock_with_conditional_return(
        'foundations_events.producers.tag_set.TagSet')
    mock_tag_set = let_mock()
    mock_message_router = let_patch_mock(
        'foundations_contrib.global_state.message_router')

    @let
    def job_id(self):
        return self.faker.uuid4()

    @let
    def project_name(self):
        return self.faker.name()

    @let
    def controller(self):
        return ProjectMetricsController()

    @let
    def project_metric_logger(self):
        from foundations_events.consumers.project_metrics import ProjectMetrics
        return ProjectMetrics(self.mock_redis)

    @let
    def single_project_metric_logger(self):
        from foundations_events.consumers.single_project_metric import SingleProjectMetric
        return SingleProjectMetric(self.mock_redis)

    @set_up
    def set_up(self):
        self.patch('foundations_rest_api.global_state.redis_connection',
                   self.mock_redis)

    def test_index_returns_timestamp_ordered_metrics(self):
        self.controller.params = {'project_name': self.project_name}
        self._log_metric(33, 'job13', 'metric77', 123.4)

        expected_output = {
            'all_metric_names': ['metric77'],
            'metric_query': [{
                'metric_name': 'metric77',
                'values': [['job13', 123.4]]
            }]
        }

        self.assertEqual(expected_output, self.controller.index().as_json())

    def test_index_returns_timestamp_ordered_metrics_when_metric_name_has_colons(
            self):
        self.controller.params = {'project_name': self.project_name}
        self._log_metric(33, 'job13', 'metric77:007', 123.4)

        expected_output = {
            'all_metric_names': ['metric77:007'],
            'metric_query': [{
                'metric_name': 'metric77:007',
                'values': [['job13', 123.4]]
            }]
        }

        self.assertEqual(expected_output, self.controller.index().as_json())

    def test_index_returns_timestamp_ordered_metrics_different_metrics(self):
        self.controller.params = {'project_name': self.project_name}
        self._log_metric(321, 'job1', 'metric1', 432)
        self._log_metric(123, 'job2', 'metric1', 843)
        self._log_metric(123, 'job1', 'metric2', 221)

        expected_output = {
            'all_metric_names': ['metric1', 'metric2'],
            'metric_query': [{
                'metric_name': 'metric1',
                'values': [['job2', 843], ['job1', 432]]
            }, {
                'metric_name': 'metric2',
                'values': [['job1', 221]]
            }]
        }

        self.assertEqual(expected_output, self.controller.index().as_json())

    def test_index_returns_timestamp_ordered_metrics_metric_filter(self):
        metric_name = self.faker.name()
        self.controller.params = {
            'project_name': self.project_name,
            'metric_name': metric_name
        }
        self._log_metric(321, 'job1', metric_name, 432)
        self._log_metric(123, 'job2', metric_name, 843)
        self._log_metric(123, 'job1', 'metric2', 221)

        expected_output = {
            'all_metric_names': [metric_name, 'metric2'],
            'metric_query': [{
                'metric_name': metric_name,
                'values': [['job2', 843], ['job1', 432]]
            }]
        }

        self.assertEqual(expected_output, self.controller.index().as_json())

    def _log_metric(self, timestamp, job_id, key, value):

        message = {
            'project_name': self.project_name,
            'job_id': job_id,
            'key': key,
            'value': value
        }
        self.project_metric_logger.call(message, timestamp, None)
        self.single_project_metric_logger.call(message, timestamp, None)
class TestProjectNoteListingController(Spec):

    mock_redis = let_fake_redis()

    @let_now
    def auth_client(self):
        constructor = self.patch(
            "foundations_rest_api.utils.AuthenticationClient",
            autospec=True,
        )
        return constructor("conf", "redirect")

    @let
    def message(self):
        return self.faker.sentence()

    @let
    def userid(self):
        return self.faker.word()

    @let
    def username(self):
        return self.faker.word()

    @let
    def controller(self):
        from foundations_rest_api.v2beta.controllers.project_note_listing_controller import (
            ProjectNoteListingController, )
        return ProjectNoteListingController()

    @let
    def project_name(self):
        return self.faker.word()

    @set_up
    def set_up(self):
        self.patch("foundations_rest_api.global_state.redis_connection",
                   self.mock_redis)
        # mock the redis used by authentication
        self.patch("foundations_contrib.global_state.redis_connection",
                   self.mock_redis)
        self.controller.params = {
            "author": self.userid,
            "message": self.message,
            "project_name": self.project_name,
        }

    def test_post_returns_a_confirmation_message(self):
        expected_result = (
            f"Note with author: {self.userid} created with message: {self.message}"
        )
        self.assertEqual(expected_result, self.controller.post().as_json())

    def test_index_returns_empty(self):
        expected_result = []
        self.assertEqual(expected_result, self.controller.index().as_json())

    def test_post_then_index_returns_same_data(self):
        from foundations_rest_api.global_state import app_manager

        headers = {"Authorization": "bearer token"}
        self.auth_client.users_info = ConditionalReturn()
        self.auth_client.users_info.return_when({self.userid: self.username},
                                                "token")

        with app_manager.app().test_request_context(headers=headers):
            self.controller.post()
            index_result = self.controller.index().as_json()
            self.assertEqual(1, len(index_result))
            self.assertEqual(self.message, index_result[0]["message"])
Example #8
0
    def test_is_a_let(self):
        from foundations_spec.helpers import let

        fake_redis = let_fake_redis()
        self.assertTrue(isinstance(fake_redis, let))
Example #9
0
    def test_returns_a_fake_redis(self):
        from fakeredis import FakeRedis

        fake_redis = let_fake_redis().evaluate(self)
        self.assertTrue(isinstance(fake_redis, FakeRedis))