def test_experiment_logs_path_creation_deletion(self): experiment_logs_path = get_experiment_logs_path(self.experiment.unique_name) filepath = get_experiment_logs_path(self.experiment.unique_name) open(filepath, '+w') # Should be true, created by the signal assert os.path.exists(experiment_logs_path) is True assert os.path.exists(filepath) is True delete_experiment_logs(self.experiment.unique_name) assert os.path.exists(filepath) is False
def handle_events_job_logs(experiment_name, experiment_uuid, job_uuid, log_line, task_type=None, task_idx=None): # Must persist resources if logs according to the config if not Experiment.objects.filter(uuid=experiment_uuid).exists(): return logger.debug('handling log event for %s %s', experiment_uuid, job_uuid) if task_type and task_idx: log_line = '{}.{} -- {}'.format(task_type, int(task_idx) + 1, log_line) xp_logger = logging.getLogger(experiment_name) log_path = get_experiment_logs_path(experiment_name) try: log_handler = logging.FileHandler(log_path) log_formatter = logging.Formatter('%(asctime)s %(levelname)s %(message)s') log_handler.setFormatter(log_formatter) xp_logger.addHandler(log_handler) xp_logger.setLevel(logging.INFO) xp_logger.info(log_line) xp_logger.handlers = [] except OSError: # TODO: retry instead? pass
def test_handle_events_job_logs_create_one_handler(self): with patch('runner.tasks.experiments.build_experiment.apply_async' ) as _: # noqa experiment = ExperimentFactory() params = dict(experiment_name=experiment.unique_name, experiment_uuid=experiment.uuid.hex, job_uuid=uuid.uuid4().hex, log_line='First test', task_type=TaskType.MASTER, task_idx=0) handle_events_job_logs(**params) # Check new log path is created log_path = get_experiment_logs_path(experiment.unique_name) assert os.path.exists(log_path) is True # Check the logger has no file handler, and one line created xp_logger = logging.getLogger(experiment.unique_name) assert len(xp_logger.handlers) == 0 # pylint:disable=len-as-condition assert self.file_line_count(log_path) == 1 # pylint:disable=len-as-condition # Calling again the task should not reuse handler, and create a new line handle_events_job_logs(**params) # Check the logger has no file handler, and one line created xp_logger = logging.getLogger(experiment.unique_name) assert len(xp_logger.handlers) == 0 # pylint:disable=len-as-condition assert self.file_line_count(log_path) == 2
def test_experiment_group_logs_path_creation_deletion(self): experiment = ExperimentFactory(user=self.project.user, project=self.project, experiment_group=self.experiment_group) experiment_logs_path = get_experiment_logs_path(experiment.unique_name) open(experiment_logs_path, '+w') experiment_group_logs_path = get_experiment_group_logs_path( self.experiment_group.unique_name) # Should be true, created by the signal assert os.path.exists(experiment_logs_path) is True assert os.path.exists(experiment_group_logs_path) is True delete_experiment_group_logs(self.experiment_group.unique_name) assert os.path.exists(experiment_logs_path) is False assert os.path.exists(experiment_group_logs_path) is False
def test_project_logs_path_creation_deletion(self): with patch('experiments.tasks.build_experiment.apply_async') as _: experiment = ExperimentFactory(user=self.project.user, project=self.project) experiment_logs_path = get_experiment_logs_path(experiment.unique_name) open(experiment_logs_path, '+w') project_logs_path = get_project_logs_path(self.project.unique_name) project_repos_path = get_project_logs_path(self.project.unique_name) # Should be true, created by the signal assert os.path.exists(experiment_logs_path) is True assert os.path.exists(project_logs_path) is True assert os.path.exists(project_repos_path) is True delete_project_logs(self.project.unique_name) assert os.path.exists(experiment_logs_path) is False assert os.path.exists(project_logs_path) is False assert os.path.exists(project_repos_path) is False
def setUp(self): super().setUp() with patch('runner.tasks.experiments.start_experiment.delay') as _: project = ProjectFactory(user=self.auth_client.user) experiment = ExperimentFactory(project=project) self.url = '/{}/{}/{}/experiments/{}/logs'.format( API_V1, project.user.username, project.name, experiment.sequence) log_path = get_experiment_logs_path(experiment.unique_name) fake = Faker() self.logs = [] for _ in range(self.num_log_lines): self.logs.append(fake.sentence()) with open(log_path, 'w') as file: for line in self.logs: file.write(line) file.write('\n')
def get_config_map(namespace, project_name, experiment_group_name, experiment_name, project_uuid, experiment_group_uuid, experiment_uuid, original_name, cloning_strategy, cluster_def, declarations, log_level): name = constants.CONFIG_MAP_NAME.format(experiment_uuid=experiment_uuid) labels = get_map_labels(project_name, experiment_group_name, experiment_name, project_uuid, experiment_group_uuid, experiment_uuid) metadata = client.V1ObjectMeta(name=name, labels=labels, namespace=namespace) experiment_outputs_path = get_experiment_outputs_path( experiment_name=experiment_name, original_name=original_name, cloning_strategy=cloning_strategy) experiment_logs_path = get_experiment_logs_path(experiment_name) experiment_data_path = get_project_data_path(project_name) data = { constants.CONFIG_MAP_CLUSTER_KEY_NAME: json.dumps(cluster_def), constants.CONFIG_MAP_DECLARATIONS_KEY_NAME: json.dumps(declarations) or '{}', constants.CONFIG_MAP_EXPERIMENT_INFO_KEY_NAME: json.dumps(labels), constants.CONFIG_MAP_LOG_LEVEL_KEY_NAME: log_level, constants.CONFIG_MAP_API_KEY_NAME: 'http://{}:{}'.format(settings.POLYAXON_K8S_API_HOST, settings.POLYAXON_K8S_API_PORT), constants.CONFIG_MAP_EXPERIMENT_OUTPUTS_PATH_KEY_NAME: experiment_outputs_path, constants.CONFIG_MAP_EXPERIMENT_LOGS_PATH_KEY_NAME: experiment_logs_path, constants.CONFIG_MAP_EXPERIMENT_DATA_PATH_KEY_NAME: experiment_data_path, } return client.V1ConfigMap(api_version=k8s_constants.K8S_API_VERSION_V1, kind=k8s_constants.K8S_CONFIG_MAP_KIND, metadata=metadata, data=data)
def get(self, request, *args, **kwargs): experiment = self.get_experiment() auditor.record(event_type=EXPERIMENT_LOGS_VIEWED, instance=self.experiment, actor_id=request.user.id) log_path = get_experiment_logs_path(experiment.unique_name) filename = os.path.basename(log_path) chunk_size = 8192 try: wrapped_file = FileWrapper(open(log_path, 'rb'), chunk_size) response = StreamingHttpResponse( wrapped_file, content_type=mimetypes.guess_type(log_path)[0]) response['Content-Length'] = os.path.getsize(log_path) response['Content-Disposition'] = "attachment; filename={}".format( filename) return response except FileNotFoundError: logger.warning('Log file not found: log_path=%s', log_path) return Response( status=status.HTTP_404_NOT_FOUND, data='Log file not found: log_path={}'.format(log_path))