Esempio n. 1
0
    def build_workflow(self, split_graph: SplitGraph,
                       project_context: ProjectContext) -> Workflow:
        workflow = Workflow()
        workflow.workflow_config = current_workflow_config()
        workflow.workflow_snapshot_id = '{}.{}.{}'.format(
            project_context.project_name, workflow.workflow_name,
            round(time.time() * 1000))
        # add ai_nodes to workflow
        for sub in split_graph.nodes.values():
            if sub.config.job_type not in self.job_generator_registry.object_dict:
                raise Exception("job generator not support job_type {}".format(
                    sub.config.job_type))
            generator: JobGenerator = self.job_generator_registry \
                .get_object(sub.config.job_type)

            # set job resource dir
            job_resource_dir = os.path.join(
                project_context.get_generated_path(),
                workflow.workflow_snapshot_id, sub.config.job_name)
            if not os.path.exists(job_resource_dir):
                os.makedirs(job_resource_dir)

            job: Job = generator.generate(sub_graph=sub,
                                          resource_dir=job_resource_dir)
            job.resource_dir = job_resource_dir

            # set input output dataset
            for node in sub.nodes.values():
                if isinstance(node, ReadDatasetNode):
                    job.input_dataset_list.append(node.dataset())
                elif isinstance(node, WriteDatasetNode):
                    job.output_dataset_list.append(node.dataset())

            workflow.add_job(job)

        def validate_edge(head, tail):
            if head not in workflow.jobs:
                raise Exception(
                    'job: {} is not defined in workflow!'.format(head))
            if tail is not None and tail != '' and tail != '*' and tail not in workflow.jobs:
                raise Exception(
                    'job: {} is not defined in workflow!'.format(tail))

        # add edges to workflow
        for edges in split_graph.edges.values():
            for e in edges:
                control_edge = copy.deepcopy(e)
                validate_edge(control_edge.destination, control_edge.source)
                workflow.add_edge(control_edge.destination, control_edge)
        return workflow
    def build_workflow(self, split_graph: SplitGraph,
                       project_desc: ProjectDesc) -> Workflow:
        sub_id_to_job_id = {}
        workflow = Workflow()
        workflow.project_desc = project_desc
        # add ai_nodes to workflow
        for sub in split_graph.nodes.values():
            if (sub.config.platform, sub.config.engine
                ) not in self.job_generator_registry.object_dict:
                raise Exception(
                    "job generator not support platform {} engine {}".format(
                        sub.config.platform, sub.config.engine))
            generator: BaseJobGenerator = self.job_generator_registry \
                .get_object((sub.config.platform, sub.config.engine))
            job: BaseJob = generator.generate(sub_graph=sub,
                                              project_desc=project_desc)
            job.job_config.project_desc = project_desc
            if job.job_name is None:
                job.job_name = job.job_config.job_name
            workflow.add_job(job)
            sub_id_to_job_id[sub.instance_id] = job.instance_id
        # add edges to workflow
        for edges in split_graph.edges.values():
            for e in edges:
                control_edge = copy.deepcopy(e)
                control_edge.source_node_id = sub_id_to_job_id[
                    e.source_node_id]
                control_edge.target_node_id = sub_id_to_job_id[
                    e.target_node_id]
                job_edge: JobControlEdge = control_edge_to_job_edge(
                    control_edge=control_edge)
                workflow.add_edge(sub_id_to_job_id[e.source_node_id], job_edge)

        self.register_workflow_execution(workflow)

        for job in workflow.jobs.values():
            job.job_config.project_path = project_desc.project_path
            if job.job_config.exec_mode is None:
                job.job_config.exec_mode = ExecutionMode.BATCH
            generator: BaseJobGenerator = self.job_generator_registry \
                .get_object((job.platform, job.exec_engine))
            generator.generate_job_resource(job)
        return workflow
 def test_workflow_serde(self):
     workflow_config_file = os.path.join(os.path.dirname(__file__), 'workflow_1.yaml')
     workflow_config = load_workflow_config(workflow_config_file)
     workflow = Workflow()
     workflow.workflow_config = workflow_config
     jobs = []
     for job_config in workflow_config.job_configs.values():
         job = Job(job_config=job_config)
         workflow.add_job(job)
         jobs.append(job)
     edge = ControlEdge(destination=jobs[0].job_name,
                        scheduling_rule=SchedulingRule(MeetAnyEventCondition().add_event('a', 'a'),
                                                       JobAction.START))
     workflow.add_edge(jobs[0].job_name, edge)
     edge = ControlEdge(destination=jobs[0].job_name,
                        scheduling_rule=SchedulingRule(MeetAnyEventCondition().add_event('b', 'b'),
                                                       JobAction.START))
     workflow.add_edge(jobs[0].job_name, edge)
     json_text = json_utils.dumps(workflow)
     w: Workflow = json_utils.loads(json_text)
     self.assertEqual(3, len(w.jobs))
     self.assertEqual(2, len(w.edges.get(jobs[0].job_name)))