def _Create(self):
        """Create an un-managed yarn cluster."""
        logging.info('Should have created vms by now.')
        logging.info(str(self.vms))

        def InstallSpark(vm):
            vm.Install('spark')

        vm_util.RunThreaded(
            InstallSpark, self.vms['worker_group'] + self.vms['master_group'])
        self.leader = self.vms['master_group'][0]
        spark.ConfigureAndStart(self.leader, self.vms['worker_group'])
Ejemplo n.º 2
0
    def _Create(self):
        """Create an un-managed yarn cluster."""
        logging.info('Should have created vms by now.')
        logging.info(str(self.vms))

        def InstallSpark(vm):
            vm.Install('spark')
            if self.cloud == 'GCP':
                hadoop.InstallGcsConnector(vm)
            if self.cloud == 'AWS':
                hadoop.InstallS3Connector(vm)

        vm_util.RunThreaded(
            InstallSpark, self.vms['worker_group'] + self.vms['master_group'])
        self.leader = self.vms['master_group'][0]
        spark.ConfigureAndStart(self.leader,
                                self.vms['worker_group'],
                                configure_s3=self.cloud == 'AWS')
Ejemplo n.º 3
0
    def _Create(self):
        """Create an un-managed yarn cluster."""
        logging.info('Should have created vms by now.')
        logging.info(str(self.vms))

        def InstallSpark(vm):
            vm.Install('spark')
            if self.cloud == 'GCP':
                hadoop.InstallGcsConnector(vm)

        if 'worker_group' not in self.vms:
            raise errors.Resource.CreationError(
                'UnmanagedDpbSparkCluster requires VMs in a worker_group.')

        vm_util.RunThreaded(
            InstallSpark, self.vms['worker_group'] + self.vms['master_group'])
        self.leader = self.vms['master_group'][0]
        spark.ConfigureAndStart(self.leader,
                                self.vms['worker_group'],
                                configure_s3=self.cloud == 'AWS')