Exemplo n.º 1
0
    def __init__(self, id, tag, work_dir, max_work_dir_size_bytes,
                 max_images_bytes, shared_file_system, slots, bundle_service,
                 docker):
        self.id = id
        self._tag = tag
        self.shared_file_system = shared_file_system
        self._bundle_service = bundle_service
        self._docker = docker
        self._slots = slots

        self._worker_state_manager = WorkerStateManager(
            work_dir, self.shared_file_system)

        if not self.shared_file_system:
            # Manages which dependencies are available.
            self._dependency_manager = DependencyManager(
                work_dir, max_work_dir_size_bytes,
                self._worker_state_manager.previous_runs.keys())
        self._image_manager = DockerImageManager(self._docker, work_dir,
                                                 max_images_bytes)
        self._max_images_bytes = max_images_bytes

        self._exiting_lock = threading.Lock()
        self._exiting = False
        self._should_upgrade = False
        self._last_checkin_successful = False
Exemplo n.º 2
0
    def __init__(self, id, tag, work_dir, max_work_dir_size_bytes,
                 max_images_bytes, shared_file_system, slots, bundle_service,
                 docker):
        self.id = id
        self._tag = tag
        self.shared_file_system = shared_file_system
        self._bundle_service = bundle_service
        self._docker = docker
        self._slots = slots

        if not self.shared_file_system:
            # Manages which dependencies are available.
            self._dependency_manager = DependencyManager(
                work_dir, max_work_dir_size_bytes)
        self._image_manager = DockerImageManager(self._docker, work_dir,
                                                 max_images_bytes)
        self._max_images_bytes = max_images_bytes

        # Dictionary from UUID to Run that keeps track of bundles currently
        # running. These runs are added to this dict inside _run, and removed
        # when the Run class calls finish_run.
        self._runs_lock = threading.Lock()
        self._runs = {}

        self._exiting_lock = threading.Lock()
        self._exiting = False
        self._should_upgrade = False
        self._last_checkin_successful = False
Exemplo n.º 3
0
    def __init__(self, worker_id, tag, work_dir,
                 max_work_dir_size_bytes, max_dependencies_serialized_length,
                 shared_file_system, bundle_service, create_run_manager):
        self.id = worker_id
        self._tag = tag
        self.shared_file_system = shared_file_system
        self._bundle_service = bundle_service
        self._resource_lock = threading.Lock() # lock for cpuset and gpuset
        self._run_manager = create_run_manager(self)

        self._worker_state_manager = WorkerStateManager(
            work_dir=work_dir,
            run_manager=self._run_manager,
            shared_file_system=self.shared_file_system
        )

        if not self.shared_file_system:
            # Manages which dependencies are available.
            self._dependency_manager = DependencyManager(
                work_dir,
                max_work_dir_size_bytes,
                max_dependencies_serialized_length,
                self._worker_state_manager.previous_runs.keys()
            )

        self._exiting_lock = threading.Lock()
        self._exiting = False
        self._should_upgrade = False
        self._last_checkin_successful = False
Exemplo n.º 4
0
class Worker(object):
    """
    This class is responsible for:

        1) Registering with the bundle service and receiving all messages
           sent to the worker.
        2) Managing all the runs currently executing on the worker and
           forwarding messages associated with those runs to the appropriate
           instance of the Run class.
        3) Spawning classes and threads that manage other worker resources,
           specifically the storage of bundles (both running bundles as well as
           their dependencies) and the cache of Docker images.
        4) Upgrading the worker.
    """
    def __init__(self, id, tag, work_dir, max_work_dir_size_bytes,
                 max_images_bytes, shared_file_system, slots, bundle_service,
                 docker):
        self.id = id
        self._tag = tag
        self.shared_file_system = shared_file_system
        self._bundle_service = bundle_service
        self._docker = docker
        self._slots = slots

        if not self.shared_file_system:
            # Manages which dependencies are available.
            self._dependency_manager = DependencyManager(
                work_dir, max_work_dir_size_bytes)
        self._image_manager = DockerImageManager(self._docker, work_dir,
                                                 max_images_bytes)
        self._max_images_bytes = max_images_bytes

        # Dictionary from UUID to Run that keeps track of bundles currently
        # running. These runs are added to this dict inside _run, and removed
        # when the Run class calls finish_run.
        self._runs_lock = threading.Lock()
        self._runs = {}

        self._exiting_lock = threading.Lock()
        self._exiting = False
        self._should_upgrade = False
        self._last_checkin_successful = False

    def run(self):
        if self._max_images_bytes is not None:
            self._image_manager.start_cleanup_thread()
        if not self.shared_file_system:
            self._dependency_manager.start_cleanup_thread()

        while self._should_run():
            try:
                self._checkin()
                if not self._last_checkin_successful:
                    logger.info('Connected! Successful check in.')
                self._last_checkin_successful = True

            except Exception:
                self._last_checkin_successful = False
                traceback.print_exc()
                time.sleep(1)

        self._checkout()

        if self._max_images_bytes is not None:
            self._image_manager.stop_cleanup_thread()
        if not self.shared_file_system:
            self._dependency_manager.stop_cleanup_thread()

        if self._should_upgrade:
            self._upgrade()

    def signal(self):
        logger.info(
            'Exiting: Will wait for exiting jobs to finish, but will not start any new jobs.'
        )
        with self._exiting_lock:
            self._exiting = True

    def _is_exiting(self):
        with self._exiting_lock:
            return self._exiting

    def _should_run(self):
        if not self._is_exiting():
            return True
        with self._runs_lock:
            if self._runs:
                return True
        return False

    def _get_memory_bytes(self):
        try:
            return os.sysconf('SC_PAGE_SIZE') * os.sysconf('SC_PHYS_PAGES')
        except ValueError:
            # Fallback to sysctl when os.sysconf('SC_PHYS_PAGES') fails on OS X
            return int(check_output(['sysctl', '-n', 'hw.memsize']).strip())

    def _get_gpu_count(self):
        if not self._docker._use_nvidia_docker:
            return 0

        container_id = self._docker.run_nvidia_smi('-L',
                                                   'nvidia/cuda:8.0-runtime')
        out, err = self._docker.get_logs(container_id)
        count = len(re.findall('^GPU \d', out))
        self._docker.delete_container(container_id)
        return count

    def _checkin(self):
        request = {
            'version':
            VERSION,
            'will_upgrade':
            self._should_upgrade,
            'tag':
            self._tag,
            'slots':
            self._slots if not self._is_exiting() else 0,
            'cpus':
            multiprocessing.cpu_count(),
            'gpus':
            self._get_gpu_count(),
            'memory_bytes':
            self._get_memory_bytes(),
            'dependencies': [] if self.shared_file_system else
            self._dependency_manager.dependencies()
        }
        response = self._bundle_service.checkin(self.id, request)
        if response:
            type = response['type']
            logger.debug('Received %s message: %s', type, response)
            if type == 'run':
                self._run(response['bundle'], response['resources'])
            elif type == 'read':
                self._read(response['socket_id'], response['uuid'],
                           response['path'], response['read_args'])
            elif type == 'write':
                self._write(response['uuid'], response['subpath'],
                            response['string'])
            elif type == 'kill':
                self._kill(response['uuid'])
            elif type == 'upgrade':
                with self._exiting_lock:
                    self._exiting = True
                self._should_upgrade = True

    def _run(self, bundle, resources):
        if self.shared_file_system:
            bundle_path = bundle['location']
        else:
            bundle_path = self._dependency_manager.get_run_path(bundle['uuid'])
        run = Run(self._bundle_service, self._docker, self._image_manager,
                  self, bundle, bundle_path, resources)
        if run.run():
            with self._runs_lock:
                self._runs[bundle['uuid']] = run

    def add_dependency(self, parent_uuid, parent_path, uuid, loop_callback):
        """
        Registers that the run with UUID uuid depends on path parent_path in
        bundle with UUID parent_uuid. Downloads the dependency if necessary, and
        returns the path to the dependency. Note, remove_dependency should be
        called for every dependency added.

        loop_callback is a method that is called repeatedly while downloading
        the dependency. If that method throws an exception, the download gets
        interrupted and add_dependency fails with that same exception.
        """
        assert (not self.shared_file_system)
        dependency_path, should_download = (
            self._dependency_manager.add_dependency(parent_uuid, parent_path,
                                                    uuid))
        if should_download:
            logger.debug('Downloading dependency %s/%s', parent_uuid,
                         parent_path)
            try:
                download_success = False
                fileobj, target_type = (
                    self._bundle_service.get_bundle_contents(
                        parent_uuid, parent_path))
                with closing(fileobj):
                    # "Bug" the fileobj's read function so that we can keep
                    # track of the number of bytes downloaded so far.
                    old_read_method = fileobj.read
                    bytes_downloaded = [0]

                    def interruptable_read(*args, **kwargs):
                        data = old_read_method(*args, **kwargs)
                        bytes_downloaded[0] += len(data)
                        loop_callback(bytes_downloaded[0])
                        return data

                    fileobj.read = interruptable_read

                    self._store_dependency(dependency_path, fileobj,
                                           target_type)
                    download_success = True
            finally:
                logger.debug('Finished downloading dependency %s/%s',
                             parent_uuid, parent_path)
                self._dependency_manager.finish_download(
                    parent_uuid, parent_path, download_success)

        return dependency_path

    def _store_dependency(self, dependency_path, fileobj, target_type):
        try:
            if target_type == 'directory':
                un_tar_directory(fileobj, dependency_path, 'gz')
            else:
                with open(dependency_path, 'wb') as f:
                    shutil.copyfileobj(fileobj, f)
        except:
            remove_path(dependency_path)
            raise

    def remove_dependency(self, parent_uuid, parent_path, uuid):
        """
        Unregisters that the run with UUID uuid depends on path parent_path in
        bundle with UUID parent_uuid. This method is safe to call on
        dependencies that were never added with add_dependency.
        """
        assert (not self.shared_file_system)
        self._dependency_manager.remove_dependency(parent_uuid, parent_path,
                                                   uuid)

    def _read(self, socket_id, uuid, path, read_args):
        run = self._get_run(uuid)
        if run is None:
            Run.read_run_missing(self._bundle_service, self, socket_id)
        else:
            # Reads may take a long time, so do the read in a separate thread.
            threading.Thread(target=Run.read,
                             args=(run, socket_id, path, read_args)).start()

    def _write(self, uuid, subpath, string):
        run = self._get_run(uuid)
        if run is not None:
            run.write(subpath, string)

    def _kill(self, uuid):
        run = self._get_run(uuid)
        if run is not None:
            run.kill('Kill requested')

    def _get_run(self, uuid):
        with self._runs_lock:
            return self._runs.get(uuid)

    def finish_run(self, uuid):
        """
        Registers that the run with the given UUID has finished.
        """
        with self._runs_lock:
            del self._runs[uuid]
        if not self.shared_file_system:
            self._dependency_manager.finish_run(uuid)

    def _checkout(self):
        try:
            self._bundle_service.checkout(self.id)
        except BundleServiceException:
            traceback.print_exc()

    def _upgrade(self):
        logger.debug('Upgrading')
        worker_dir = os.path.dirname(os.path.realpath(__file__))

        while True:
            try:
                with closing(self._bundle_service.get_code()) as code:
                    remove_path(worker_dir)
                    un_tar_directory(code, worker_dir, 'gz')
                    break
            except Exception:
                traceback.print_exc()
                time.sleep(1)

        exit(123)
Exemplo n.º 5
0
    def __init__(self,
                 id,
                 tag,
                 work_dir,
                 cpuset,
                 gpuset,
                 max_work_dir_size_bytes,
                 max_dependencies_serialized_length,
                 max_images_bytes,
                 shared_file_system,
                 bundle_service,
                 docker,
                 docker_network_prefix='codalab_worker_network'):
        self.id = id
        self._tag = tag
        self.shared_file_system = shared_file_system
        self._bundle_service = bundle_service
        self._docker = docker
        self._docker_network_prefix = docker_network_prefix

        self._resource_lock = threading.Lock()  # lock for cpuset and gpuset
        self._cpuset = cpuset
        self._gpuset = gpuset
        self._cpuset_free = set(
            self._cpuset)  # make a copy of self._cpuset as initial value
        self._gpuset_free = set(
            self._gpuset)  # make a copy of self._gpuset as initial value

        self._worker_state_manager = WorkerStateManager(
            work_dir, self.shared_file_system)

        if not self.shared_file_system:
            # Manages which dependencies are available.
            self._dependency_manager = DependencyManager(
                work_dir, max_work_dir_size_bytes,
                max_dependencies_serialized_length,
                self._worker_state_manager.previous_runs.keys())
        self._image_manager = DockerImageManager(self._docker, work_dir,
                                                 max_images_bytes)
        self._max_images_bytes = max_images_bytes

        self._exiting_lock = threading.Lock()
        self._exiting = False
        self._should_upgrade = False
        self._last_checkin_successful = False

        # set up docker networks for running bundles: one with external network access and one without
        self.docker_network_external_name = self._docker_network_prefix + "_ext"
        if self.docker_network_external_name not in self._docker.list_networks(
        ):
            logger.debug('Creating docker network: {}'.format(
                self.docker_network_external_name))
            self._docker.create_network(self.docker_network_external_name,
                                        internal=False)
        else:
            logger.debug(
                'Docker network already exists, not creating: {}'.format(
                    self.docker_network_external_name))

        self.docker_network_internal_name = self._docker_network_prefix + "_int"
        if self.docker_network_internal_name not in self._docker.list_networks(
        ):
            logger.debug('Creating docker network: {}'.format(
                self.docker_network_internal_name))
            self._docker.create_network(self.docker_network_internal_name)
        else:
            logger.debug(
                'Docker network already exists, not creating: {}'.format(
                    self.docker_network_internal_name))
Exemplo n.º 6
0
class Worker(object):
    """
    This class is responsible for:

        1) Registering with the bundle service and receiving all messages
           sent to the worker.
        2) Managing all the runs currently executing on the worker and
           forwarding messages associated with those runs to the appropriate
           instance of the Run class.
        3) Spawning classes and threads that manage other worker resources,
           specifically the storage of bundles (both running bundles as well as
           their dependencies) and the cache of Docker images.
        4) Upgrading the worker.
    """
    def __init__(self,
                 id,
                 tag,
                 work_dir,
                 cpuset,
                 gpuset,
                 max_work_dir_size_bytes,
                 max_dependencies_serialized_length,
                 max_images_bytes,
                 shared_file_system,
                 bundle_service,
                 docker,
                 docker_network_prefix='codalab_worker_network'):
        self.id = id
        self._tag = tag
        self.shared_file_system = shared_file_system
        self._bundle_service = bundle_service
        self._docker = docker
        self._docker_network_prefix = docker_network_prefix

        self._resource_lock = threading.Lock()  # lock for cpuset and gpuset
        self._cpuset = cpuset
        self._gpuset = gpuset
        self._cpuset_free = set(
            self._cpuset)  # make a copy of self._cpuset as initial value
        self._gpuset_free = set(
            self._gpuset)  # make a copy of self._gpuset as initial value

        self._worker_state_manager = WorkerStateManager(
            work_dir, self.shared_file_system)

        if not self.shared_file_system:
            # Manages which dependencies are available.
            self._dependency_manager = DependencyManager(
                work_dir, max_work_dir_size_bytes,
                max_dependencies_serialized_length,
                self._worker_state_manager.previous_runs.keys())
        self._image_manager = DockerImageManager(self._docker, work_dir,
                                                 max_images_bytes)
        self._max_images_bytes = max_images_bytes

        self._exiting_lock = threading.Lock()
        self._exiting = False
        self._should_upgrade = False
        self._last_checkin_successful = False

        # set up docker networks for running bundles: one with external network access and one without
        self.docker_network_external_name = self._docker_network_prefix + "_ext"
        if self.docker_network_external_name not in self._docker.list_networks(
        ):
            logger.debug('Creating docker network: {}'.format(
                self.docker_network_external_name))
            self._docker.create_network(self.docker_network_external_name,
                                        internal=False)
        else:
            logger.debug(
                'Docker network already exists, not creating: {}'.format(
                    self.docker_network_external_name))

        self.docker_network_internal_name = self._docker_network_prefix + "_int"
        if self.docker_network_internal_name not in self._docker.list_networks(
        ):
            logger.debug('Creating docker network: {}'.format(
                self.docker_network_internal_name))
            self._docker.create_network(self.docker_network_internal_name)
        else:
            logger.debug(
                'Docker network already exists, not creating: {}'.format(
                    self.docker_network_internal_name))

    def run(self):
        if self._max_images_bytes is not None:
            self._image_manager.start_cleanup_thread()
        if not self.shared_file_system:
            self._dependency_manager.start_cleanup_thread()

        resumed_prev_runs = False

        def resume_previous_runs():
            # resume previous runs
            self._worker_state_manager.resume_previous_runs(
                lambda run_info: Run.deserialize(self._bundle_service, self.
                                                 _docker, self._image_manager,
                                                 self, run_info))

            # for each resumed run, remove the assigned cpu and gpus from the free sets
            with self._resource_lock:
                run_sets = self._worker_state_manager.map_runs(
                    lambda run: (run._cpuset, run._gpuset))
                for cpuset, gpuset in run_sets:
                    for k in cpuset:
                        if k in self._cpuset:
                            self._cpuset_free.remove(k)
                        else:
                            logger.debug(
                                'Warning: cpu {} not in worker cpuset'.format(
                                    k))

                    for k in gpuset:
                        if k in self._gpuset:
                            self._gpuset_free.remove(k)
                        else:
                            logger.debug(
                                'Warning: gpu {} not in worker gpuset'.format(
                                    k))

            self._worker_state_manager.save_state()

        while self._should_run():
            try:
                self._checkin()

                # resume previous runs once in the beginning, but after checkin
                # this is not an ideal design because initial checkin can assign worker with new runs
                # but resources from previous runs are not re-allocated to the worker yet; this can cause
                # performance problems
                if not resumed_prev_runs:
                    resume_previous_runs()
                    resumed_prev_runs = True
                self._worker_state_manager.save_state()
                if not self._last_checkin_successful:
                    logger.info('Connected! Successful check in.')
                self._last_checkin_successful = True

            except Exception:
                self._last_checkin_successful = False
                traceback.print_exc()
                time.sleep(1)

        self._checkout()
        self._worker_state_manager.save_state()

        if self._max_images_bytes is not None:
            self._image_manager.stop_cleanup_thread()
        if not self.shared_file_system:
            self._dependency_manager.stop_cleanup_thread()

        if self._should_upgrade:
            self._upgrade()

    def signal(self):
        logger.info(
            'Exiting: Will wait for exiting jobs to finish, but will not start any new jobs.'
        )
        with self._exiting_lock:
            self._exiting = True

    def _is_exiting(self):
        with self._exiting_lock:
            return self._exiting

    def _should_run(self):
        if not self._is_exiting():
            return True
        return self._worker_state_manager.has_runs()

    def _get_installed_memory_bytes(self):
        try:
            return os.sysconf('SC_PAGE_SIZE') * os.sysconf('SC_PHYS_PAGES')
        except ValueError:
            # Fallback to sysctl when os.sysconf('SC_PHYS_PAGES') fails on OS X
            return int(check_output(['sysctl', '-n', 'hw.memsize']).strip())

    def _get_allocated_memory_bytes(self):
        return sum(
            self._worker_state_manager.map_runs(
                lambda run: run.requested_memory_bytes))

    def _get_memory_bytes(self):
        return max(
            0,
            self._get_installed_memory_bytes() -
            self._get_allocated_memory_bytes())

    def _get_gpu_count(self):
        info = self._docker.get_nvidia_devices_info()
        count = 0 if info is None else len(info['Devices'])
        return count

    def _checkin(self):
        request = {
            'version':
            VERSION,
            'will_upgrade':
            self._should_upgrade,
            'tag':
            self._tag,
            'cpus':
            len(self._cpuset),
            'gpus':
            len(self._gpuset),
            'memory_bytes':
            self._get_memory_bytes(),
            'dependencies': [] if self.shared_file_system else
            self._dependency_manager.dependencies()
        }
        response = self._bundle_service.checkin(self.id, request)
        if response:
            type = response['type']
            logger.debug('Received %s message: %s', type, response)
            if type == 'run':
                self._run(response['bundle'], response['resources'])
            elif type == 'read':
                self._read(response['socket_id'], response['uuid'],
                           response['path'], response['read_args'])
            elif type == 'netcat':
                self._netcat(response['socket_id'], response['uuid'],
                             response['port'], response['message'])
            elif type == 'write':
                self._write(response['uuid'], response['subpath'],
                            response['string'])
            elif type == 'kill':
                self._kill(response['uuid'])
            elif type == 'upgrade':
                with self._exiting_lock:
                    self._exiting = True
                self._should_upgrade = True

    def _allocate_cpu_and_gpu_sets(self, request_cpus, request_gpus):
        """
        Allocate a cpuset and gpuset to assign to a bundle based on given requested resources.
        Side effects: updates the free sets, self._cpuset_free and self._gpuset_free

        Arguments:
            request_cpus: integer
            request_gpus: integer

        Returns a 2-tuple:
            cpuset: Allocated cpuset. Empty set if allocation was unsuccessful
            gpuset: Allocated gpuset. Empty set if allocation was unsuccessful
        """
        cpuset, gpuset = set(), set()

        with self._resource_lock:
            if len(self._cpuset_free) < request_cpus or len(
                    self._gpuset_free) < request_gpus:
                return cpuset, gpuset

            for i in range(request_cpus):
                cpuset.add(self._cpuset_free.pop())
            for j in range(request_gpus):
                gpuset.add(self._gpuset_free.pop())
            return cpuset, gpuset

    def _deallocate_cpu_and_sets(self, cpuset, gpuset):
        """
        Release held up cpus and gpus

        Re-add cpuset and gpuset back to their respective free sets
        """
        with self._resource_lock:
            self._cpuset_free |= cpuset
            self._gpuset_free |= gpuset

    def _run(self, bundle, resources):
        if self.shared_file_system:
            bundle_path = bundle['location']
        else:
            bundle_path = self._dependency_manager.get_run_path(bundle['uuid'])

        cpuset, gpuset = self._allocate_cpu_and_gpu_sets(
            resources['request_cpus'], resources['request_gpus'])

        if len(cpuset) == 0 and len(
                gpuset
        ) == 0:  # revert self._cpuset_free and self._gpuset_free in-place
            logger.debug(
                'Unsuccessful allocation of cpu and gpu sets for bundle %s',
                bundle['uuid'])
            return

        run = Run(self._bundle_service, self._docker, self._image_manager,
                  self, bundle, bundle_path, resources, cpuset, gpuset)
        if run.run():
            self._worker_state_manager.add_run(bundle['uuid'], run)
        else:  # revert self._cpuset_free and self._gpuset_free in-place
            self._deallocate_cpu_and_sets(cpuset, gpuset)

    def add_dependency(self, parent_uuid, parent_path, uuid, loop_callback):
        """
        Registers that the run with UUID uuid depends on path parent_path in
        bundle with UUID parent_uuid. Downloads the dependency if necessary, and
        returns the path to the dependency. Note, remove_dependency should be
        called for every dependency added.

        loop_callback is a method that is called repeatedly while downloading
        the dependency. If that method throws an exception, the download gets
        interrupted and add_dependency fails with that same exception.
        """
        assert (not self.shared_file_system)
        dependency_path, should_download = (
            self._dependency_manager.add_dependency(parent_uuid, parent_path,
                                                    uuid))
        if should_download:
            logger.debug('Downloading dependency %s/%s', parent_uuid,
                         parent_path)
            try:
                download_success = False
                fileobj, target_type = (
                    self._bundle_service.get_bundle_contents(
                        parent_uuid, parent_path))
                with closing(fileobj):
                    # "Bug" the fileobj's read function so that we can keep
                    # track of the number of bytes downloaded so far.
                    old_read_method = fileobj.read
                    bytes_downloaded = [0]

                    def interruptable_read(*args, **kwargs):
                        data = old_read_method(*args, **kwargs)
                        bytes_downloaded[0] += len(data)
                        loop_callback(bytes_downloaded[0])
                        return data

                    fileobj.read = interruptable_read

                    self._store_dependency(dependency_path, fileobj,
                                           target_type)
                    download_success = True
            finally:
                logger.debug('Finished downloading dependency %s/%s',
                             parent_uuid, parent_path)
                self._dependency_manager.finish_download(
                    parent_uuid, parent_path, download_success)

        return dependency_path

    def _store_dependency(self, dependency_path, fileobj, target_type):
        try:
            if target_type == 'directory':
                un_tar_directory(fileobj, dependency_path, 'gz')
            else:
                with open(dependency_path, 'wb') as f:
                    shutil.copyfileobj(fileobj, f)
        except:
            remove_path(dependency_path)
            raise

    def remove_dependency(self, parent_uuid, parent_path, uuid):
        """
        Unregisters that the run with UUID uuid depends on path parent_path in
        bundle with UUID parent_uuid. This method is safe to call on
        dependencies that were never added with add_dependency.
        """
        assert (not self.shared_file_system)
        self._dependency_manager.remove_dependency(parent_uuid, parent_path,
                                                   uuid)

    def _read(self, socket_id, uuid, path, read_args):
        run = self._worker_state_manager._get_run(uuid)
        if run is None:
            Run.read_run_missing(self._bundle_service, self, socket_id)
        else:
            # Reads may take a long time, so do the read in a separate thread.
            threading.Thread(target=Run.read,
                             args=(run, socket_id, path, read_args)).start()

    def _netcat(self, socket_id, uuid, port, message):
        run = self._worker_state_manager._get_run(uuid)
        if run is None:
            Run.read_run_missing(self._bundle_service, self, socket_id)
        else:
            # Reads may take a long time, so do the read in a separate thread.
            threading.Thread(target=Run.netcat,
                             args=(run, socket_id, port, message)).start()

    def _write(self, uuid, subpath, string):
        run = self._worker_state_manager._get_run(uuid)
        if run is not None:
            run.write(subpath, string)

    def _kill(self, uuid):
        run = self._worker_state_manager._get_run(uuid)
        if run is not None:
            run.kill('Kill requested')

    def finish_run(self, uuid):
        """
        Registers that the run with the given UUID has finished.
        """
        self._worker_state_manager.finish_run(uuid)
        if not self.shared_file_system:
            self._dependency_manager.finish_run(uuid)

    def _checkout(self):
        try:
            self._bundle_service.checkout(self.id)
        except BundleServiceException:
            traceback.print_exc()

    def _upgrade(self):
        logger.debug('Upgrading')
        worker_dir = os.path.dirname(os.path.realpath(__file__))

        while True:
            try:
                with closing(self._bundle_service.get_code()) as code:
                    remove_path(worker_dir)
                    un_tar_directory(code, worker_dir, 'gz')
                    break
            except Exception:
                traceback.print_exc()
                time.sleep(1)

        exit(123)
Exemplo n.º 7
0
class Worker(object):
    """
    This class is responsible for:

        1) Registering with the bundle service and receiving all messages
           sent to the worker.
        2) Managing all the runs currently executing on the worker and
           forwarding messages associated with those runs to the appropriate
           instance of the Run class.
        3) Spawning classes and threads that manage other worker resources,
           specifically the storage of bundles (both running bundles as well as
           their dependencies) and the cache of Docker images.
        4) Upgrading the worker.
    """

    def __init__(self, worker_id, tag, work_dir,
                 max_work_dir_size_bytes, max_dependencies_serialized_length,
                 shared_file_system, bundle_service, create_run_manager):
        self.id = worker_id
        self._tag = tag
        self.shared_file_system = shared_file_system
        self._bundle_service = bundle_service
        self._resource_lock = threading.Lock() # lock for cpuset and gpuset
        self._run_manager = create_run_manager(self)

        self._worker_state_manager = WorkerStateManager(
            work_dir=work_dir,
            run_manager=self._run_manager,
            shared_file_system=self.shared_file_system
        )

        if not self.shared_file_system:
            # Manages which dependencies are available.
            self._dependency_manager = DependencyManager(
                work_dir,
                max_work_dir_size_bytes,
                max_dependencies_serialized_length,
                self._worker_state_manager.previous_runs.keys()
            )

        self._exiting_lock = threading.Lock()
        self._exiting = False
        self._should_upgrade = False
        self._last_checkin_successful = False


    def run(self):
        if not self.shared_file_system:
            self._dependency_manager.start_cleanup_thread()

        self._run_manager.worker_did_start()

        while self._should_run():
            try:
                self._checkin()
                self._worker_state_manager.resume_previous_runs()
                self._worker_state_manager.save_state()
                if not self._last_checkin_successful:
                    logger.info('Connected! Successful check in.')
                self._last_checkin_successful = True

            except Exception:
                self._last_checkin_successful = False
                traceback.print_exc()
                time.sleep(1)

        self._checkout()
        self._worker_state_manager.save_state()

        if not self.shared_file_system:
            self._dependency_manager.stop_cleanup_thread()

        self._run_manager.worker_will_stop()

        if self._should_upgrade:
            self._upgrade()

    def signal(self):
        logger.info('Exiting: Will wait for exiting jobs to finish, but will not start any new jobs.')
        with self._exiting_lock:
            self._exiting = True

    def _is_exiting(self):
        with self._exiting_lock:
            return self._exiting

    def _should_run(self):
        if not self._is_exiting():
            return True
        return self._worker_state_manager.has_runs()

    def _get_allocated_memory_bytes(self):
        return sum(self._worker_state_manager.map_runs(lambda run: run.requested_memory_bytes))

    def _get_memory_bytes(self):
        return max(0, self._run_manager.memory_bytes - self._get_allocated_memory_bytes())

    def _checkin(self):
        request = {
            'version': VERSION,
            'will_upgrade': self._should_upgrade,
            'tag': self._tag,
            'cpus': self._run_manager.cpus,
            'gpus': self._run_manager.gpus,
            'memory_bytes': self._get_memory_bytes(),
            'dependencies': [] if self.shared_file_system else self._dependency_manager.dependencies()
        }
        response = self._bundle_service.checkin(self.id, request)
        if response:
            type = response['type']
            logger.debug('Received %s message: %s', type, response)
            if type == 'run':
                self._run(response['bundle'], response['resources'])
            elif type == 'read':
                self._read(response['socket_id'], response['uuid'], response['path'],
                           response['read_args'])
            elif type == 'netcat':
                self._netcat(response['socket_id'], response['uuid'], response['port'],
                           response['message'])
            elif type == 'write':
                self._write(response['uuid'], response['subpath'],
                            response['string'])
            elif type == 'kill':
                self._kill(response['uuid'])
            elif type == 'upgrade':
                with self._exiting_lock:
                    self._exiting = True
                self._should_upgrade = True

    def _run(self, bundle, resources):
        if self.shared_file_system:
            assert 'location' in bundle, \
                "Bundle location not provided by master with shared file system. Are you running as the master user?"
            bundle_path = bundle['location']
        else:
            bundle_path = self._dependency_manager.get_run_path(bundle['uuid'])

        run = self._run_manager.create_run(
            bundle=bundle,
            bundle_path=bundle_path,
            resources=resources
        )
        try:
            run.pre_start()
            if run.start():
                self._worker_state_manager.add_run(bundle['uuid'], run)
        except Exception as e:
            run.kill('Problem starting run.')
            run.post_stop()
            raise e

    def add_dependency(self, parent_uuid, parent_path, uuid, loop_callback):
        """
        Registers that the run with UUID uuid depends on path parent_path in
        bundle with UUID parent_uuid. Downloads the dependency if necessary, and
        returns the path to the dependency. Note, remove_dependency should be
        called for every dependency added.

        loop_callback is a method that is called repeatedly while downloading
        the dependency. If that method throws an exception, the download gets
        interrupted and add_dependency fails with that same exception.
        """
        assert(not self.shared_file_system)
        dependency_path, should_download = (
            self._dependency_manager.add_dependency(parent_uuid, parent_path, uuid))
        if should_download:
            logger.debug('Downloading dependency %s/%s', parent_uuid, parent_path)
            try:
                download_success = False
                fileobj, target_type = (
                    self._bundle_service.get_bundle_contents(parent_uuid, parent_path))
                with closing(fileobj):
                    # "Bug" the fileobj's read function so that we can keep
                    # track of the number of bytes downloaded so far.
                    old_read_method = fileobj.read
                    bytes_downloaded = [0]
                    def interruptable_read(*args, **kwargs):
                        data = old_read_method(*args, **kwargs)
                        bytes_downloaded[0] += len(data)
                        loop_callback(bytes_downloaded[0])
                        return data
                    fileobj.read = interruptable_read

                    self._store_dependency(dependency_path, fileobj, target_type)
                    download_success = True
            finally:
                logger.debug('Finished downloading dependency %s/%s', parent_uuid, parent_path)
                self._dependency_manager.finish_download(
                    parent_uuid, parent_path, download_success)

        return dependency_path

    def _store_dependency(self, dependency_path, fileobj, target_type):
        try:
            if target_type == 'directory':
                un_tar_directory(fileobj, dependency_path, 'gz')
            else:
                with open(dependency_path, 'wb') as f:
                    shutil.copyfileobj(fileobj, f)
        except:
            remove_path(dependency_path)
            raise

    def remove_dependency(self, parent_uuid, parent_path, uuid):
        """
        Unregisters that the run with UUID uuid depends on path parent_path in
        bundle with UUID parent_uuid. This method is safe to call on
        dependencies that were never added with add_dependency.
        """
        assert(not self.shared_file_system)
        self._dependency_manager.remove_dependency(parent_uuid, parent_path, uuid)

    def _read(self, socket_id, uuid, path, read_args):
        socket = self._bundle_service.socket(worker_id=self.id, socket_id=socket_id)
        run = self._worker_state_manager._get_run(uuid)
        if run is None:
            message = {
                'error_code': httplib.INTERNAL_SERVER_ERROR,
                'error_message': BUNDLE_NO_LONGER_RUNNING_MESSAGE,
            }
            socket.reply(message)
        else:
            run.read(path=path, read_args=read_args, socket=socket)

    def _netcat(self, socket_id, uuid, port, message):
        # TODO Netcat isn't supported because it cannot be by a general execution framework; this shouldn't be allowed.
        pass

    def _write(self, uuid, subpath, string):
        run = self._worker_state_manager._get_run(uuid)
        if run is not None:
            run.write(subpath, string)

    def _kill(self, uuid):
        run = self._worker_state_manager._get_run(uuid)
        if run is not None:
            run.kill('Kill requested.')

    def finish_run(self, uuid):
        """
        Registers that the run with the given UUID has finished.
        """
        run = self._worker_state_manager._get_run(uuid)
        if run:
            run.post_stop()
            self._worker_state_manager.finish_run(uuid)
            if not self.shared_file_system:
                self._dependency_manager.finish_run(uuid)

    def _checkout(self):
        try:
            self._bundle_service.checkout(self.id)
        except BundleServiceException:
            traceback.print_exc()

    def _upgrade(self):
        logger.debug('Upgrading')
        worker_dir = os.path.dirname(os.path.realpath(__file__))

        while True:
            try:
                with closing(self._bundle_service.get_code()) as code:
                    remove_path(worker_dir)
                    un_tar_directory(code, worker_dir, 'gz')
                    break
            except Exception:
                traceback.print_exc()
                time.sleep(1)

        exit(123)