示例#1
0
    def test_write_exception(self):

        with self.assertRaises(Exception):
            QueueSink(BrokenQueue(Exception())).write(1)

        QueueSink(BrokenQueue(EOFError())).write(1)
        QueueSink(BrokenQueue(BrokenPipeError())).write(1)
示例#2
0
    def test_keyboard_interrupt(self):
        stdin = QueueSource(Queue())
        stdout = QueueSink(Queue())
        stderr = QueueSink(Queue())

        filter = ExceptionFilter(KeyboardInterrupt())

        stdin._queue.put(pickle.dumps(1))
        stdin._queue.put(pickle.dumps(2))
        stdin._queue.put(None)

        self.assertEqual(0, stdout._queue.qsize())
        self.assertEqual(0, stderr._queue.qsize())
        PipesPool.worker(filter, stdin, stdout, stderr, None)
        self.assertEqual(0, stdout._queue.qsize())
        self.assertEqual(0, stderr._queue.qsize())
示例#3
0
    def test_attribute_error(self):
        stdin = QueueSource(Queue())
        stdout = QueueSink(Queue())
        stderr = QueueSink(Queue())

        filter = ExceptionFilter(Exception("Can't get attribute"))

        stdin._queue.put(pickle.dumps(1))
        stdin._queue.put(pickle.dumps(2))
        stdin._queue.put(None)

        self.assertEqual(0, stdout._queue.qsize())
        self.assertEqual(0, stderr._queue.qsize())
        PipesPool.worker(filter, stdin, stdout, stderr, None)
        self.assertEqual(0, stdout._queue.qsize())
        self.assertEqual(1, stderr._queue.qsize())
        self.assertIn("We attempted to evaluate", stderr._queue.get())
示例#4
0
    def test_worker_iterable_filter_return(self):

        stdin = QueueSource(Queue())
        stdout = QueueSink(Queue())
        stderr = QueueSink(Queue())

        filter = ProcessNameFilter()

        stdin._queue.put(pickle.dumps(1))
        stdin._queue.put(None)

        self.assertEqual(0, stdout._queue.qsize())
        self.assertEqual(0, stderr._queue.qsize())
        PipesPool.worker(filter, stdin, stdout, stderr, None)
        self.assertEqual(1, stdout._queue.qsize())
        self.assertEqual(1, len(stdout._queue.get()))
        self.assertEqual(0, stderr._queue.qsize())
示例#5
0
    def test_worker_stderr(self):

        stdin = QueueSource(Queue())
        stdout = QueueSink(Queue())
        stderr = QueueSink(Queue())

        filter = ExceptionFilter()

        stdin._queue.put(pickle.dumps(1))
        stdin._queue.put(None)

        self.assertEqual(0, stdout._queue.qsize())
        self.assertEqual(0, stderr._queue.qsize())
        PipesPool.worker(filter, stdin, stdout, stderr, None)
        self.assertEqual(0, stdout._queue.qsize())
        self.assertEqual(1, stderr._queue.qsize())
        self.assertEqual("Exception Filter", str(stderr._queue.get()[2]))
示例#6
0
    def test_worker_stdin_stdout_max(self):

        stdin = QueueSource(Queue())
        stdout = QueueSink(Queue())
        stderr = QueueSink(Queue())

        filter = Identity()

        stdin._queue.put(pickle.dumps(1))
        stdin._queue.put(pickle.dumps(2))
        stdin._queue.put(None)

        self.assertEqual(0, stdout._queue.qsize())
        self.assertEqual(0, stderr._queue.qsize())
        PipesPool.worker(filter, stdin, stdout, stderr, 1)
        self.assertEqual(1, stdout._queue.qsize())
        self.assertEqual(1, stdout._queue.get())
        self.assertEqual(0, stderr._queue.qsize())
示例#7
0
    def test_write(self):

        sink = QueueSink()

        sink.write("a")
        sink.write("b")
        sink.write("c")

        self.assertEqual("a", sink._queue.get())
        self.assertEqual("b", sink._queue.get())
        self.assertEqual("c", sink._queue.get())
示例#8
0
    def filter(self, items: Iterable[Any]) -> Iterable[Any]:

        if len(self._filters) == 0:
            return items

        try:
            with Manager() as manager:

                stdout_queue = manager.Queue()  #type: ignore
                stdlog_queue = manager.Queue()  #type: ignore

                stdout_writer, stdout_reader = QueueSink(
                    stdout_queue), QueueSource(stdout_queue)
                stdlog_writer, stdlog_reader = QueueSink(
                    stdlog_queue), QueueSource(stdlog_queue)

                class MyPool(multiprocessing.pool.Pool):

                    _missing_error_definition_error_is_new = True

                    def _join_exited_workers(self):

                        for worker in self._pool:
                            if worker.exitcode == 1000 and MyPool._missing_error_definition_error_is_new:
                                #this is a hack... This only works so long as we just
                                #process one job at a time... This is true in our case.
                                #this is necessary because multiprocessing can get stuck
                                #waiting for failed workers and that is frustrating for users.

                                MyPool._missing_error_definition_error_is_new = False

                                message = (
                                    "Coba attempted to evaluate your benchmark in multiple processes but the pickle module was unable to "
                                    "find all the definitions needed to pass the tasks to the processes. The two most common causes of "
                                    "this error are: 1) a learner or simulation is defined in a Jupyter Notebook cell or 2) a necessary "
                                    "class definition exists inside the `__name__=='__main__'` code block in the main execution script. In "
                                    "either case there are two simple solutions: 1) evalute your benchmark in a single processed with no "
                                    "limit on child tasks or 2) define all you classes in a separate python file that is imported when "
                                    "evaluating.")

                                CobaConfig.Logger.log(message)

                            if worker.exitcode is not None and worker.exitcode != 0:
                                #A worker exited in an uncontrolled manner and was unable to clean its job
                                #up. We therefore mark one of the jobs as "finished" but failed to prevent an
                                #infinite wait on a failed job to finish that is actually no longer running.
                                list(self._cache.values())[0]._set(
                                    None, (False, None))

                        return super()._join_exited_workers()

                with MyPool(self._processes,
                            maxtasksperchild=self._maxtasksperchild) as pool:

                    # handle not picklable (this is handled by done_or_failed)
                    # handle empty list (this is done by checking result.ready())
                    # handle exceptions in process (unhandled exceptions can cause children to hang so we pass them to stderr)
                    # handle ctrl-c without hanging
                    #   > don't call result.get when KeyboardInterrupt has been hit
                    #   > handle EOFError,BrokenPipeError errors with queue since ctr-c kills manager
                    # handle AttributeErrors. These occure when... (this is handled by shadowing several pool methods)
                    #   > a class that is defined in a Jupyter Notebook cell is pickled
                    #   > a class that is defined inside the __name__=='__main__' block is pickeled
                    # handle Benchmark.evaluate not being called inside of __name__=='__main__' (this is handled by a big try/catch)

                    def done_or_failed(results_or_exception=None):
                        #This method is called one time at the completion of map_async
                        #in the case that one of our jobs threw an exception the argument
                        #will contain an exception otherwise it will be the returned results
                        #of all the jobs. This method is executed on a thread in the Main context.

                        if isinstance(results_or_exception, Exception):
                            from coba.config import CobaConfig

                            if "Can't pickle" in str(
                                    results_or_exception) or "Pickling" in str(
                                        results_or_exception):

                                message = (
                                    str(results_or_exception) +
                                    ". Coba attempted to process your Benchmark on multiple processes and "
                                    "the named class was not able to be pickled. This problem can be fixed in one of two ways: 1) "
                                    "evaluate the benchmark in question on a single process with no limit on the tasks per child or 2) "
                                    "modify the named class to be picklable. The easiest way to make the given class picklable is to "
                                    "add `def __reduce__ (self) return (<the class in question>, (<tuple of constructor arguments>))` to "
                                    "the class. For more information see https://docs.python.org/3/library/pickle.html#object.__reduce__."
                                )

                                CobaConfig.Logger.log(message)
                            else:
                                CobaConfig.Logger.log_exception(
                                    results_or_exception)

                        stdout_writer.write([None])
                        stdlog_writer.write([None])

                    log_thread = Thread(target=Pipe.join(
                        stdlog_reader, [], CobaConfig.Logger.sink).run)
                    log_thread.daemon = True
                    log_thread.start()

                    processor = MultiprocessFilter.Processor(
                        self._filters, stdout_writer, stdlog_writer,
                        self._processes)
                    result = pool.map_async(processor.process,
                                            items,
                                            callback=done_or_failed,
                                            error_callback=done_or_failed,
                                            chunksize=1)

                    # When items is empty finished_callback will not be called and we'll get stuck waiting for the poison pill.
                    # When items is empty ready() will be true immediately and this check will place the poison pill into the queues.
                    if result.ready(): done_or_failed()

                    try:
                        for item in stdout_reader.read():
                            yield item
                        pool.close()
                    except (KeyboardInterrupt, Exception):
                        try:
                            pool.terminate()
                        except:
                            pass
                        raise
                    finally:
                        pool.join()
                        log_thread.join()

        except RuntimeError as e:
            #This happens when importing main causes this code to run again
            raise CobaFatal(str(e))