Esempio n. 1
0
def do_initial_works():
    """
    works after db initialized
    :return:
    """
    cli = pyredis_pool.acquire()

    from core.crawler import spider_keymap
    from service.spider.functions import key_prefix

    existed_keys = cli.keys('%s*' % key_prefix)

    for _type, _map in spider_keymap.items():
        for k, v in _map.items():
            if k in existed_keys:
                existed_keys.remove(k)
            cli.hmset(k,
                      'status', 'stopped',
                      'name', v.name)
    for k in existed_keys:
        cli.delete(k)
    logger.info('Finished database initial works.')
Esempio n. 2
0
 def __init__(self):
     self.cli = pyredis_pool.acquire()
     self.cli.hmset_dict = self.hmset_dict
Esempio n. 3
0
 def __init__(self):
     from core.db.redis import pyredis_pool
     self.cli = pyredis_pool.acquire()
     self.cli.hmset_dict = self.hmset_dict
Esempio n. 4
0
class SpiderServer:
    # TODO: move these to config file?
    enabled_crawlers = {
        'crawler': (
            A3464Spider,
            CoderbusySpider,
            Data5uSpider,
            Ip66Spider,
            KuaidailiSpider,
            MixSpider,
            XicidailiSpider,
            YundailiSpider,
            # Ip89Spider,
            # IhuanSpider,
        ),
        'checker': (CheckerSpider, )
    }
    cli = aioredis_pool
    blocking_cli = pyredis_pool.acquire()  # XXX: It's ugly

    async def check(self, key=None, keys=None, detail=False):
        """
        Argument:
            if detail:
                return detailed info
            else:
                return value(s)
        """

        if (key and keys) or (not key and not keys):
            raise AssertionError

        results = []
        _keys = keys or [key]

        for k in _keys:
            st = await self.cli.get(k)

            if detail:
                total_time = None

                if st:
                    total_time = int(time.time()) - int(st)

                results.append(
                    {k: {
                        "start_time": st,
                        "total_time": total_time
                    }})
            else:
                results.append(st)

        return results if keys else results[0]

    async def all_status(self):
        """
        get all status about spiders
        """
        keys = await self.cli.keys('%s*' % key_prefix)

        if keys:
            return await self.check(keys=keys, detail=True)

        return []

    async def register_status(self, key):
        """

        :param key: will be deleted after crawling or last for some hours
        :param detail:
        :return:
        """
        t = int(time.time())

        await self.cli.set(key, t)
        # await self.cli.expire(key, 4 * 3600)

        return t

    async def unregister_status(self, st, key, *args, **kw):
        """

        :param st: start time
        :param key: key in db
        :param args: preserved
        :param kw: preserved
        :return: preserved
        """
        st = await self.cli.get(key)
        total_time = int(time.time()) - int(st)

        return await self.cli.delete(key), total_time

    def callback_unregister_status(self, _, st, key, *args, **kw):
        """
        as callback of crawling
        * Twisted doesn't support aioredis

        :param _: preserved for fired deffer
        :param st: start time
        :param key: key in db
        :param args: preserved
        :param kw: preserved
        :return: preserved
        """
        total_time = int(time.time()) - int(st)
        logger.info("One spider finished working. "
                    "Delete key: %s. Total time: %s" % (key, total_time))

        return self.blocking_cli.delete(key), total_time

    async def crawler_condition(self):
        keys = await self.cli.keys('%s*' % proxy_key_prefix) or []

        return len(keys) < config.PROXY_STORE_NUM

    async def start_crawling(self, _type='crawler'):
        """
        trigger spiders
        """

        started = []

        if _type == 'crawler' and not await self.crawler_condition():
            return

        for spider in self.enabled_crawlers[_type]:
            key = build_key(spider)
            running = bool(await self.check(key=key))

            if running:
                continue
            # register
            st = await self.register_status(key)
            # TODO: specify settings
            logger.info('Started %s at %s. Key: %s.' % (st, spider, key))

            # build a new thread
            IOLoop.current().run_in_executor(None, self.run_crawler, spider,
                                             st, key)
            started.append(key)

        return started

    def run_crawler(self, spider, st, key):
        """ run a crawler, then unregister it
        * moved to another thread

        :st: start time
        :key: key in redis
        """
        crawler = self.get_crawler(spider)
        logger.info('Got crawler: %s' % crawler)
        d = crawler_runner.crawl(crawler)
        # unregister
        d.addBoth(self.callback_unregister_status, st=st, key=key)

    def get_crawler(self, spider):
        """
        do some specific settings

        :param spider: spider class
        :return: crawler
        """
        settings = crawler_runner.settings

        # FIXME !!!
        # conf = {}
        # log_file = crawler_runner.settings.get('LOG_FILE')
        # if log_file:
        #     conf['LOG_FILE'] = '%s.%s' % (log_file, spider.name)
        #     conf['LOG_FILE'] = None
        #     conf['LOG_FORMAT'] = ('%(levelname)1.1s [%(asctime)s]'
        #                           ' [spider-{spider}]'
        #                           ' %(message)s'
        #                           ).format(spider=spider.name)
        #     settings = updated_crawler_settings(settings, conf)
        # configure_logging(settings)

        return Crawler(spider, settings)
Esempio n. 5
0
 def __init__(self):
     self.cli = aioredis_pool
     self.blocking_cli = pyredis_pool.acquire()  # XXX: It's ugly