def get_defaults_spider_mw(self): crawler = get_crawler(Spider) spider = crawler._create_spider('foo') defaults = { to_bytes(k): [to_bytes(v)] for k, v in crawler.settings.get('DEFAULT_REQUEST_HEADERS').items() } return defaults, spider, DefaultHeadersMiddleware.from_crawler(crawler)
def get_defaults_spider_mw(self): crawler = get_crawler(Spider) spider = crawler._create_spider('foo') defaults = { to_bytes(k): [to_bytes(v)] for k, v in crawler.settings.get('DEFAULT_REQUEST_HEADERS').items() } return defaults, spider, DefaultHeadersMiddleware.from_crawler(crawler)
def get_defaults_spider_mw(self): crawler = get_crawler(Spider) spider = crawler._create_spider('foo') defaults = dict([(k, [v]) for k, v in \ six.iteritems(crawler.settings.get('DEFAULT_REQUEST_HEADERS'))]) return defaults, spider, DefaultHeadersMiddleware.from_crawler(crawler)
def get_defaults_spider_mw(self): crawler = get_crawler(Spider) spider = crawler._create_spider('foo') defaults = dict([(k, [v]) for k, v in \ six.iteritems(crawler.settings.get('DEFAULT_REQUEST_HEADERS'))]) return defaults, spider, DefaultHeadersMiddleware.from_crawler(crawler)