Example #1
0
class TestOffsiteMiddleware(TestCase):

    def setUp(self):
        self.spider = self._get_spider()
        self.mw = OffsiteMiddleware()
        self.mw.spider_opened(self.spider)

    def _get_spider(self):
        return BaseSpider('foo', allowed_domains=['scrapytest.org', 'scrapy.org'])

    def test_process_spider_output(self):
        res = Response('http://scrapytest.org')

        onsite_reqs = [Request('http://scrapytest.org/1'),
                       Request('http://scrapy.org/1'),
                       Request('http://sub.scrapy.org/1'),
                       Request('http://offsite.tld/letmepass', dont_filter=True)]
        offsite_reqs = [Request('http://scrapy2.org'),
                       Request('http://offsite.tld/')]
        reqs = onsite_reqs + offsite_reqs

        out = list(self.mw.process_spider_output(res, reqs, self.spider))
        self.assertEquals(out, onsite_reqs)

    def tearDown(self):
        self.mw.spider_closed(self.spider)
    def setUp(self):
        self.spider = BaseSpider('foo')
        self.spider.name = 'scrapytest.org'
        self.spider.allowed_domains = ['scrapytest.org', 'scrapy.org']

        self.mw = OffsiteMiddleware()
        self.mw.spider_opened(self.spider)
    def setUp(self):
        self.spider = BaseSpider()
        self.spider.domain_name = 'scrapytest.org'
        self.spider.extra_domain_names = ['scrapy.org']

        self.mw = OffsiteMiddleware()
        self.mw.spider_opened(self.spider)
class TestOffsiteMiddleware(TestCase):

    def setUp(self):
        self.spider = BaseSpider('foo')
        self.spider.name = 'scrapytest.org'
        self.spider.allowed_domains = ['scrapytest.org', 'scrapy.org']

        self.mw = OffsiteMiddleware()
        self.mw.spider_opened(self.spider)

    def test_process_spider_output(self):
        res = Response('http://scrapytest.org')

        onsite_reqs = [Request('http://scrapytest.org/1'),
                       Request('http://scrapy.org/1'),
                       Request('http://sub.scrapy.org/1')]
        offsite_reqs = [Request('http://scrapy2.org')]
        reqs = onsite_reqs + offsite_reqs

        out = list(self.mw.process_spider_output(res, reqs, self.spider))
        self.assertEquals(out, onsite_reqs)

    def tearDown(self):
        self.mw.spider_closed(self.spider)
 def setUp(self):
     self.spider = self._get_spider()
     crawler = get_crawler()
     self.mw = OffsiteMiddleware.from_crawler(crawler)
     self.mw.spider_opened(self.spider)
 def setUp(self):
     self.spider = self._get_spider()
     crawler = get_crawler()
     self.mw = OffsiteMiddleware.from_crawler(crawler)
     self.mw.spider_opened(self.spider)
Example #7
0
 def setUp(self):
     self.spider = self._get_spider()
     self.mw = OffsiteMiddleware()
     self.mw.spider_opened(self.spider)
Example #8
0
 def setUp(self):
     crawler = get_crawler(Spider)
     self.spider = crawler._create_spider(**self._get_spiderargs())
     self.mw = OffsiteMiddleware.from_crawler(crawler)
     self.mw.spider_opened(self.spider)
 def setUp(self):
     crawler = get_crawler(Spider)
     self.spider = crawler._create_spider(**self._get_spiderargs())
     self.mw = OffsiteMiddleware.from_crawler(crawler)
     self.mw.spider_opened(self.spider)