Esempio n. 1
0
    def run_module(self, url, home, depth_limit):
        
        confine_prefix = url
        exclude = ''

        if home[-1] != '/': home = home + '/'

        self.mprint('[%s] Crawling paths in %s (depth %i)' % (self.name, url, depth_limit))
        
        try:
            crawler = Crawler(url, depth_limit, confine_prefix, exclude)
            crawler.crawl()
        except Exception, e:
            raise ModuleException(self.name, "Crawler exception: %s" % str(e))
Esempio n. 2
0
 def _prepare(self):
 
     if not url_validator.match(self.args['url']):
         raise ProbeException(self.name, '\'%s\': %s' % (self.args['url'], WARN_NOT_URL) )
     if not url_validator.match(self.args['baseurl']):
         raise ProbeException(self.name, '\'%s\': %s' % (self.args['baseurl'], WARN_NOT_URL) )
 
     url = self.args['url']    
     baseurl = self.args['baseurl']
     rpath = self.args['rpath']
     
     urls = []
 
     try:
         crawler = Crawler(url, self.args['depth'], '', '')
         crawler.crawl()
     except ModuleException, e:
         raise