def __init__(self): self.manager = URLManager() self.downloader = HtmlDownloader() self.parser = HtmlParser() self.output = DataOutput()
import URLManager import fetcher """ Feed the fetcher with url """ if __name__ == '__main__': config = configparser.ConfigParser() config.read('crawler.config') print("Master started. Initial page {}".format( config["SITE"]["initial_page"])) # init start_time = datetime.datetime.now() url_manager = URLManager.URLManager() url_manager.insert_url(config["SITE"]["initial_page"], 0, 0, -1) fetcher = fetcher.Fetcher(url_manager) end_time = datetime.datetime.now() delta = end_time - start_time print("Init time", delta) # start crawling while url_manager.has_next_url(): # TODO: change if parallel print("queue size", url_manager.get_size()) next_url = url_manager.get_next_url() print("fetching", next_url) fetcher.get_page(next_url)