Esempio n. 1
0
def downloadAndSave(url, savePath):
    try:
        response = requests.get(url, proxies=proxies)
        html_source = response.text
        soup = BeautifulSoup(html_source, "html.parser")
        # 获取文章名称
        title_entity = soup.select('head > title')[0]
        title = title_entity.text
        seed_urls = re.findall(r_seed_link_match, html_source)
        seed_save = []
        for link in seed_urls:
            print(link)
            seed_save.append(head_seed + link)

        file_task = FileWriter.FileWriteMission(
            FileWriter.FileWriteMission.MISSION_NOMAL, title, url, seed_save)
        FileWriter.add_file_write_task(file_task)
    except Exception as e:
        print(e)
        pass