try: spider_main.crawl(url, path) except Exception as e: with open('error/error.txt', 'a+') as f: f.write('94'+str(e) + '\n') end = time.time() times = int(end - start) if times > 60: mins = times//60 second = times - mins * 60 print('搜索用户所用时间为%d分%d秒' % (mins, second)) else: print('搜索用户所用时间为%d秒' % times) else: # 加载先前下载好的文件 spider_main.load_users(path, used_path) # 启动服务 server(spider_main) print('关闭服务,显示数据') # 搜索用户结束,则重新搜索之前搜索失败的用户集合 if not spider_main.urlmanager.get_user_size(): spider_main.crawl_error_user() if not spider_main.displaydata.display(): os.chdir('html') os.system(name + '.html') print('程序结束')