def ProcessMain(): global db,Cqvip db = HCJ_MySQL() Cqvip = Cqvip_Crawler(db=db) multiprocessing.freeze_support() # 多进程打包的话必须加上 init_main() if '0' in (Read_buff(file_buff="Config.ini", settion=SearchDBName, info='stopflag')) : main()
def ProcessMain(): global db, Wanfang multiprocessing.freeze_support() # db = HCJ_MySQL() Wanfang = WanFangCrawler(db=db) init_main() if '0' in str( Read_buff(file_buff="Config.ini", settion=SearchDBName, info='stopflag')): main()
def run(self): _db = HCJ_MySQL() _Cqvip = Cnki_Crawler(db=_db) _Cqvip.WriteAllUrlIntoDBMain() print("Cnki:获取全部url结束")
def run(self): _db = HCJ_MySQL() _Cqvip = self.Crawler(db=_db) _Cqvip.WriteAllUrlIntoDBMain() print("结束")
def init_main(): if '1' in str(Read_buff(file_buff="Config.ini", settion=SearchDBName, info='restart')): CreatResultDBTable(db,Dbresult) CreatUrlBuffTable(db,DbDatabuff) time.sleep(0.05) Write_buff(file_buff="Config.ini", settion=SearchDBName, info="restart", state=0) Write_buff(file_buff="Config.ini", settion=SearchDBName, info="startpage", state=1) Write_buff(file_buff="Config.ini", settion=SearchDBName, info="stopflag", state=0) Write_buff(file_buff="Config.ini", settion=SearchDBName, info="flag_get_all_url", state=0) if '0' in str(Read_buff(file_buff="Config.ini", settion=SearchDBName, info='restart')): db.upda_sql("Update `%s` set `State`=0 where `State`=10"%DbDatabuff) time.sleep(1) ex_dbname = Read_buff(file_buff="Config.ini", settion=SearchDBName, info='ex_dbname') DbDatabuff="databuff"+str(ex_dbname) Dbresult="result"+str(ex_dbname) def ProcessMain(): global db,Cqvip db = HCJ_MySQL() Cqvip = Cqvip_Crawler(db=db) multiprocessing.freeze_support() # 多进程打包的话必须加上 init_main() if '0' in (Read_buff(file_buff="Config.ini", settion=SearchDBName, info='stopflag')) : main() if __name__ == '__main__': # ProcessMain() db = HCJ_MySQL() Cqvip = Cqvip_Crawler(db=db) url="http://www.cqvip.com/%22/QK/93419A/200901/29132001.html%22" g=GetSoup(url) parse(url,g)
def run(self): # 在调用子进程的start() 的时候,默认会执行run() 方法 _db = HCJ_MySQL() _Wanfang = WanFangCrawler(db=_db) _Wanfang.GetAllUrl() print("采集链接结束")