def _Main(): All_Line_Arr = Creep_Tools.Read_Line_by_Line._Read_Line_by_Line("goods_id.log") for All_Line in All_Line_Arr: Line_arr = All_Line.split('@@@') commodity_index = Line_arr[0] commodity_url = Line_arr[1] ID_arr = commodity_url.split('.') ID = ID_arr[2] ID_arr = ID.split('/') ID = ID_arr[1] # 2.创建对应的文件夹 code_picture_path = "E:/JD/" + commodity_index + "/" + ID + "/code_picture" detail_picture_path = "E:/JD/" + commodity_index + "/" + ID + "/detail_picture" Creep_Tools._mkdir(code_picture_path) # 3.获取网页的HTML文件 import commodity_list # 4.解析网页的soup文件 # 获取商品图片的 soup = Creep_Tools._Analyze_Soup(commodity_url) ul = soup.find('ul', {'class': "lh"}) if ul: all_li = ul.find_all('li') num = 0 for li in all_li: num = num +1 img = dict(li.contents[0].attrs)['data-url'] img_url = 'https://img11.360buyimg.com/popWaterMark/%s' % (img) save_path = code_picture_path ID = ID + str(num) Creep_Tools._Download_Picture(img_url, ID, save_path) # 获取商品详情 product_detail = get_product_detail(soup) print(product_detail) # 获取商品价格 price = get_product_price(soup) print(price) savePath = get_datail_picture(soup) print(savePath)