Esempio n. 1
0
def _Main():
	All_Line_Arr = Creep_Tools.Read_Line_by_Line._Read_Line_by_Line("goods_id.log")
	for All_Line in All_Line_Arr:
		Line_arr = All_Line.split('@@@')
		commodity_index = Line_arr[0]
		commodity_url = Line_arr[1]
		ID_arr = commodity_url.split('.')
		ID = ID_arr[2]
		ID_arr = ID.split('/')
		ID = ID_arr[1]
		# 2.创建对应的文件夹
		code_picture_path = "E:/JD/" + commodity_index + "/" + ID + "/code_picture"
		detail_picture_path = "E:/JD/" + commodity_index + "/" + ID + "/detail_picture"
		Creep_Tools._mkdir(code_picture_path)
		# 3.获取网页的HTML文件
		import commodity_list
		# 4.解析网页的soup文件

		# 获取商品图片的
		soup = Creep_Tools._Analyze_Soup(commodity_url)
		ul = soup.find('ul', {'class': "lh"})
		if ul:
			all_li = ul.find_all('li')
			num = 0
			for li in all_li:
				num = num +1
				img = dict(li.contents[0].attrs)['data-url']
				img_url = 'https://img11.360buyimg.com/popWaterMark/%s' % (img)
				save_path = code_picture_path
				ID = ID + str(num)
				Creep_Tools._Download_Picture(img_url, ID, save_path)
		# 获取商品详情
		product_detail = get_product_detail(soup)
		print(product_detail)

		# 获取商品价格
		price = get_product_price(soup)
		print(price)
		savePath = get_datail_picture(soup)
		print(savePath)