Ejemplo n.º 1
0
def crawl(pageNumger):
    url = 'https://search.shopping.naver.com/search/all.nhn?query=텀플러&cat_id=&pagingIndex={}&frm=NVSHATC'.format(
        pageNumger)
    data = requests.get(url)
    print(data, url)
    return data.content


# pageString = crawl(4)
# products = parse(pageString)
# print(products)
# print(len(products))

totalProducts = []

# for pageNo in range(0, 10+1):
#     pageString = crawl(pageNo)
#     product = parse(pageString)
#     totalProducts += product

for pageString in map(crawl, range(1, 11)):
    totalProducts += parse(pageString)

# pageString = list())
# totalProducts += list(map(parse, pageString))
print(len(totalProducts))

file = open('./tumbler.json', 'w+')
file.write(json.dumps(totalProducts))
Ejemplo n.º 2
0
from libs.naver_shopping.crawler import crawl
from libs.naver_shopping.parser import parse
import json

pageString = crawl('')
products = parse(pageString)
print(len(products))

#json파일로 내보내기
##file = open("./products.json", "w+")
##file.write(json.dumps(products))
Ejemplo n.º 3
0
import xlrd
from libs.naver_shopping.crawler import crawl
from libs.naver_shopping.parser import parse

workbook = xlrd.open_workbook('user_info.xlsx')
worksheet = workbook.sheet_by_name('db')
list = worksheet._cell_values

for row in list[1:]:
    keyword = row[0]
    result = parse(crawl(keyword))
    print(keyword, result)