def search_news(keyword): newsdao = NewsDAO() #memcache = MemCache() #user_id = request.args.get('user_id') #apikey = request.args.get('apikey') # print user_id, apikey # if memcache.auth_user(user_id, apikey): # data = newsdao.get_news_by_keyword_in_content(str(keyword)) # else: # data = {'result' : '인증에 실패하였습니다.'} data = newsdao.get_news_by_keyword_in_content(str(keyword)) #return json.dumps(data, ensure_ascii=False).encode('utf8') return jsonify(data)
soup = BeautifulSoup(content) table = soup.find('table', attrs={'class': 'container'}) for a in table.find_all(find_a): link = a['href'] self.crawl_title_content(link) def crawl_title_content(self, link): res = requests.get(link) content = res.content soup = BeautifulSoup(content) title = soup.find('h3', attrs={'id': 'articleTitle'}).get_text() content = soup.find('div', attrs={ 'id': 'articleBodyContents' }).get_text().strip() #print link #print str(title) #print str(content) self.newsdao.save_news(link, str(title), str(content)) urls = ['http://news.naver.com/main/main.nhn?mode=LSD&mid=shm&sid1=105'] newsdao = NewsDAO() crawler = NaverNewsCrawler(newsdao, urls) crawler.crawl_link()
def search_news(keyword): print type(keyword) newsdao = NewsDAO() data = newsdao.get_news_by_keyword_in_content(str(keyword)) return jsonify(data)
def search_news_title(keyword): session = Session() newsdao = NewsDAO() data = newsdao.get_news_by_keyword_in_title(str(keyword)) return jsonify(data)