Exemple #1
0
def search_news(keyword):
    newsdao = NewsDAO()
    #memcache = MemCache()

    #user_id = request.args.get('user_id')
    #apikey = request.args.get('apikey')

    # print user_id, apikey
    # if memcache.auth_user(user_id, apikey):
    #     data = newsdao.get_news_by_keyword_in_content(str(keyword))
    # else:
    #     data = {'result' : '인증에 실패하였습니다.'}
    data = newsdao.get_news_by_keyword_in_content(str(keyword))
    #return json.dumps(data, ensure_ascii=False).encode('utf8')
    return jsonify(data)
Exemple #2
0
            soup = BeautifulSoup(content)

            table = soup.find('table', attrs={'class': 'container'})
            for a in table.find_all(find_a):
                link = a['href']
                self.crawl_title_content(link)

    def crawl_title_content(self, link):
        res = requests.get(link)
        content = res.content

        soup = BeautifulSoup(content)

        title = soup.find('h3', attrs={'id': 'articleTitle'}).get_text()
        content = soup.find('div', attrs={
            'id': 'articleBodyContents'
        }).get_text().strip()
        #print link
        #print str(title)
        #print str(content)

        self.newsdao.save_news(link, str(title), str(content))


urls = ['http://news.naver.com/main/main.nhn?mode=LSD&mid=shm&sid1=105']
newsdao = NewsDAO()

crawler = NaverNewsCrawler(newsdao, urls)
crawler.crawl_link()
Exemple #3
0
def search_news(keyword):
    print type(keyword)
    newsdao = NewsDAO()
    data = newsdao.get_news_by_keyword_in_content(str(keyword))
    return jsonify(data)
def search_news_title(keyword):
    session = Session()
    newsdao = NewsDAO()
    data = newsdao.get_news_by_keyword_in_title(str(keyword))

    return jsonify(data)