def add_to_source(self): # check if already exsit source = News_source_m() source.url =('http://' + self.link.split('/')[2]).decode('utf-8') if store.find(News_source_m, News_source_m.url == source.url).count() == 0: print "saving" source.save()
def init_record(): init_db() init_record() # mnews.get_info() # crawl_news("http://www.solidot.org/story?sid=40747") # pool.wait() print "my source" News_source_m.print_all() print "error page" UnparsePage_m.print_all()