def parse_source(self, existing_ids=None): news = [] feed_content = get_rss(self.ZURNAL_RSS_URL) for feed_entry in feed_content.entries: link = feed_entry["link"] if existing_ids and (get_hash(link) in existing_ids or get_sha_hash(link) in existing_ids): logger.debug("Skipping %s", link) continue published_date = time_to_datetime(feed_entry["published_parsed"]) news.append((link, {"published": published_date})) return news
def parse_source(self, existing_ids=None): news = [] for rss_feed in self.RTV_RSS_URLS: logger.debug("Parsing %s", rss_feed) feed_content = get_rss(rss_feed) for feed_entry in feed_content.entries: # Download article link = feed_entry["link"] if existing_ids and (get_hash(link) in existing_ids or get_sha_hash(link) in existing_ids): logger.debug("Skipping %s", link) continue published_date = time_to_datetime( feed_entry["published_parsed"]) news.append((link, {"published": published_date})) return news
def parse_source(self, existing_ids=None): news = [] feed_content = get_rss(self.DNEVNIK_RSS_URL) max_counter = 30 for feed_entry in feed_content.entries: link = feed_entry["link"] if existing_ids and (get_hash(link) in existing_ids or get_sha_hash(link) in existing_ids): logger.debug("Skipping %s", link) continue published_date = time_to_datetime(feed_entry["published_parsed"]) title = feed_entry["title"] news.append((link, {"published": published_date, "title": title})) max_counter -= 1 if max_counter <= 0: break return news