Exemplo n.º 1
0
    def _clean_article(self):
        """
        remove html tags, images, links from the article, and encode it
        appropriately
        """
        try:
            # convert to normal encoding
            self._article = str(
                urllib2.unquote(hparser.unescape(self._article)))

            # remove unnecessary parts
            html_stripper = html2text.HTML2Text()
            html_stripper.ignore_links = True
            html_stripper.ignore_images = True
            html_stripper.ignore_emphasis = True
            # body_width = 0 disables text wrapping
            html_stripper.body_width = 0
            self._article = html_stripper.handle(
                self._article).strip().strip("#").strip()

            # convert to appropriate encoding
            if isinstance(self._article, str):
                self._article = self._article.decode(
                    chardet.detect(self._article)['encoding'], 'ignore')
        except Exception as k:
            logger.error(str(k))
            return None
Exemplo n.º 2
0
    def _analyze(self, image_url=None, referer=None):
        """
        remove CDN prefix, if any; and read image data
        """
        if not image_url:
            logger.error('Method malformed!')
            raise Exception('Method malformed!')

        image_url = image_url.replace("\/", "/").strip()
        image_url = urllib2.unquote(hparser.unescape(image_url))

        # as the name could be http://xxx.com/yyy--http://zzz.jpg
        # or http://xxx.com/yyy--https://zzz.jpg
        last_http_index = image_url.rfind('http')
        image_url = image_url[last_http_index:]

        response = None
        if referer:
            HEADERS['Referer'] = referer
        try:
            response = requests.get(image_url,
                                    headers=HEADERS,
                                    timeout=UCK_TIMEOUT)
            # avoid redirected URL
            image_url = response.url
            # either exception or wrong HTTP code
            if response.status_code >= 400:
                raise Exception('Response code %s' % response.status_code)
        except Exception as k:
            logger.info('%s for %s' % (str(k), str(image_url)))
            try:
                # CDN URL could be formed as http:/xxxx.jpg
                path = re.split('https?://?', image_url)[-1]
                scheme = requests.utils.urlparse(image_url).scheme
                image_url = '%s://%s' % (scheme, path)

                response = requests.get(image_url,
                                        headers=HEADERS,
                                        timeout=UCK_TIMEOUT)
                # avoid redirected URL
                image_url = response.url
                if response.status_code >= 400:
                    raise Exception('Response code %s' % response.status_code)
            except Exception as k:
                logger.error('%s for %s' % (str(k), str(image_url)))
                raise Exception('%s for %s' % (str(k), str(image_url)))

        if response and response.status_code < 400 and response.content:
            # GIF is not supported yet
            #pr = requests.utils.urlparse(image_url)
            #image_url_address = pr.netloc + pr.path
            # if image_url_address.lower().endswith('.gif'):
            #    raise Exception('GIF is not supported! %s' % str(image_url))
            # else:
            image_html = response.content
            image_url = self._check_image(image_url, image_html)
            return str(image_url), str(image_html)
        else:
            logger.error('Cannot parse %s' % str(image_url))
            raise Exception('Cannot parse %s' % str(image_url))
Exemplo n.º 3
0
    def _analyze(self, image_url=None, referer=None):
        """
        remove CDN prefix, if any; and read image data
        """
        if not image_url:
            logger.error('Method malformed!')
            raise Exception('Method malformed!')

        image_url = image_url.replace("\/", "/").strip()
        image_url = urllib2.unquote(hparser.unescape(image_url))

        # as the name could be http://xxx.com/yyy--http://zzz.jpg
        # or http://xxx.com/yyy--https://zzz.jpg
        last_http_index = image_url.rfind('http')
        image_url = image_url[last_http_index:]

        response = None
        if referer:
            HEADERS['Referer'] = referer
        try:
            response = requests.get(
                image_url, headers=HEADERS, timeout=UCK_TIMEOUT)
            # avoid redirected URL
            image_url = response.url
            # either exception or wrong HTTP code
            if response.status_code >= 400:
                raise Exception('Response code %s' % response.status_code)
        except Exception as k:
            logger.info('%s for %s' % (str(k), str(image_url)))
            try:
                # CDN URL could be formed as http:/xxxx.jpg
                path = re.split('https?://?', image_url)[-1]
                scheme = requests.utils.urlparse(image_url).scheme
                image_url = '%s://%s' % (scheme, path)

                response = requests.get(
                    image_url, headers=HEADERS, timeout=UCK_TIMEOUT)
                # avoid redirected URL
                image_url = response.url
                if response.status_code >= 400:
                    raise Exception('Response code %s' % response.status_code)
            except Exception as k:
                logger.error('%s for %s' % (str(k), str(image_url)))
                raise Exception('%s for %s' % (str(k), str(image_url)))

        if response and response.status_code < 400 and response.content:
            # GIF is not supported yet
            #pr = requests.utils.urlparse(image_url)
            #image_url_address = pr.netloc + pr.path
            # if image_url_address.lower().endswith('.gif'):
            #    raise Exception('GIF is not supported! %s' % str(image_url))
            # else:
            image_html = response.content
            image_url = self._check_image(image_url, image_html)
            return str(image_url), str(image_html)
        else:
            logger.error('Cannot parse %s' % str(image_url))
            raise Exception('Cannot parse %s' % str(image_url))
Exemplo n.º 4
0
def _transcode(link):
    """
    send link to uck and get the data
    """
    try:
        html = urllib2.urlopen('%s%s' % (UCK_TRANSCODING_NEW, link),
                               timeout=UCK_TIMEOUT).read()
        data = urllib2.unquote(hparser.unescape(html))
        return data
    except Exception as k:
        logger.info('Problem:[%s] Source:[%s]' % (str(k), link))
        return None
Exemplo n.º 5
0
def _transcode(link):
    """
    send link to uck and get the data
    """
    try:
        html = urllib2.urlopen(
            '%s%s' % (UCK_TRANSCODING_NEW, link), timeout=UCK_TIMEOUT).read()
        data = urllib2.unquote(hparser.unescape(html))
        return data
    except Exception as k:
        logger.info('Problem:[%s] Source:[%s]' % (str(k), link))
        return None
Exemplo n.º 6
0
def _transcode(link):
    """
    send link to uck server
    """
    try:
        uck_url = '%s%s' % (UCK_TRANSCODING, link)
        # timeout set to UCK_TIMEOUT, currently
        html = urllib2.urlopen(uck_url, timeout=UCK_TIMEOUT).read()
        # free data from html encoding
        data = urllib2.unquote(hparser.unescape(html))
        return data
    except Exception as k:
        logger.info('Problem:[%s] Source:[%s]' % (str(k), link))
        return None
Exemplo n.º 7
0
def _transcode(link):
    """
    send link to uck server
    """
    try:
        uck_url = '%s%s' % (UCK_TRANSCODING, link)
        # timeout set to UCK_TIMEOUT, currently
        html = urllib2.urlopen(uck_url, timeout=UCK_TIMEOUT).read()
        # free data from html encoding
        data = urllib2.unquote(hparser.unescape(html))
        return data
    except Exception as k:
        logger.info('Problem:[%s] Source:[%s]' % (str(k), link))
        return None
Exemplo n.º 8
0
def _save(data, path):
    """
    save the file on local disk and return web and local path
    """
    if not data or not path:
        logger.error('Method malformed!')
        return None, None

    try:
        local_path = '%s%s.html' % (TRANSCODED_LOCAL_DIR, path)
        web_path = '%s%s.html' % (TRANSCODED_PUBLIC_DIR, path)

        f = open(local_path, 'w')
        f.write(urllib2.unquote(hparser.unescape(data)).replace(u'\xa0', ' '))
        f.close()
        return web_path, local_path
    except Exception as k:
        logger.error(str(k))
        return None, None
Exemplo n.º 9
0
def _save(data, path):
    """
    save the file on local disk and return web and local path
    """
    if not data or not path:
        logger.error('Method malformed!')
        return None, None

    try:
        local_path = '%s%s.html' % (TRANSCODED_LOCAL_DIR, path)
        web_path = '%s%s.html' % (TRANSCODED_PUBLIC_DIR, path)

        f = open(local_path, 'w')
        f.write(urllib2.unquote(hparser.unescape(data)).replace(u'\xa0', ' '))
        f.close()
        return web_path, local_path
    except Exception as k:
        logger.error(str(k))
        return None, None
Exemplo n.º 10
0
def find_images(content=None, referer=None):
    """
    find out all images from content and its size info
    """
    if not content:
        logger.error('Content/HTML is found VOID!')
        return None, content

    try:
        if isinstance(content, str) or isinstance(content, unicode):
            soup = BeautifulSoup(content.decode('utf-8', 'ignore'))
            normalized_images = []

            element_replaced = False
            for image in soup.findAll('img'):
                if image.get('src'):
                    normalized_image = find_image(image.get('src'), referer)
                    if normalized_image:
                        # replace original image link with clean and (local)
                        # copy
                        if 'original_url' in normalized_image and \
                                normalized_image['original_url']:
                            image['src'] = str(normalized_image['url'])
                            element_replaced = True
                        normalized_images.append(normalized_image)

            content_new = soup.prettify(encoding='utf-8')
            if element_replaced and content_new:
                content = str(
                    html_slimmer(
                        urllib2.unquote(hparser.unescape(content_new))))
            return normalized_images, content
        else:
            logger.info("Wrong format %s" % content)
            return None, content
    except Exception as k:
        logger.error("Problem [%s] Source [%s]" % (str(k), content))
        return None, content
Exemplo n.º 11
0
def find_images(content=None, referer=None):
    """
    find out all images from content and its size info
    """
    if not content:
        logger.error('Content/HTML is found VOID!')
        return None, content

    try:
        if isinstance(content, str) or isinstance(content, unicode):
            soup = BeautifulSoup(content.decode('utf-8', 'ignore'))
            normalized_images = []

            element_replaced = False
            for image in soup.findAll('img'):
                if image.get('src'):
                    normalized_image = find_image(image.get('src'), referer)
                    if normalized_image:
                        # replace original image link with clean and (local)
                        # copy
                        if 'original_url' in normalized_image and \
                                normalized_image['original_url']:
                            image['src'] = str(normalized_image['url'])
                            element_replaced = True
                        normalized_images.append(normalized_image)

            content_new = soup.prettify(encoding='utf-8')
            if element_replaced and content_new:
                content = str(
                    html_slimmer(urllib2.unquote(
                        hparser.unescape(content_new))))
            return normalized_images, content
        else:
            logger.info("Wrong format %s" % content)
            return None, content
    except Exception as k:
        logger.error("Problem [%s] Source [%s]" % (str(k), content))
        return None, content
Exemplo n.º 12
0
def prepare_link(url):
    """
    decode with the correct encoding
    """
    if not url:
        logger.error('Method malformed!')
        return None

    try:
        resp = requests.get(url, timeout=UCK_TIMEOUT)
        html = resp.content if resp.ok else None
        if html:
            detected = chardet.detect(html)
            encoding = detected['encoding'] if detected else 'utf-8'
            encoding = 'windows-1252' if 'folha.uol.com.br' in url else encoding
            data = html.decode(encoding, 'ignore')
            return hparser.unescape(urllib2.unquote(data)).replace(u'\xa0', ' ')
        else:
            logger.warning("Cannot read %s" % url)
            return None
    except Exception as k:
        logger.info('Problem:[%s] Source:[%s]' % (str(k), url))
        return None
Exemplo n.º 13
0
def prepare_link(url):
    """
    decode with the correct encoding
    """
    if not url:
        logger.error('Method malformed!')
        return None

    try:
        resp = requests.get(url, timeout=UCK_TIMEOUT)
        html = resp.content if resp.ok else None
        if html:
            detected = chardet.detect(html)
            encoding = detected['encoding'] if detected else 'utf-8'
            encoding = 'windows-1252' if 'folha.uol.com.br' in url else encoding
            data = html.decode(encoding, 'ignore')
            return hparser.unescape(urllib2.unquote(data)).replace(
                u'\xa0', ' ')
        else:
            logger.warning("Cannot read %s" % url)
            return None
    except Exception as k:
        logger.info('Problem:[%s] Source:[%s]' % (str(k), url))
        return None
Exemplo n.º 14
0
def parse(feed_link=None, feed_id=None, feed_title=None, language=None,
          categories=None, etag=None, modified=None):
    """
    read rss/atom data from a given feed
    feed_id is the feed ObjectId in MongoDB
    Etag and Modified are used to save rss http server's bandwidth
    Note: category should be added to feed table/database
    """
    if not feed_link or not feed_id or not language or not categories:
        logger.error("Method malformed!")
        return None, None, feed_title, etag, modified, "Method malformed!"
    if language not in LANGUAGES:
        logger.error("Language not supported for %s!" % feed_link)
        return None, None, feed_title, etag, modified, "Language not "
        "supported for %s!" % feed_link

    def _validate_time(entry):
        """
        see if the entry's updated time is earlier than needed
        """
        deadline = datetime.utcfromtimestamp(
            entry['updated']) + timedelta(days=DATABASE_REMOVAL_DAYS)
        return True if deadline > datetime.now() else False

    try:
        # variables d and e follow feedparser tradition
        feedparser.USER_AGENT = "newsman"
        d = feedparser.parse(feed_link, etag=etag, modified=modified)
        if d:
            # http://pythonhosted.org/feedparser/reference-status.html
            # http://pythonhosted.org/feedparser/http-etag.html#http-etag
            status = d.status if 'status' in d else None

            if status == 301:
                logger.critical(
                    '%s has been permantently moved to a %s!' % (
                        feed_link, d.href))
                return None, status, feed_title, etag, modified, '%s has been '
                'permantently moved to a %s!' % (
                    feed_link, d.href)
            elif status == 304:
                logger.warning(
                    '%s server has not updated its feeds' % feed_link)
                return None, status, feed_title, etag, modified, '%s server '
                'has not updated its feeds' % feed_link
            elif status == 410:
                logger.critical(
                    '%s is gone! Admin should check the feed availability!' %
                    feed_link)
                return None, status, feed_title, etag, modified, '%s is gone! '
                'Admin should check the feed availability!' % feed_link
            elif status == 200 or status == 302:
                # no need to worry.
                if status == 302:
                    logger.info(
                        '%s url has been temp moved to a new place' % feed_link)

                if not feed_title:
                    # if title were not found in feed, an AttributeError would
                    # be raised.
                    feed_title = urllib2.unquote(
                        hparser.unescape(d.feed.title)).strip()
                else:
                    feed_title = feed_title.strip()
                    if 'title' in d.feed:
                        feed_title_latest = urllib2.unquote(
                            hparser.unescape(d.feed.title)).strip()
                        if feed_title != feed_title_latest:
                            # change feed title
                            logger.info(
                                '%s title changed! Please update feed '
                                'table/database' % feed_link)
                            logger.info('old title: %s' % feed_title)
                            logger.info('new title: %s' % feed_title_latest)
                            #feed_title = feed_title_latest
                    else:
                        logger.info(
                            '%s[%s] has no title in its latest RSS' % (
                                feed_title, feed_link))

                # update etag/modified
                etag = None
                modified = None
                try:
                    etag = d.etag
                except AttributeError:
                    try:
                        modified = d.modified
                    except AttributeError:
                        pass

                if 'entries' in d:
                    language = language if 'language' not in d else d.language
                    # an Exception might be raised from _read_entry
                    entries = []
                    logger.error('%s begins processing' % feed_title)
                    for i, e in enumerate(d.entries):
                        if e:
                            entry = _read_entry(
                                e, feed_id, feed_title, language, categories)
                            if entry:
                                entries.append(entry)
                            else:
                                logger.info('Cannot parse %s' % e['link'])
                                continue
                        else:
                            logger.info(
                                'No infomation found for %s-th entry' % i)
                            continue

                    if entries:
                        # the FINAL return
                        # the last one indicates nothing wrong happended in
                        # parsing
                        return filter(_validate_time,
                                      entries), status, feed_title, etag, \
                               modified, 'OK'
                    else:
                        logger.info('Feed parsing goes wrong!')
                        return None, status, feed_title, etag, modified, \
                               'Feed parsing goes wrong!'
                else:
                    logger.info("Feed %s has no items!" % feed_id)
                    return None, status, feed_title, etag, modified, 'Feed %s '
                    'has no items!' % feed_id
            else:
                logger.info(
                    'HTTP Error Code [%s] for %s' % (status, feed_link))
                return None, status, feed_title, etag, modified, 'HTTP Error '
                'Code [%s] for %s' % (
                    status, feed_link)
        else:
            logger.info("Cannot parse %s correctly!" % feed_id)
            return None, None, feed_title, etag, modified, "Cannot parse %s "
            "correctly!" % feed_id
    except Exception as k:
        logger.exception('%s for %s' % (str(k), feed_id))
        return None, None, feed_title, etag, modified, '%s for %s' % (
            str(k), feed_id)
Exemplo n.º 15
0
def _read_entry(e=None, feed_id=None, feed_title=None, language=None,
                categories=None):
    """
    read a specific entry item from a feed 
    Note. categories are ids of category item
    """
    if not e or not feed_title or not language or not categories:
        logger.error('Method malformed!')
        return None
    if language not in LANGUAGES:
        logger.error("Language not supported for %s!" % feed_title)
        return None

    try:
        entry = {}
        entry['feed_id'] = feed_id
        entry['feed'] = feed_title.strip()
        entry['language'] = language.strip()
        entry['categories'] = categories

        # the easy part: the must-have
        entry['error'] = []

        # article original link
        if e.link:
            original_link = e.link.strip()
            if not original_link.startswith(AD_LINKS):
                # print 'original', original_link
                # print 'unescaped', hparser.unescape(original_link)
                # print 'unquoted', urllib2.unquote(original_link)
                # print 'unescaped-unquoted', urllib2.unquote(hparser
                # .unescape(original_link))
                # print 'unquoted-unescaped', hparser.unescape(urllib2
                # .unquote(original_link))
                # find the real link from redirection
                # the sequence of the following two steps are IMPORTANT!
                original_link = _find_redirected_link(original_link)
                # print 'anti-redirected', original_link
                # clean the URL
                # original_link = urllib2.unquote(hparser.unescape(
                # original_link))
                # print 'unescaped-unquoted', original_link
                # print '------------------------------------------------'


                # find the redirected link
                matched_prefix = [
                    link for link in HIDDEN_LINKS if
                    original_link.startswith(link)]
                found_prefix = matched_prefix[0] if matched_prefix else None
                if found_prefix:
                    actual_link = _get_actual_link(found_prefix, original_link)
                    if actual_link:
                        entry['link'] = actual_link
                    else:
                        logger.error(
                            'No actual link found for %s!' % original_link)
                        return None
                else:
                    entry['link'] = original_link
            else:
                logger.info('Advertising link %s' % original_link)
                return None
        else:
            logger.info('Feed malformed! No link found!')
            return None

        # article title
        if e.title_detail.type != 'text/plain':
            entry['title'] = urllib2.unquote(hparser.unescape(e.title.strip()))
        elif 'title' in e:
            entry['title'] = e.title.strip()
        else:
            entry['title'] = None
        # remove possible htmlized title
        entry['title'] = re.sub("<.*?>", " ", entry[
            'title']) if 'title' in entry and entry['title'] else None

        # article published time
        # first try parsed time info
        try:
            entry['updated'] = calendar.timegm(e.updated_parsed)
            entry['updated_human'] = e.updated
        except AttributeError as k:
            try:
                entry['updated'] = calendar.timegm(e.published_parsed)
                entry['updated_human'] = e.published
            except AttributeError as k:
                entry['error'] = ['%s\n%s' % (
                    entry['error'],
                    "no 'updated_parsed' or 'published_parsed'")]
                # then try unparsed time info
                # this is rarely possible.
                try:
                    updated = e.updated if 'updated' in e else e.published
                    if updated:
                        # get time zone
                        offset = int(updated[-5:])
                        delta = timedelta(hours=int(offset) / 100)
                        format = "%a, %d %b %Y %H:%M:%S"
                        if updated[-8:-5] != 'UTC':
                            updated = datetime.strptime(updated[:-6], format)
                        else:
                            updated = datetime.strptime(updated[:-9], format)
                        updated -= delta
                        entry['updated'] = time.mktime(updated.timetuple())
                    else:
                        logger.info(
                            "Attribute updated/published has no value")
                        return None
                except ValueError as k:
                    logger.info(str(k))
                    entry['error'].append('%s\n%s' % (entry['error'], k))
                    return None
                except AttributeError as k:
                    logger.info(str(k))
                    entry['error'].append('no update or published\n')
                    return None

        # article's summary
        try:
            # its possible summary is html-based
            summary = urllib2.unquote(hparser.unescape(e.summary))
            if isinstance(summary, str):
                summary_encoding = chardet.detect(summary)['encoding']
                summary = summary.decode(summary_encoding, 'ignore')
            # a <div, for example, and a </div
            is_html = True if len(
                re.findall(u'</?a|</?p|</?strong|</?img|</?html|</?div',
                           summary)) > 1 else False
            if is_html:
                h = html2text.HTML2Text()
                h.ignore_images = True
                h.ignore_links = True
                h.ignore_emphasis = True
                paragraphs = (h.handle(summary)).strip().strip(
                    '#').strip().split('\n\n')
                paragraphs_above_limit = []
                # remove paragraphs that contain less than x number of words
                for paragraph in paragraphs:
                    if entry['language'].startswith('zh') or entry[
                        'language'] == 'ja':
                        if len(paragraph) > 18:
                            paragraphs_above_limit.append(paragraph)
                    else:
                        words = paragraph.split()
                        if len(words) > 12:
                            paragraphs_above_limit.append(paragraph)
                entry['summary'] = '\n\n'.join(paragraphs_above_limit)
            else:
                entry['summary'] = summary
        except AttributeError as k:
            entry['summary'] = None
        entry['summary'] = None if not entry['summary'] else entry['summary']

        # article's images
        # e.g. [{'url':'http://image.com/tests.jpg, 'width': u'130', 'height':
        # u'86'}]
        entry['images'] = []
        try:
            images, media_content_new = illustrator.find_images(
                e.media_content, entry['link'])
            if images:
                entry['images'].extend(images)
        except AttributeError as k:
            pass
        try:
            images, media_content_new = illustrator.find_images(
                e.media_thumbnail, entry['link'])
            if images:
                entry['images'].extend(images)
        except AttributeError as k:
            pass
        for attribute in e:
            if 'thumbnail' in attribute:
                # currently set thumbnail to None if its a dictionary
                image = e[attribute] if isinstance(e[attribute], str) else None
                image = illustrator.find_image(image, entry['link'])
                if image:
                    entry['images'].append(image)
        try:
            links = e.links
            for link in links:
                if 'type' in link and 'image' in link.type:
                    if 'href' in link:
                        image = illustrator.find_image(
                            link.href, entry['link'])
                        if image:
                            entry['images'].append(image)
        except AttributeError as k:
            pass

        if entry.has_key('summary') and entry['summary']:
            images, entry['summary'] = illustrator.find_images(
                entry['summary'], entry['link'])
            if images:
                entry['images'].extend(images)
        # dedup images is processed at rss.py

        # article's author
        # e.g. Yuan Jin
        try:
            # i guess this could be a string or a list
            entry['author'] = e.author
        except AttributeError as k:
            entry['author'] = None

        # article's source
        # e.g. {'href': u'http://www.reuters.com/', 'title': u'Reuters'}
        try:
            entry['source'] = e.source
        except AttributeError as k:
            entry['source'] = None

        # article's tags
        # e.g. [{'term': u'Campus Party', 'scheme': None, 'label': None}]
        # term is usually combined with scheme to form a url; label is
        # the name of term
        try:
            entry['tags'] = e.tag
        except AttributeError as k:
            entry['tags'] = None

        # the FINAL return
        return entry
    except Exception as k:
        logger.error(str(k))
        return None
Exemplo n.º 16
0
def _read_entry(e=None,
                feed_id=None,
                feed_title=None,
                language=None,
                categories=None):
    """
    read a specific entry item from a feed 
    Note. categories are ids of category item
    """
    if not e or not feed_title or not language or not categories:
        logger.error('Method malformed!')
        return None
    if language not in LANGUAGES:
        logger.error("Language not supported for %s!" % feed_title)
        return None

    try:
        entry = {}
        entry['feed_id'] = feed_id
        entry['feed'] = feed_title.strip()
        entry['language'] = language.strip()
        entry['categories'] = categories

        # the easy part: the must-have
        entry['error'] = []

        # article original link
        if e.link:
            original_link = e.link.strip()
            if not original_link.startswith(AD_LINKS):
                # print 'original', original_link
                # print 'unescaped', hparser.unescape(original_link)
                # print 'unquoted', urllib2.unquote(original_link)
                # print 'unescaped-unquoted', urllib2.unquote(hparser
                # .unescape(original_link))
                # print 'unquoted-unescaped', hparser.unescape(urllib2
                # .unquote(original_link))
                # find the real link from redirection
                # the sequence of the following two steps are IMPORTANT!
                original_link = _find_redirected_link(original_link)
                # print 'anti-redirected', original_link
                # clean the URL
                # original_link = urllib2.unquote(hparser.unescape(
                # original_link))
                # print 'unescaped-unquoted', original_link
                # print '------------------------------------------------'

                # find the redirected link
                matched_prefix = [
                    link for link in HIDDEN_LINKS
                    if original_link.startswith(link)
                ]
                found_prefix = matched_prefix[0] if matched_prefix else None
                if found_prefix:
                    actual_link = _get_actual_link(found_prefix, original_link)
                    if actual_link:
                        entry['link'] = actual_link
                    else:
                        logger.error('No actual link found for %s!' %
                                     original_link)
                        return None
                else:
                    entry['link'] = original_link
            else:
                logger.info('Advertising link %s' % original_link)
                return None
        else:
            logger.info('Feed malformed! No link found!')
            return None

        # article title
        if e.title_detail.type != 'text/plain':
            entry['title'] = urllib2.unquote(hparser.unescape(e.title.strip()))
        elif 'title' in e:
            entry['title'] = e.title.strip()
        else:
            entry['title'] = None
        # remove possible htmlized title
        entry['title'] = re.sub(
            "<.*?>", " ",
            entry['title']) if 'title' in entry and entry['title'] else None

        # article published time
        # first try parsed time info
        try:
            entry['updated'] = calendar.timegm(e.updated_parsed)
            entry['updated_human'] = e.updated
        except AttributeError as k:
            try:
                entry['updated'] = calendar.timegm(e.published_parsed)
                entry['updated_human'] = e.published
            except AttributeError as k:
                entry['error'] = [
                    '%s\n%s' % (entry['error'],
                                "no 'updated_parsed' or 'published_parsed'")
                ]
                # then try unparsed time info
                # this is rarely possible.
                try:
                    updated = e.updated if 'updated' in e else e.published
                    if updated:
                        # get time zone
                        offset = int(updated[-5:])
                        delta = timedelta(hours=int(offset) / 100)
                        format = "%a, %d %b %Y %H:%M:%S"
                        if updated[-8:-5] != 'UTC':
                            updated = datetime.strptime(updated[:-6], format)
                        else:
                            updated = datetime.strptime(updated[:-9], format)
                        updated -= delta
                        entry['updated'] = time.mktime(updated.timetuple())
                    else:
                        logger.info("Attribute updated/published has no value")
                        return None
                except ValueError as k:
                    logger.info(str(k))
                    entry['error'].append('%s\n%s' % (entry['error'], k))
                    return None
                except AttributeError as k:
                    logger.info(str(k))
                    entry['error'].append('no update or published\n')
                    return None

        # article's summary
        try:
            # its possible summary is html-based
            summary = urllib2.unquote(hparser.unescape(e.summary))
            if isinstance(summary, str):
                summary_encoding = chardet.detect(summary)['encoding']
                summary = summary.decode(summary_encoding, 'ignore')
            # a <div, for example, and a </div
            is_html = True if len(
                re.findall(u'</?a|</?p|</?strong|</?img|</?html|</?div',
                           summary)) > 1 else False
            if is_html:
                h = html2text.HTML2Text()
                h.ignore_images = True
                h.ignore_links = True
                h.ignore_emphasis = True
                paragraphs = (
                    h.handle(summary)).strip().strip('#').strip().split('\n\n')
                paragraphs_above_limit = []
                # remove paragraphs that contain less than x number of words
                for paragraph in paragraphs:
                    if entry['language'].startswith(
                            'zh') or entry['language'] == 'ja':
                        if len(paragraph) > 18:
                            paragraphs_above_limit.append(paragraph)
                    else:
                        words = paragraph.split()
                        if len(words) > 12:
                            paragraphs_above_limit.append(paragraph)
                entry['summary'] = '\n\n'.join(paragraphs_above_limit)
            else:
                entry['summary'] = summary
        except AttributeError as k:
            entry['summary'] = None
        entry['summary'] = None if not entry['summary'] else entry['summary']

        # article's images
        # e.g. [{'url':'http://image.com/tests.jpg, 'width': u'130', 'height':
        # u'86'}]
        entry['images'] = []
        try:
            images, media_content_new = illustrator.find_images(
                e.media_content, entry['link'])
            if images:
                entry['images'].extend(images)
        except AttributeError as k:
            pass
        try:
            images, media_content_new = illustrator.find_images(
                e.media_thumbnail, entry['link'])
            if images:
                entry['images'].extend(images)
        except AttributeError as k:
            pass
        for attribute in e:
            if 'thumbnail' in attribute:
                # currently set thumbnail to None if its a dictionary
                image = e[attribute] if isinstance(e[attribute], str) else None
                image = illustrator.find_image(image, entry['link'])
                if image:
                    entry['images'].append(image)
        try:
            links = e.links
            for link in links:
                if 'type' in link and 'image' in link.type:
                    if 'href' in link:
                        image = illustrator.find_image(link.href,
                                                       entry['link'])
                        if image:
                            entry['images'].append(image)
        except AttributeError as k:
            pass

        if entry.has_key('summary') and entry['summary']:
            images, entry['summary'] = illustrator.find_images(
                entry['summary'], entry['link'])
            if images:
                entry['images'].extend(images)
        # dedup images is processed at rss.py

        # article's author
        # e.g. Yuan Jin
        try:
            # i guess this could be a string or a list
            entry['author'] = e.author
        except AttributeError as k:
            entry['author'] = None

        # article's source
        # e.g. {'href': u'http://www.reuters.com/', 'title': u'Reuters'}
        try:
            entry['source'] = e.source
        except AttributeError as k:
            entry['source'] = None

        # article's tags
        # e.g. [{'term': u'Campus Party', 'scheme': None, 'label': None}]
        # term is usually combined with scheme to form a url; label is
        # the name of term
        try:
            entry['tags'] = e.tag
        except AttributeError as k:
            entry['tags'] = None

        # the FINAL return
        return entry
    except Exception as k:
        logger.error(str(k))
        return None
Exemplo n.º 17
0
def parse(feed_link=None,
          feed_id=None,
          feed_title=None,
          language=None,
          categories=None,
          etag=None,
          modified=None):
    """
    read rss/atom data from a given feed
    feed_id is the feed ObjectId in MongoDB
    Etag and Modified are used to save rss http server's bandwidth
    Note: category should be added to feed table/database
    """
    if not feed_link or not feed_id or not language or not categories:
        logger.error("Method malformed!")
        return None, None, feed_title, etag, modified, "Method malformed!"
    if language not in LANGUAGES:
        logger.error("Language not supported for %s!" % feed_link)
        return None, None, feed_title, etag, modified, "Language not "
        "supported for %s!" % feed_link

    def _validate_time(entry):
        """
        see if the entry's updated time is earlier than needed
        """
        deadline = datetime.utcfromtimestamp(
            entry['updated']) + timedelta(days=DATABASE_REMOVAL_DAYS)
        return True if deadline > datetime.now() else False

    try:
        # variables d and e follow feedparser tradition
        feedparser.USER_AGENT = "newsman"
        d = feedparser.parse(feed_link, etag=etag, modified=modified)
        if d:
            # http://pythonhosted.org/feedparser/reference-status.html
            # http://pythonhosted.org/feedparser/http-etag.html#http-etag
            status = d.status if 'status' in d else None

            if status == 301:
                logger.critical('%s has been permantently moved to a %s!' %
                                (feed_link, d.href))
                return None, status, feed_title, etag, modified, '%s has been '
                'permantently moved to a %s!' % (feed_link, d.href)
            elif status == 304:
                logger.warning('%s server has not updated its feeds' %
                               feed_link)
                return None, status, feed_title, etag, modified, '%s server '
                'has not updated its feeds' % feed_link
            elif status == 410:
                logger.critical(
                    '%s is gone! Admin should check the feed availability!' %
                    feed_link)
                return None, status, feed_title, etag, modified, '%s is gone! '
                'Admin should check the feed availability!' % feed_link
            elif status == 200 or status == 302:
                # no need to worry.
                if status == 302:
                    logger.info('%s url has been temp moved to a new place' %
                                feed_link)

                if not feed_title:
                    # if title were not found in feed, an AttributeError would
                    # be raised.
                    feed_title = urllib2.unquote(hparser.unescape(
                        d.feed.title)).strip()
                else:
                    feed_title = feed_title.strip()
                    if 'title' in d.feed:
                        feed_title_latest = urllib2.unquote(
                            hparser.unescape(d.feed.title)).strip()
                        if feed_title != feed_title_latest:
                            # change feed title
                            logger.info('%s title changed! Please update feed '
                                        'table/database' % feed_link)
                            logger.info('old title: %s' % feed_title)
                            logger.info('new title: %s' % feed_title_latest)
                            #feed_title = feed_title_latest
                    else:
                        logger.info('%s[%s] has no title in its latest RSS' %
                                    (feed_title, feed_link))

                # update etag/modified
                etag = None
                modified = None
                try:
                    etag = d.etag
                except AttributeError:
                    try:
                        modified = d.modified
                    except AttributeError:
                        pass

                if 'entries' in d:
                    language = language if 'language' not in d else d.language
                    # an Exception might be raised from _read_entry
                    entries = []
                    logger.error('%s begins processing' % feed_title)
                    for i, e in enumerate(d.entries):
                        if e:
                            entry = _read_entry(e, feed_id, feed_title,
                                                language, categories)
                            if entry:
                                entries.append(entry)
                            else:
                                logger.info('Cannot parse %s' % e['link'])
                                continue
                        else:
                            logger.info('No infomation found for %s-th entry' %
                                        i)
                            continue

                    if entries:
                        # the FINAL return
                        # the last one indicates nothing wrong happended in
                        # parsing
                        return filter(_validate_time,
                                      entries), status, feed_title, etag, \
                               modified, 'OK'
                    else:
                        logger.info('Feed parsing goes wrong!')
                        return None, status, feed_title, etag, modified, \
                               'Feed parsing goes wrong!'
                else:
                    logger.info("Feed %s has no items!" % feed_id)
                    return None, status, feed_title, etag, modified, 'Feed %s '
                    'has no items!' % feed_id
            else:
                logger.info('HTTP Error Code [%s] for %s' %
                            (status, feed_link))
                return None, status, feed_title, etag, modified, 'HTTP Error '
                'Code [%s] for %s' % (status, feed_link)
        else:
            logger.info("Cannot parse %s correctly!" % feed_id)
            return None, None, feed_title, etag, modified, "Cannot parse %s "
            "correctly!" % feed_id
    except Exception as k:
        logger.exception('%s for %s' % (str(k), feed_id))
        return None, None, feed_title, etag, modified, '%s for %s' % (str(k),
                                                                      feed_id)