Ejemplo n.º 1
0
def retrieveEspnGameIdsForDay(date, root_url):
    """

    """
    url = root_url + date
    url = "http://scores.espn.go.com/nba/scoreboard?date=20150202"
    scores_page = soupypages.soupFromUrl(url)
    if scores_page['pass']:
        scores_page = scores_page['soup']
        game_data = scores_page.body.findAll(id="main-container")[0].find('div', class_='scoreboards').attrs['data-data']
        game_data = re.subn('true', "'true'", game_data)[0]
        game_data = re.subn('false', "'false'", game_data)[0]
        game_data = eval(game_data)
        game_ids = [game['id'] for game in game_data['events']]
        return game_ids
    
##    scores_page = makePage(url, hdr=True)
##    if scores_page['pass']:
##        game_ids = key_phrase.findall(scores_page['page'])
##        return game_ids
    
    else:
        err_msg = scores_page['err']
        if type(err_msg) == str:
            print("Failed to retrieve scores page for %s. Error: %s" %\
                  (date, err_msg))
        else:
            print("Failed to retrieve scores page for %s. Unknown error." %\
                  date)
        return []
Ejemplo n.º 2
0
def retrieveEspnGameIdsForDay(date, root_url):
    """

    """
    url = root_url + date
    
##    url = "http://scores.espn.go.com/nba/scoreboard?date=20150202"

##    scores_page = soupypages.soupFromUrl(url)
##    if scores_page['pass']:
##        scores_page = scores_page['soup']
##        game_data = scores_page.body.findAll(id="main-container")[0].find('div', class_='scoreboards').attrs['data-data']
##        game_data = re.subn('true', "'true'", game_data)[0]
##        game_data = re.subn('false', "'false'", game_data)[0]
##        game_data = eval(game_data)
##        game_ids = [game['id'] for game in game_data['events']]
##        return game_ids

    scores_page = soupypages.soupFromUrl(url)
    if scores_page['pass']:
        scores_page = scores_page['soup']
        scripts = scores_page.head.findAll('script')
        game_data = [s for s in scripts if s.contents and s.contents[0].startswith('window.espn.scoreboardData')][0]
        game_data = re.search(r'{.*}', game_data.contents[0]).group()
        game_data = re.subn('true', '"true"', game_data)[0]
        game_data = re.subn('false', '"false"', game_data)[0]
        game_data = re.subn(r'null', '""', game_data)[0]
        game_data = game_data.split(';window.espn.scoreboardSettings')[0]
        game_data = eval(game_data)
        game_ids = [game['id'] for game in game_data['events']]
        return(game_ids)
    

    
##    scores_page = makePage(url, hdr=True)
##    if scores_page['pass']:
##        game_ids = key_phrase.findall(scores_page['page'])
##        return game_ids
    
    else:
        err_msg = scores_page['err']
        if type(err_msg) == str:
            print("Failed to retrieve scores page for %s. Error: %s" %\
                  (date, err_msg))
        else:
            print("Failed to retrieve scores page for %s. Unknown error." %\
                  date)
        return []
def retrieveEspnGameIdsForDay(url):

    page = soupypages.soupFromUrl(url)
    
    if page['pass']:
        gid_info = page['soup'].find('div', id='nbaSSOuter')
        game_divs = gid_info.findAll('div', class_="Recap GameLine")
        game_ids = [gd['id'] for gd in game_divs]
        game_ids = [gid.strip('nbaGL') for gid in game_ids]
        return(game_ids)
    else:
        err_msg = page['err']
        if type(err_msg) == str:
            print("Failed to retrieve scores page for %s. Error: %s" %\
                  (date, err_msg))
        else:
            print("Failed to retrieve scores page for %s. Unknown error." %\
                  date)
        return []