def findMissingData(): GB_ID = 5 games = db_utils.customQuery( "SELECT game.title,game.id,game.alt_title,(SELECT JSON_ARRAYAGG(platform.short) FROM platform LEFT JOIN gameplatform on platform.id=gameplatform.platformID WHERE gameID=game.id) as 'platforms',gameplatform.* FROM `gameplatform` LEFT join game on gameID=game.id WHERE cover_platform_link is null and platformID=5" ) listpro = scrape_utils.findSuitableProxy() while (len(listpro) == 0): print("trying to extract proxylist") time.sleep(5) listpro = scrape_utils.findSuitableProxy() counter = 10 for count in range(0, len(games)): counter = counter - 1 if (counter == 0): listpro = scrape_utils.findSuitableProxy() while (len(listpro) == 0): print("trying to extract proxylist") time.sleep(5) listpro = scrape_utils.findSuitableProxy() counter = 10 game = games[count] title = re.sub('\(.*\)', '', game['title']) print('doing ' + str(game['id']) + ":" + title) data = scrape_utils.getGamefaqsDescAndImage(title, 'GAMEBOY', listpro) if (data != -1 and data is not None): db_utils.saveCoverPlatformLink(game['id'], GB_ID, data['img']) db_utils.saveDescription(game['id'], data['desc']) db_utils.saveWikiCoverLink(game['id'], data['img']) elif (game['alt_title'] is not None): data = scrape_utils.getGamefaqsDescAndImage( game['alt_title'], 'GAMEBOY', listpro) if (data != -1 and data is not None): db_utils.saveCoverPlatformLink(game['id'], GB_ID, data['img']) db_utils.saveDescription(game['id'], data['desc']) db_utils.saveWikiCoverLink(game['id'], data['img'])
def findMissingData(): games = db_utils.customQuery( "SELECT * FROM `game` WHERE description is null and game.id>4750;") listpro = scrape_utils.findSuitableProxy() while (len(listpro) == 0): print("trying to extract proxylist") listpro = scrape_utils.findSuitableProxy() for count in range(0, len(games)): game = games[count] if (r' (SMS)' in game['title']): cleanTitle = re.sub(' \(SMS\)', '', game['title']) data = scrape_utils.getGamefaqsDescAndImage( cleanTitle, 'SMS', listpro) elif (r'Starring' in game['title']): cleanTitleParts = re.split(' Starring', game['title']) cleanTitle = cleanTitleParts[1] + ": " + cleanTitleParts[0] data = scrape_utils.getGamefaqsDescAndImage( cleanTitle, 'SMS', listpro) elif ('Sonic' in game['title']): cleanTitle = re.sub('Sonic', 'Sonic The Hedgehog', game['title']) data = scrape_utils.getGamefaqsDescAndImage( cleanTitle, 'SMS', listpro) elif ('Solomon' in game['title']): cleanTitle = 'Solomon\'s Key' data = scrape_utils.getGamefaqsDescAndImage( cleanTitle, 'SMS', listpro) else: data = scrape_utils.getGamefaqsDescAndImage( game['title'], 'SMS', listpro) if (data != -1 and data is not None): db_utils.saveDescription(game['id'], data['desc']) db_utils.saveWikiCoverLink(game['id'], data['img']) db_utils.saveCoverPlatformLink(game['id'], 3, data['img'])
def findMissingData(): games = db_utils.customQuery( "SELECT * FROM `game` WHERE cover_wikipedia_link is null or description is null and game.id>4033;" ) listpro = scrape_utils.findSuitableProxy() for count in range(0, len(games)): game = games[count] data = scrape_utils.getGamefaqsDescAndImage(game['title'], 'NES', listpro) if (data != -1 and data is not None): db_utils.saveDescription(game['id'], data['desc']) db_utils.saveWikiCoverLink(game['id'], data['img'])
def resolveConflicts(conflicting_indexes, titles, developers, publishers, dateJP, dateEU, dateUS): #in database fix titles like final fantasy legends #delete info games on gb #check title with (GB) GB_ID = 5 print("conflicts nr", len(conflicting_indexes)) decisions = dict() newTitles = dict() for i in conflicting_indexes: if (i in decisions): continue titleClean = re.sub('(?<=[a-z]|[1-9])(JP|PAL|NA)', '', titles[i]) possibleTitles = cleanupTitles( re.split( '(?<=[a-z]|[1-9])([A-Z])|JP\/EU|NA|EU|JP|EU\/AU|\/|NA\/PAL', titleClean)) conflicts = db_utils.gameExistsMultiple(possibleTitles[0]) print("DECISION " + str(i) + " for title " + titles[i]) print("Considering " + possibleTitles[0]) for index in range(0, len(conflicts)): print("Enter " + str(index) + " to merge with: " + conflicts[index]['title'] + "(" + str(conflicts[index]['similarity']) + ")" + " - " + str(conflicts[index]['platforms'])) decisions[i] = input( "Enter new or newapp(newtitle to input custom title) or new to create new entry or NUMERICAL index to merge with existing: " ) while (not validDecision(decisions[i])): decisions[i] = input( "Invalid input, try again(new, newapp or index of game to merge): " ) if (decisions[i] == "newtitle"): newTitles[i] = input("Input new desired game title:") decisions[i] = "new" print(decisions) print(newTitles) counter = 20 for count in conflicting_indexes: if (count not in decisions): continue counter = counter - 1 if (counter == 0): listpro = scrape_utils.findSuitableProxy() while (len(listpro) == 0): print("trying to extract proxylist") time.sleep(5) listpro = scrape_utils.findSuitableProxy() counter = 20 try: print("doing iteration " + str(count) + " for title " + titles[count]) titleClean = re.sub('(?<=[a-z]|[1-9])(JP|PAL|NA)', '', titles[count]) possibleTitles = cleanupTitles( re.split( '(?<=[a-z]|[1-9])([A-Z])|JP\/EU|NA|EU|JP|EU\/AU|\/|NA\/PAL', titleClean)) conflicts = db_utils.gameExistsMultiple(possibleTitles[0]) if (count in newTitles): possibleTitles[0] = newTitles[count] #print(possibleTitles,dateJP[count],dateUS[count],dateEU[count],developers[count],publishers[count]) newGame = True gameDetails = [] gameDetails.append(GB_ID) decision = decisions[count] if (str(decision) == "new" or str(decision) == "newapp"): print("Creating new game out of conflict") if (str(decision) == "newapp"): newGameID = db_utils.insertGame(possibleTitles[0] + " (GB)") else: newGameID = db_utils.insertGame(possibleTitles[0]) if (type(newGameID) is list): gameDetails.append(newGameID[0]) else: gameDetails.append(newGameID) elif (str.isdigit(decision)): print("Merging with " + str(conflicts[int(decision)])) if r')' in conflicts[int(decision)]['title']: newTitle = re.sub('\)', '/GB)', conflicts[int(decision)]['title']) db_utils.customQuery( 'UPDATE game SET title="' + newTitle + '" WHERE game.id=' + str(conflicts[int(decision)]['gameid'])) gameDetails.append(conflicts[int(decision)]['gameid']) newGame = False if (len(possibleTitles) == 2 and possibleTitles[1] != ""): db_utils.customQuery('UPDATE game SET alt_title="' + possibleTitles[1] + '" WHERE id=' + str(gameDetails[1]) + ' AND alt_title is null') if (len(possibleTitles) == 3 and possibleTitles[2] != ""): db_utils.customQuery('UPDATE game SET alt_title2="' + possibleTitles[2] + '" WHERE id=' + str(gameDetails[1]) + ' AND alt_title2 is null') if (len(possibleTitles) == 4 and possibleTitles[3] != ""): db_utils.customQuery('UPDATE game SET alt_title2="' + possibleTitles[3] + '" WHERE id=' + str(gameDetails[1]) + ' AND alt_title2 is null') #depending on regions if (type(dateUS[count]) is float): gameDetails.append("Unreleased") else: gameDetails.append(dateUS[count]) if (type(dateEU[count]) is float): gameDetails.append("Unreleased") else: gameDetails.append(dateEU[count]) if (type(dateJP[count]) is float): gameDetails.append("Unreleased") else: gameDetails.append(dateJP[count]) gameDetails.append("") #dateGEN db_utils.insertGamePlatform(gameDetails) except Exception as e: print(str(e)) return try: if (newGame): if (type(publishers[count]) is not float): cleanPublishers = re.sub('\[.{0,3}\]', '', publishers[count]) publishersSplit = re.split( '(?<=[a-z])[A-Z].*|JP\/EU|NA|EU|JP|EU\/AU|\/|PAL|NA\/PAL|JP\/PAL|NA\/JP|NA\/EU', cleanPublishers) pubIDs = db_utils.insertPublishers(publishersSplit) db_utils.insertGamePublishers(gameDetails[1], pubIDs) if (type(developers[count]) is not float and developers[count] != '???'): devIDs = db_utils.insertDevelopers( re.split('(?<=[a-z]|[1-9])[A-Z]', developers[count])) db_utils.insertGameDevelopers(gameDetails[1], devIDs) infobox = scrape_utils.wikipediaInfoboxScraping( possibleTitles[0]) if (infobox is not None): db_utils.saveInfoboxData(infobox, gameDetails[1], platformID=GB_ID) if ('boxart' in infobox): gamefaqsScraping(possibleTitles[0], gameDetails[1], False, newGame) else: gamefaqsScraping(possibleTitles[0], gameDetails[1], True, newGame) else: gamefaqsScraping(possibleTitles[0], gameDetails[1], True, newGame) except Exception as e: print(str(e)) continue
def scrapeGBGames(): """Scrapes GB games info from wikipedia lists""" global GB_ID global listpro listpro = scrape_utils.findSuitableProxy() #listpro=set(['193.178.246.248:8080','27.98.206.187:3128']) if (len(listpro) == 0): print(listpro) return GB_ID = 5 url = r'https://en.wikipedia.org/wiki/List_of_Game_Boy_games' tables = pd.read_html(url) # Returns list of all tables on page games = tables[0] titles = games[games.columns[0]].tolist() developers = games[games.columns[1]].tolist() publishers = games[games.columns[2]].tolist() dateJP = games[games.columns[4]].tolist() dateUS = games[games.columns[6]].tolist() dateEU = games[games.columns[8]].tolist() conflicting_indexes = [] counter = 20 for count in range(len(titles) - 1, len(titles)): if (type(dateJP[count]) is float and type(dateEU[count]) is float and type(dateUS[count]) is float): continue counter = counter - 1 if (counter == 0): listpro = scrape_utils.findSuitableProxy() while (len(listpro) == 0): print("trying to extract proxylist") time.sleep(5) listpro = scrape_utils.findSuitableProxy() counter = 20 if (len(conflicting_indexes) > 0): print("CONFLICTING INDEXES ARE:") print(conflicting_indexes) if (count == len(titles) - 1): resolveConflicts(conflicting_indexes, titles, developers, publishers, dateJP, dateEU, dateUS) return if (titles[count] == ''): return print("doing iteration " + str(count) + " for title " + titles[count]) titleClean = re.sub('(?<=[a-z]|[1-9])(JP|PAL|NA)', '', titles[count]) possibleTitles = cleanupTitles( re.split( '(?<=[a-z]|[1-9])([A-Z])|JP\/EU|NA|EU|JP|EU\/AU|\/|NA\/PAL', titleClean)) #print(possibleTitles,dateJP[count],dateUS[count],dateEU[count],developers[count],publishers[count]) newGame = True gameDetails = [] gameDetails.append(GB_ID) conflicts = db_utils.gameExistsMultiple(possibleTitles[0]) if (conflicts != -1 and len(conflicts) > 0): conflicting_index = count conflicting_indexes.append(conflicting_index) continue else: newGameID = db_utils.insertGame(possibleTitles[0]) if (type(newGameID) is list): gameDetails.append(newGameID[0]) else: gameDetails.append(newGameID) if (len(possibleTitles) == 2 and possibleTitles[1] != ""): db_utils.customQuery('UPDATE game SET alt_title="' + possibleTitles[1] + '" WHERE id=' + str(gameDetails[1]) + ' AND alt_title is null') if (len(possibleTitles) == 3 and possibleTitles[2] != ""): db_utils.customQuery('UPDATE game SET alt_title2="' + possibleTitles[2] + '" WHERE id=' + str(gameDetails[1]) + ' AND alt_title2 is null') if (len(possibleTitles) == 4 and possibleTitles[3] != ""): db_utils.customQuery('UPDATE game SET alt_title2="' + possibleTitles[3] + '" WHERE id=' + str(gameDetails[1]) + ' AND alt_title2 is null') #depending on regions if (type(dateUS[count]) is float): gameDetails.append("Unreleased") else: gameDetails.append(dateUS[count]) if (type(dateEU[count]) is float): gameDetails.append("Unreleased") else: gameDetails.append(dateEU[count]) if (type(dateJP[count]) is float): gameDetails.append("Unreleased") else: gameDetails.append(dateJP[count]) gameDetails.append("") #dateGEN db_utils.insertGamePlatform(gameDetails) if (newGame): if (type(publishers[count]) is not float): cleanPublishers = re.sub('\[.{0,3}\]', '', publishers[count]) publishersSplit = re.split( '(?<=[a-z])[A-Z].*|JP\/EU|NA|EU|JP|EU\/AU|\/|PAL|NA\/PAL|JP\/PAL|NA\/JP|NA\/EU', cleanPublishers) pubIDs = db_utils.insertPublishers(publishersSplit) db_utils.insertGamePublishers(gameDetails[1], pubIDs) if (type(developers[count]) is not float and developers[count] != '???'): devIDs = db_utils.insertDevelopers( re.split('(?<=[a-z]|[1-9])[A-Z]', developers[count])) db_utils.insertGameDevelopers(gameDetails[1], devIDs) infobox = scrape_utils.wikipediaInfoboxScraping(possibleTitles[0]) if (infobox is not None): db_utils.saveInfoboxData(infobox, gameDetails[1], platformID=GB_ID) if ('boxart' in infobox): gamefaqsScraping(possibleTitles[0], gameDetails[1], False, newGame) else: gamefaqsScraping(possibleTitles[0], gameDetails[1], True, newGame) else: gamefaqsScraping(possibleTitles[0], gameDetails[1], True, newGame)
def scrapeSmsGames(): """Scrapes SMS games info from wikipedia lists""" global SMS_ID global listpro listpro = scrape_utils.findSuitableProxy() #listpro=set(['23.89.198.227:3128','206.189.238.81:80','54.39.97.250:3128','193.178.246.248:8080','27.98.206.187:3128']) if (len(listpro) == 0): print(listpro) return SMS_ID = 3 url = r'https://en.wikipedia.org/wiki/List_of_Master_System_games' tables = pd.read_html(url) # Returns list of all tables on page games = tables[0] titles = games[games.columns[0]].tolist() developers = games[games.columns[1]].tolist() publishers = games[games.columns[2]].tolist() date = games[games.columns[4]].tolist() for count in range(2, len(titles)): if (titles[count] == ''): return if (publishers[count] == "Tec Toy"): continue if (type(date[count]) is float or date[count] == "Unreleased"): continue if ("Unlicensed" in titles[count]): continue if (', The' in titles[count]): cleanTitle = cleanTitle = re.sub(', The', '', titles[count]) cleanTitle = 'The ' + cleanTitle cleanTitle = re.sub('\(.*\)|(?<=[a-z]|[1-9])[A-Z].*', '', cleanTitle) elif ('Cop' in titles[count]): cleanTitle = re.sub('\(.*\)', '', cleanTitle) else: cleanTitle = re.sub('\(.*\)|(?<=[a-z]|[1-9])[A-Z].*', '', titles[count]) cleanTitle = cleanTitle.replace(u'\xa0', '') newGame = True gameDetails = [] gameDetails.append(SMS_ID) existingID = db_utils.gameExists(cleanTitle) if (existingID != -1): var = input( "Enter n or new to create new entry on this console or anything else to merge with existing entry: " ) print("You entered " + str(var)) if (str(var) == "new" or str(var) == "n"): print("Creating new game out of conflict") newGameID = db_utils.insertGame(cleanTitle.strip() + " (SMS)") if (type(newGameID) is list): gameDetails.append(newGameID[0]) else: gameDetails.append(newGameID) else: gameDetails.append(existingID) newGame = False else: newGameID = db_utils.insertGame(cleanTitle.strip()) if (type(newGameID) is list): gameDetails.append(newGameID[0]) else: gameDetails.append(newGameID) #depending on regions gameDetails.append("") #dateUS gameDetails.append("") #dateEU gameDetails.append("") #dateJP gameDetails.append(date[count]) #dateGEN db_utils.insertGamePlatform(gameDetails) if (newGame): if (type(publishers[count]) is not float): cleanPublishers = re.sub('\[.{0,3}\]', '', publishers[count]) publishersSplit = re.split( '(?<=[a-z])[A-Z].*|JP\/EU|NA|EU|JP|EU\/AU', cleanPublishers) pubIDs = db_utils.insertPublishers(publishersSplit) db_utils.insertGamePublishers(gameDetails[1], pubIDs) if (type(developers[count]) is not float and developers[count] != '???'): devIDs = db_utils.insertDevelopers( developers[count].split(';')) db_utils.insertGameDevelopers(gameDetails[1], devIDs) infobox = scrape_utils.wikipediaInfoboxScraping(cleanTitle) if (infobox is not None): db_utils.saveInfoboxData(infobox, gameDetails[1]) if ('boxart' in infobox): gamefaqsScraping(cleanTitle, gameDetails[1], False, newGame) else: gamefaqsScraping(cleanTitle, gameDetails[1], True, newGame) else: gamefaqsScraping(cleanTitle, gameDetails[1], True, newGame)