def domain_shodan(domain): try: data = json.loads(shodandomainsearch(domain)) if 'matches' in data.keys(): for x in data['matches']: ip_d = (x['ip_str']) hostname_d = ((x['hostnames'])) for hs in hostname_d: out_hs = ("".join(map(str, hs))) domains_d = (x['domains']) for dn in domains_d: out_dn = ("".join(map(str, dn))) port_d = (x['port']) data_d = (x['data'].replace("\n", "")) location_d = (x['location']) save_data = domain_shodans_db(ip=(ip_d), host=(out_hs), domain=(out_dn), port=(port_d), data=(data_d), location=(location_d)) save_data.save() except Exception as error: print error
def main(): domain = sys.argv[1] API_URL = "https://www.censys.io/api/v1" #print cfg.zoomeyeuser #print WhoIs information print whoisnew(domain) print "\n-----------------------------\n" #print DNS Information dns_records = parse_dns_records(domain) for x in dns_records.keys(): print x if "No" in dns_records[x] and "Found" in dns_records[x]: print "\t%s" % (dns_records[x]) else: for y in dns_records[x]: print "\t%s" % (y) #print type(dns_records[x]) print "\n-----------------------------\n" #convert domain to reverse_domain for passing to checkpunkspider() reversed_domain = "" for x in reversed(domain.split(".")): reversed_domain = reversed_domain + "." + x reversed_domain = reversed_domain[1:] checkpunkspider(reversed_domain) print "\n-----------------------------\n" #make proper URL with domain. Check on ssl as well as 80. print "\t\t\t[+] Wapplyzing " + domain print "Hitting HTTP:\n", try: targeturl = "http://" + domain wappalyzeit(targeturl) except: print "[-] HTTP connection was unavailable" print "Hitting HTTPS:\n", try: targeturl = "https://" + domain wappalyzeit(targeturl) except: print "[-] HTTPS connection was unavailable" print "\n-----------------------------\n" #make Search github code for the given domain. print github_search(domain, 'Code') print "\n-----------------------------\n" #collecting emails for the domain and adding information in master email list. emailhunter(domain) print "\t\t\t[+] Finding Email Ids\n" for x in collected_emails: print str(x) + ", ", print "\n-----------------------------\n" dns_history = netcraft_domain_history(domain) for x in dns_history.keys(): print "%s: %s" % (dns_history[x], x) print "\n-----------------------------\n" #subdomains [to be called before pagelinks so as to avoid repititions.] print "\t\t\t[+] Finding Subdomains and appending\n" subdomains(domain) ##print "\t\t\t[+] Check_subdomains from wolframalpha" ##find_subdomains_from_wolfram(domain) print "\n-----------------------------\n" #domain pagelinks print "\t\t\t[+] Pagelinks\n" links = pagelinks(domain) for x in links: print x print "\n-----------------------------\n" #calling and printing subdomains after pagelinks. subdomains_from_netcraft(domain) print "\n\t\t\t[+] List of subdomains found\n" for sub in subdomain_list: print sub #wikileaks print "\t\t\t[+] Associated WikiLeaks\n" leaklinks = wikileaks(domain) for tl, lnk in leaklinks.items(): print "%s (%s)" % (lnk, tl) print "For all results, visit: " + 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date' % ( domain) print "\n-----------------------------\n" print "\t\t\t[+] Associated Forum Links\n" links = boardsearch_forumsearch(domain) for tl, lnk in links.items(): print "%s (%s)" % (lnk, tl) print "\n-----------------------------\n" results = check_ssl_htbsecurity(domain) if 'ERROR' in results.keys(): print results['ERROR'] elif 'TOKEN' in results.keys(): print 'Picking up One IP from bunch of IPs returned: %s' % results[ 'MULTIPLE_IPS'][0] results_new = check_ssl_htbsecurity(results['MULTIPLE_IPS'][0]) print "OverAll Rating: %s" % results_new['GRADE'] print 'Check https://www.htbridge.com/ssl/ for more information' for x in results_new['VALUE'].keys(): if str("[5]") in str(results_new['VALUE'][x]) or str("[3]") in str( results_new['VALUE'][x]): if x == 'httpHeaders': pass else: print results_new['VALUE'][x] else: print "OverAll Rating: %s" % results['GRADE'] for x in results['VALUE'].keys(): if str("[5]") in str(results['VALUE'][x]) or str("[3]") in str( results['VALUE'][x]): if x == 'httpHeaders': pass else: print results['VALUE'][x] print "\n-----------------------------\n" #checks results from zoomeye #filters need to be applied if cfg.zoomeyeuser != "" and cfg.zoomeyepass != "": zoomeye_results = search_zoomeye(domain) dict_zoomeye_results = json.loads(zoomeye_results) if 'matches' in dict_zoomeye_results.keys(): for x in dict_zoomeye_results['matches']: if x['site'].split('.')[-2] == domain.split('.')[-2]: print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % ( x['ip'], x['site'], x['title'], x['headers'].replace( "\n", ""), x['geoinfo']) print "\n-----------------------------\n" print "[+]\t Kicking off Censys Search. This may take a while.." censys_search(domain) for x in censys_list: print x print "\n-----------------------------\n" #checks for host ip and other details in shodan. ''' #Code for shodan Ip search. now we are doing Hostname search. #converts domain to IP. Prints a statement for the same. ip_addr = socket.gethostbyname(domain) #checks for information at shodan, and comes back with whatever available. ## need to apply filter here (lot of noise coming in) res_from_shodan = json.loads(shodansearch(ip_addr)) #print res_from_shodan for iterate_shodan_list in res_from_shodan['data']: print "ISP: %s \n Hosts: %s \n IP: %s \n Data: %s\n" % (iterate_shodan_list['isp'], iterate_shodan_list['hostnames'], iterate_shodan_list['ip_str'], iterate_shodan_list['data'].strip("\n")) print "\n-----------------------------\n" ''' res_from_shodan = json.loads(shodandomainsearch(domain)) if 'matches' in res_from_shodan.keys(): for x in res_from_shodan['matches']: print "IP: %s\nHosts: %s\nDomain: %s\nPort: %s\nData: %s\nLocation: %s\n" % ( x['ip_str'], x['hostnames'], x['domains'], x['port'], x['data'].replace("\n", ""), x['location']) print "-----------------------------\n"
def main(): domain = sys.argv[1] API_URL = "https://www.censys.io/api/v1" #print cfg.zoomeyeuser #print WhoIs information print whoisnew(domain) print "\n-----------------------------\n" #print DNS Information dns_records = parse_dns_records(domain) for x in dns_records.keys(): print x if "No" in dns_records[x] and "Found" in dns_records[x]: print "\t%s" % (dns_records[x]) else: for y in dns_records[x]: print "\t%s" % (y) #print type(dns_records[x]) print "\n-----------------------------\n" #convert domain to reverse_domain for passing to checkpunkspider() reversed_domain = "" for x in reversed(domain.split(".")): reversed_domain = reversed_domain + "." + x reversed_domain = reversed_domain[1:] checkpunkspider(reversed_domain) print "\n-----------------------------\n" #make proper URL with domain. Check on ssl as well as 80. print "\t\t\t[+] Wapplyzing " + domain print "Hitting HTTP:\n", try: targeturl = "http://" + domain wappalyzeit(targeturl) except: print "[-] HTTP connection was unavailable" print "Hitting HTTPS:\n", try: targeturl = "https://" + domain wappalyzeit(targeturl) except: print "[-] HTTPS connection was unavailable" print "\n-----------------------------\n" #make Search github code for the given domain. print github_search(domain, 'Code') print "\n-----------------------------\n" #collecting emails for the domain and adding information in master email list. emailhunter(domain) print "\t\t\t[+] Finding Email Ids\n" for x in collected_emails: print str(x) + ", ", print "\n-----------------------------\n" dns_history = netcraft_domain_history(domain) for x in dns_history.keys(): print "%s: %s" % (dns_history[x], x) print "\n-----------------------------\n" #subdomains [to be called before pagelinks so as to avoid repititions.] print "\t\t\t[+] Finding Subdomains and appending\n" subdomains(domain) ##print "\t\t\t[+] Check_subdomains from wolframalpha" ##find_subdomains_from_wolfram(domain) print "\n-----------------------------\n" #domain pagelinks print "\t\t\t[+] Pagelinks\n" links=pagelinks(domain) for x in links: print x print "\n-----------------------------\n" #calling and printing subdomains after pagelinks. subdomains_from_netcraft(domain) print "\n\t\t\t[+] List of subdomains found\n" for sub in subdomain_list: print sub #wikileaks print "\t\t\t[+] Associated WikiLeaks\n" leaklinks=wikileaks(domain) for tl,lnk in leaklinks.items(): print "%s (%s)" % (lnk, tl) print "For all results, visit: "+ 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date'%(domain) print "\n-----------------------------\n" print "\t\t\t[+] Associated Forum Links\n" links=boardsearch_forumsearch(domain) for tl,lnk in links.items(): print "%s (%s)" % (lnk, tl) print "\n-----------------------------\n" results = check_ssl_htbsecurity(domain) if 'ERROR' in results.keys(): print results['ERROR'] elif 'TOKEN' in results.keys(): print 'Picking up One IP from bunch of IPs returned: %s' % results['MULTIPLE_IPS'][0] results_new = check_ssl_htbsecurity(results['MULTIPLE_IPS'][0]) print "OverAll Rating: %s" % results_new['GRADE'] print 'Check https://www.htbridge.com/ssl/ for more information' for x in results_new['VALUE'].keys(): if str("[5]") in str(results_new['VALUE'][x]) or str("[3]") in str(results_new['VALUE'][x]): if x == 'httpHeaders': pass else: print results_new['VALUE'][x] else: print "OverAll Rating: %s" % results['GRADE'] for x in results['VALUE'].keys(): if str("[5]") in str(results['VALUE'][x]) or str("[3]") in str(results['VALUE'][x]): if x == 'httpHeaders': pass else: print results['VALUE'][x] print "\n-----------------------------\n" #checks results from zoomeye #filters need to be applied zoomeye_results = search_zoomeye(domain) dict_zoomeye_results = json.loads(zoomeye_results) if 'matches' in dict_zoomeye_results.keys(): for x in dict_zoomeye_results['matches']: if x['site'].split('.')[-2] == domain.split('.')[-2]: print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (x['ip'], x['site'], x['title'], x['headers'].replace("\n",""), x['geoinfo']) print "\n-----------------------------\n" print "[+]\t Kicking off Censys Search. This may take a while.." censys_search(domain) for x in censys_list: print x print "\n-----------------------------\n" #checks for host ip and other details in shodan. ''' #Code for shodan Ip search. now we are doing Hostname search. #converts domain to IP. Prints a statement for the same. ip_addr = socket.gethostbyname(domain) #checks for information at shodan, and comes back with whatever available. ## need to apply filter here (lot of noise coming in) res_from_shodan = json.loads(shodansearch(ip_addr)) #print res_from_shodan for iterate_shodan_list in res_from_shodan['data']: print "ISP: %s \n Hosts: %s \n IP: %s \n Data: %s\n" % (iterate_shodan_list['isp'], iterate_shodan_list['hostnames'], iterate_shodan_list['ip_str'], iterate_shodan_list['data'].strip("\n")) print "\n-----------------------------\n" ''' res_from_shodan = json.loads(shodandomainsearch(domain)) if 'matches' in res_from_shodan.keys(): for x in res_from_shodan['matches']: print "IP: %s\nHosts: %s\nDomain: %s\nPort: %s\nData: %s\nLocation: %s\n" % (x['ip_str'], x['hostnames'], x['domains'], x['port'], x['data'].replace("\n",""), x['location']) print "-----------------------------\n"
def do_everything(domain): dict_to_apend['targetname'] = domain API_URL = "https://www.censys.io/api/v1" #print cfg.zoomeyeuser #print WhoIs information whoisdata = whoisnew(domain) print whoisdata dict_to_apend['whois'] = whoisdata #print DNS Information dns_records = parse_dns_records(domain) #dict_to_apend['dns_records'] = dns_records > not working #bson.errors.InvalidDocument: Cannot encode object: <DNS IN A rdata: 54.208.84.166> for x in dns_records.keys(): print x if "No" in dns_records[x] and "Found" in dns_records[x]: print "\t%s" % (dns_records[x]) else: for y in dns_records[x]: print "\t%s" % (y) #print type(dns_records[x]) print colored(style.BOLD + '\n---> Finding Paste(s)..\n' + style.END, 'blue') if cfg.google_cse_key != "" and cfg.google_cse_key != "XYZ" and cfg.google_cse_cx != "" and cfg.google_cse_cx != "XYZ": total_results = google_search(domain, 1) if (total_results != 0 and total_results > 10): more_iters = (total_results / 10) if more_iters >= 10: print colored( style.BOLD + '\n---> Too many results, Daily API limit might exceed\n' + style.END, 'red') for x in xrange(1, more_iters + 1): google_search(domain, (x * 10) + 1) print "\n\n-----------------------------\n" else: print colored( style.BOLD + '\n[-] google_cse_key and google_cse_cx not configured. Skipping paste(s) search.\nPlease refer to http://datasploit.readthedocs.io/en/latest/apiGeneration/.\n' + style.END, 'red') #convert domain to reverse_domain for passing to checkpunkspider() reversed_domain = "" for x in reversed(domain.split(".")): reversed_domain = reversed_domain + "." + x reversed_domain = reversed_domain[1:] res = checkpunkspider(reversed_domain) if 'data' in res.keys() and len(res['data']) >= 1: dict_to_apend['punkspider'] = res['data'] print colored("[+] Few vulnerabilities found at Punkspider", 'green') for x in res['data']: print "==> ", x['bugType'] print "Method:", x['verb'].upper() print "URL:\n" + x['vulnerabilityUrl'] print "Param:", x['parameter'] else: print colored("[-] No Vulnerabilities found on PunkSpider", 'red') print colored( style.BOLD + '\n---> Wapplyzing web page of base domain:\n' + style.END, 'blue') wappalyze_results = {} #make proper URL with domain. Check on ssl as well as 80. print "Hitting HTTP:\n", try: targeturl = "http://" + domain list_of_techs = wappalyzeit(targeturl) wappalyze_results['http'] = list_of_techs except: print "[-] HTTP connection was unavailable" wappalyze_results['http'] = [] print "\nHitting HTTPS:\n", try: targeturl = "https://" + domain list_of_techs = wappalyzeit(targeturl) wappalyze_results['https'] = list_of_techs except: print "[-] HTTPS connection was unavailable" wappalyze_results['https'] = [] if len(wappalyze_results.keys()) >= 1: dict_to_apend['wappalyzer'] = wappalyze_results #make Search github code for the given domain. git_results = github_search(domain, 'Code') if git_results is not None: print git_results else: print colored("Sad! Nothing found on github", 'red') #collecting emails for the domain and adding information in master email list. if cfg.emailhunter != "": emails = emailhunter(domain) if len(collected_emails) >= 1: for x in collected_emails: print str(x) dict_to_apend['email_ids'] = collected_emails ''' ##### code for automated osint on enumerated email email_ids while True: a = raw_input(colored("\n\nDo you want to launch osint check for these emails? [(Y)es/(N)o/(S)pecificEmail]: ", 'red')) if a.lower() =="yes" or a.lower() == "y": for x in collected_emails: print "Checking for %s" % x print_emailosint(x) break elif a.lower() =="no" or a.lower() == "n": break elif a.lower() =="s": while True: b = raw_input("Please Enter the EmailId you want to tun OSINT.) [(C)ancel?]: ") if b.lower() =="c": break else: print_emailosint(b) break break else: print("[-] Wrong choice. Please enter Yes or No [Y/N]: \n") #print emailOsint.username_list ''' dns_ip_history = netcraft_domain_history(domain) if len(dns_ip_history.keys()) >= 1: for x in dns_ip_history.keys(): print "%s: %s" % (dns_ip_history[x], x) dict_to_apend['domain_ip_history'] = dns_ip_history #subdomains [to be called before pagelinks so as to avoid repititions.] subdomains(domain) ##print "---> Check_subdomains from wolframalpha" ##find_subdomains_from_wolfram(domain) #domain pagelinks links = pagelinks(domain) if len(links) >= 1: for x in links: print x dict_to_apend['pagelinks'] = links #calling and printing subdomains after pagelinks. subdomains_from_netcraft(domain) print colored(style.BOLD + '---> Finding subdomains: \n' + style.END, 'blue') time.sleep(0.9) if len(subdomain_list) >= 1: for sub in subdomain_list: print sub dict_to_apend['subdomains'] = subdomain_list #wikileaks leaklinks = wikileaks(domain) for tl, lnk in leaklinks.items(): print "%s (%s)" % (lnk, tl) if len(leaklinks.keys()) >= 1: dict_to_apend['wikileaks'] = leaklinks print "For all results, visit: " + 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date' % ( domain) links_brd = boardsearch_forumsearch(domain) for tl, lnk in links_brd.items(): print "%s (%s)" % (lnk, tl) if len(links_brd.keys()) >= 1: dict_to_apend['forum_links'] = links_brd if cfg.zoomeyeuser != "" and cfg.zoomeyepass != "": temp_list = [] zoomeye_results = search_zoomeye(domain) dict_zoomeye_results = json.loads(zoomeye_results) if 'matches' in dict_zoomeye_results.keys(): print len(dict_zoomeye_results['matches']) for x in dict_zoomeye_results['matches']: if x['site'].split('.')[-2] == domain.split('.')[-2]: temp_list.append(x) if 'title' in x.keys(): print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % ( x['ip'], x['site'], x['title'], x['headers'].replace("\n", ""), x['geoinfo']) else: for val in x.keys(): print "%s: %s" % (val, x[val]) if len(temp_list) >= 1: dict_to_apend['zoomeye'] = temp_list if cfg.censysio_id != "" and cfg.censysio_secret != "": print colored( style.BOLD + '\n---> Kicking off Censys Search. This may take a while..\n' + style.END, 'blue') censys_search(domain) if len(censys_list) >= 1: dict_to_apend['censys'] = censys_list for x in censys_list: if x is not None and x != 'None': print x if cfg.shodan_api != "": res_from_shodan = json.loads(shodandomainsearch(domain)) if 'matches' in res_from_shodan.keys(): dict_to_apend['shodan'] = res_from_shodan['matches'] for x in res_from_shodan['matches']: print "IP: %s\nHosts: %s\nDomain: %s\nPort: %s\nData: %s\nLocation: %s\n" % ( x['ip_str'], x['hostnames'], x['domains'], x['port'], x['data'].replace("\n", ""), x['location']) '''
def do_everything(domain): dict_to_apend['targetname'] = domain API_URL = "https://www.censys.io/api/v1" #print cfg.zoomeyeuser #print WhoIs information whoisdata = whoisnew(domain) print whoisdata dict_to_apend['whois'] = whoisdata #print DNS Information dns_records = parse_dns_records(domain) #dict_to_apend['dns_records'] = dns_records > not working #bson.errors.InvalidDocument: Cannot encode object: <DNS IN A rdata: 54.208.84.166> for x in dns_records.keys(): print x if "No" in dns_records[x] and "Found" in dns_records[x]: print "\t%s" % (dns_records[x]) else: for y in dns_records[x]: print "\t%s" % (y) #print type(dns_records[x]) print colored(style.BOLD + '\n---> Finding Paste(s)..\n' + style.END, 'blue') if cfg.google_cse_key != "" and cfg.google_cse_key != "XYZ" and cfg.google_cse_cx != "" and cfg.google_cse_cx != "XYZ": total_results = google_search(domain, 1) if (total_results != 0 and total_results > 10): more_iters = (total_results / 10) if more_iters >= 10: print colored(style.BOLD + '\n---> Too many results, Daily API limit might exceed\n' + style.END, 'red') for x in xrange(1,more_iters + 1): google_search(domain, (x*10)+1) print "\n\n-----------------------------\n" else: print colored(style.BOLD + '\n[-] google_cse_key and google_cse_cx not configured. Skipping paste(s) search.\nPlease refer to http://datasploit.readthedocs.io/en/latest/apiGeneration/.\n' + style.END, 'red') #convert domain to reverse_domain for passing to checkpunkspider() reversed_domain = "" for x in reversed(domain.split(".")): reversed_domain = reversed_domain + "." + x reversed_domain = reversed_domain[1:] res = checkpunkspider(reversed_domain) if 'data' in res.keys() and len(res['data']) >= 1: dict_to_apend['punkspider'] = res['data'] print colored("[+] Few vulnerabilities found at Punkspider", 'green' ) for x in res['data']: print "==> ", x['bugType'] print "Method:", x['verb'].upper() print "URL:\n" + x['vulnerabilityUrl'] print "Param:", x['parameter'] else: print colored("[-] No Vulnerabilities found on PunkSpider", 'red') print colored(style.BOLD + '\n---> Wapplyzing web page of base domain:\n' + style.END, 'blue') wappalyze_results = {} #make proper URL with domain. Check on ssl as well as 80. print "Hitting HTTP:\n", try: targeturl = "http://" + domain list_of_techs = wappalyzeit(targeturl) wappalyze_results['http'] = list_of_techs except: print "[-] HTTP connection was unavailable" wappalyze_results['http'] = [] print "\nHitting HTTPS:\n", try: targeturl = "https://" + domain list_of_techs = wappalyzeit(targeturl) wappalyze_results['https'] = list_of_techs except: print "[-] HTTPS connection was unavailable" wappalyze_results['https'] = [] if len(wappalyze_results.keys()) >= 1: dict_to_apend['wappalyzer'] = wappalyze_results #make Search github code for the given domain. git_results = github_search(domain, 'Code') if git_results is not None: print git_results else: print colored("Sad! Nothing found on github", 'red') #collecting emails for the domain and adding information in master email list. if cfg.emailhunter != "": emails = emailhunter(domain) if len(collected_emails) >= 1: for x in collected_emails: print str(x) dict_to_apend['email_ids'] = collected_emails ''' ##### code for automated osint on enumerated email email_ids while True: a = raw_input(colored("\n\nDo you want to launch osint check for these emails? [(Y)es/(N)o/(S)pecificEmail]: ", 'red')) if a.lower() =="yes" or a.lower() == "y": for x in collected_emails: print "Checking for %s" % x print_emailosint(x) break elif a.lower() =="no" or a.lower() == "n": break elif a.lower() =="s": while True: b = raw_input("Please Enter the EmailId you want to tun OSINT.) [(C)ancel?]: ") if b.lower() =="c": break else: print_emailosint(b) break break else: print("[-] Wrong choice. Please enter Yes or No [Y/N]: \n") #print emailOsint.username_list ''' dns_ip_history = netcraft_domain_history(domain) if len(dns_ip_history.keys()) >= 1: for x in dns_ip_history.keys(): print "%s: %s" % (dns_ip_history[x], x) dict_to_apend['domain_ip_history'] = dns_ip_history #subdomains [to be called before pagelinks so as to avoid repititions.] subdomains(domain) ##print "---> Check_subdomains from wolframalpha" ##find_subdomains_from_wolfram(domain) #domain pagelinks links=pagelinks(domain) if len(links) >= 1: for x in links: print x dict_to_apend['pagelinks'] = links #calling and printing subdomains after pagelinks. subdomains_from_netcraft(domain) print colored(style.BOLD + '---> Finding subdomains: \n' + style.END, 'blue') time.sleep(0.9) if len(subdomain_list) >= 1: for sub in subdomain_list: print sub dict_to_apend['subdomains'] = subdomain_list #wikileaks leaklinks=wikileaks(domain) for tl,lnk in leaklinks.items(): print "%s (%s)" % (lnk, tl) if len(leaklinks.keys()) >= 1: dict_to_apend['wikileaks'] = leaklinks print "For all results, visit: "+ 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date'%(domain) links_brd =boardsearch_forumsearch(domain) for tl,lnk in links_brd.items(): print "%s (%s)" % (lnk, tl) if len(links_brd.keys()) >= 1: dict_to_apend['forum_links'] = links_brd if cfg.zoomeyeuser != "" and cfg.zoomeyepass != "": temp_list =[] zoomeye_results = search_zoomeye(domain) dict_zoomeye_results = json.loads(zoomeye_results) if 'matches' in dict_zoomeye_results.keys(): print len(dict_zoomeye_results['matches']) for x in dict_zoomeye_results['matches']: if x['site'].split('.')[-2] == domain.split('.')[-2]: temp_list.append(x) if 'title' in x.keys() : print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (x['ip'], x['site'], x['title'], x['headers'].replace("\n",""), x['geoinfo']) else: for val in x.keys(): print "%s: %s" % (val, x[val]) if len(temp_list) >= 1: dict_to_apend['zoomeye'] = temp_list if cfg.censysio_id != "" and cfg.censysio_secret != "": print colored(style.BOLD + '\n---> Kicking off Censys Search. This may take a while..\n' + style.END, 'blue') censys_search(domain) if len(censys_list) >= 1: dict_to_apend['censys'] = censys_list for x in censys_list: if x is not None and x != 'None': print x if cfg.shodan_api != "": res_from_shodan = json.loads(shodandomainsearch(domain)) if 'matches' in res_from_shodan.keys(): dict_to_apend['shodan'] = res_from_shodan['matches'] for x in res_from_shodan['matches']: print "IP: %s\nHosts: %s\nDomain: %s\nPort: %s\nData: %s\nLocation: %s\n" % (x['ip_str'], x['hostnames'], x['domains'], x['port'], x['data'].replace("\n",""), x['location']) '''
def do_everything(domain): dict_to_apend['targetname'] = domain API_URL = "https://www.censys.io/api/v1" #print cfg.zoomeyeuser #print WhoIs information whoisdata = whoisnew(domain) print whoisdata dict_to_apend['whois'] = whoisdata print "\n-----------------------------\n" #print DNS Information dns_records = parse_dns_records(domain) #dict_to_apend['dns_records'] = dns_records > not working #bson.errors.InvalidDocument: Cannot encode object: <DNS IN A rdata: 54.208.84.166> for x in dns_records.keys(): print x if "No" in dns_records[x] and "Found" in dns_records[x]: print "\t%s" % (dns_records[x]) else: for y in dns_records[x]: print "\t%s" % (y) #print type(dns_records[x]) print "\n-----------------------------\n" #convert domain to reverse_domain for passing to checkpunkspider() reversed_domain = "" for x in reversed(domain.split(".")): reversed_domain = reversed_domain + "." + x reversed_domain = reversed_domain[1:] res = checkpunkspider(reversed_domain) if 'data' in res.keys() and len(res['data']) >= 1: dict_to_apend['punkspider'] = res['data'] print "[+] Few vulnerabilities found at Punkspider" for x in res['data']: print "==> ", x['bugType'] print "Method:", x['verb'].upper() print "URL:\n" + x['vulnerabilityUrl'] print "Param:", x['parameter'] else: print "[-] No Vulnerabilities found on PunkSpider" print "\n-----------------------------\n" #make proper URL with domain. Check on ssl as well as 80. #print "\t\t\t[+] Wapplyzing " + domain print "Hitting HTTP: ", wappalyze_results = {} try: targeturl = "http://" + domain list_of_techs = wappalyzeit(targeturl) if len(list_of_techs) >= 1: wappalyze_results['http'] = list_of_techs for abc in list_of_techs: print abc, else: pass print "\n" except: print "[-] HTTP connection was unavailable" print "Hitting HTTPS: ", try: targeturl = "https://" + domain list_of_techs = wappalyzeit(targeturl) if len(list_of_techs) >= 1: wappalyze_results['https'] = list_of_techs for abc in list_of_techs: print abc, else: pass print "\n" except: print "[-] HTTP connection was unavailable" if len(wappalyze_results.keys()) >= 1: dict_to_apend['wappalyzer'] = wappalyze_results #make Search github code for the given domain. git_results = github_search(domain, 'Code') if git_results is not None: print git_results else: print "Sad! Nothing found on github" print "\n-----------------------------\n" #collecting emails for the domain and adding information in master email list. if cfg.emailhunter != "": emails = emailhunter(domain) print "\t\t\t[+] Finding Email Ids\n" if len(collected_emails) >= 1: for x in collected_emails: print str(x) + ", ", dict_to_apend['email_ids'] = collected_emails print "\n-----------------------------\n" dns_ip_history = netcraft_domain_history(domain) if len(dns_ip_history.keys()) >= 1: for x in dns_ip_history.keys(): print "%s: %s" % (dns_ip_history[x], x) dict_to_apend['domain_ip_history'] = dns_ip_history print "\n-----------------------------\n" #subdomains [to be called before pagelinks so as to avoid repititions.] print "\t\t\t[+] Finding Subdomains and appending\n" subdomains(domain) ##print "\t\t\t[+] Check_subdomains from wolframalpha" ##find_subdomains_from_wolfram(domain) print "\n-----------------------------\n" #domain pagelinks print "\t\t\t[+] Pagelinks\n" links=pagelinks(domain) if len(links) >= 1: for x in links: print x dict_to_apend['pagelinks'] = links print "\n-----------------------------\n" #calling and printing subdomains after pagelinks. subdomains_from_netcraft(domain) print "\n\t\t\t[+] List of subdomains found\n" if len(subdomain_list) >= 1: for sub in subdomain_list: print sub dict_to_apend['subdomains'] = subdomain_list #wikileaks print "\t\t\t[+] Associated WikiLeaks\n" leaklinks=wikileaks(domain) for tl,lnk in leaklinks.items(): print "%s (%s)" % (lnk, tl) if len(leaklinks.keys()) >= 1: dict_to_apend['wikileaks'] = leaklinks print "For all results, visit: "+ 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date'%(domain) print "\n-----------------------------\n" print "\t\t\t[+] Associated Forum Links\n" links_brd =boardsearch_forumsearch(domain) for tl,lnk in links_brd.items(): print "%s (%s)" % (lnk, tl) if len(links_brd.keys()) >= 1: dict_to_apend['forum_links'] = links_brd print "\n-----------------------------\n" results = check_ssl_htbsecurity(domain) htb_res_dict = {} htb_res_lists = [] if 'ERROR' in results.keys(): print results['ERROR'] elif 'TOKEN' in results.keys(): if 'MULTIPLE_IPS' in results.keys(): print 'Picking up One IP from bunch of IPs returned: %s' % results['MULTIPLE_IPS'][0] results_new = check_ssl_htbsecurity(results['MULTIPLE_IPS'][0]) print "OverAll Rating: %s" % results_new['GRADE'] htb_res_dict['rating'] = results_new['GRADE'] print 'Check https://www.htbridge.com/ssl/ for more information' for x in results_new['VALUE'].keys(): if str("[5]") in str(results_new['VALUE'][x]) or str("[3]") in str(results_new['VALUE'][x]): if x == 'httpHeaders': pass else: print results_new['VALUE'][x] htb_res_lists.append(results_new['VALUE'][x]) else: print "OverAll Rating: %s" % results['GRADE'] htb_res_dict['rating'] = results['GRADE'] for x in results['VALUE'].keys(): if str("[5]") in str(results['VALUE'][x]) or str("[3]") in str(results['VALUE'][x]): if x == 'httpHeaders': pass else: print results['VALUE'][x] htb_res_lists.append(results['VALUE'][x]) htb_res_dict['issues'] = htb_res_lists dict_to_apend['ssl_scan'] = htb_res_dict print "\n-----------------------------\n" #checks results from zoomeye #filters need to be applied #zoomeye_results = search_zoomeye(domain) #dict_zoomeye_results = json.loads(zoomeye_results) #if 'matches' in dict_zoomeye_results.keys(): # for x in dict_zoomeye_results['matches']: # if x['site'].split('.')[-2] == domain.split('.')[-2]: # print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (x['ip'], x['site'], x['title'], x['headers'].replace("\n",""), x['geoinfo']) #print "\n-----------------------------\n" if cfg.zoomeyeuser != "" and cfg.zoomeyepass != "": temp_list =[] zoomeye_results = search_zoomeye(domain) dict_zoomeye_results = json.loads(zoomeye_results) if 'matches' in dict_zoomeye_results.keys(): print len(dict_zoomeye_results['matches']) for x in dict_zoomeye_results['matches']: if x['site'].split('.')[-2] == domain.split('.')[-2]: temp_list.append(x) if 'title' in x.keys() : print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (x['ip'], x['site'], x['title'], x['headers'].replace("\n",""), x['geoinfo']) else: for val in x.keys(): print "%s: %s" % (val, x[val]) if len(temp_list) >= 1: dict_to_apend['zoomeye'] = temp_list print "\n-----------------------------\n" if cfg.censysio_id != "" and cfg.censysio_secret != "": print "[+]\t Kicking off Censys Search. This may take a while.." censys_search(domain) if len(censys_list) >= 1: dict_to_apend['censys'] = censys_list for x in censys_list: print x print "\n-----------------------------\n" #Code for shodan Ip search. now we are doing Hostname search. #converts domain to IP. Prints a statement for the same. #ip_addr = socket.gethostbyname(domain) #checks for information at shodan, and comes back with whatever available. ## need to apply filter here (lot of noise coming in) #res_from_shodan = json.loads(shodansearch(ip_addr)) #print res_from_shodan #for iterate_shodan_list in res_from_shodan['data']: # print "ISP: %s \n Hosts: %s \n IP: %s \n Data: %s\n" % (iterate_shodan_list['isp'], iterate_shodan_list['hostnames'], iterate_shodan_list['ip_str'], iterate_shodan_list['data'].strip("\n")) #print "\n-----------------------------\n" if cfg.shodan_api != "": res_from_shodan = json.loads(shodandomainsearch(domain)) if 'matches' in res_from_shodan.keys(): dict_to_apend['shodan'] = res_from_shodan['matches'] for x in res_from_shodan['matches']: print "IP: %s\nHosts: %s\nDomain: %s\nPort: %s\nData: %s\nLocation: %s\n" % (x['ip_str'], x['hostnames'], x['domains'], x['port'], x['data'].replace("\n",""), x['location']) print "-----------------------------\n" #insert data into mongodb instance result = db.domaindata.insert(dict_to_apend, check_keys=False) print 'done'
def do_everything(domain, output, active): dict_to_apend['targetname'] = domain API_URL = "https://www.censys.io/api/v1" #print cfg.zoomeyeuser html = "" scroll = [] #print WhoIs information whoisdata = whoisnew(domain) print whoisdata dict_to_apend['whois'] = whoisdata html, scroll = make_html(html, "WHOIS Information", "whois", "<pre>%s</pre>" % str(whoisdata), scroll) #print DNS Information dns_records = parse_dns_records(domain) #dict_to_apend['dns_records'] = dns_records > not working #bson.errors.InvalidDocument: Cannot encode object: <DNS IN A rdata: 54.208.84.166> lhtml = "<ul class='collection'>" for x in dns_records.keys(): print x lhtml += "<li class='collection-item'>%s<ul class='collection'>" % x if "No" in dns_records[x] and "Found" in dns_records[x]: print "\t%s" % (dns_records[x]) lhtml += "<li class='collection-item'>%s</li>" % dns_records[x] else: for y in dns_records[x]: if x == "TXT Records": lhtml += "<li class='collection-item'>%s</li>" % ", ".join(y.strings) print "\t%s" % ", ".join(y.strings) else: lhtml += "<li class='collection-item'>%s</li>" % str(y) print "\t%s" % (y) #print type(dns_records[x]) lhtml += "</ul></li>" lhtml += "</ul>" html, scroll = make_html(html, "DNS Records", "dns-records", lhtml, scroll) print colored(style.BOLD + '\n---> Finding Paste(s)..\n' + style.END, 'blue') '''print colored(style.BOLD + '\n---> Finding Paste(s)..\n' + style.END, 'blue') if cfg.google_cse_key and cfg.google_cse_key != "XYZ" and cfg.google_cse_cx and cfg.google_cse_cx != "XYZ": total_results = google_search(domain, 1) if (total_results != 0 and total_results > 10): more_iters = (total_results / 10) if more_iters >= 10: print colored(style.BOLD + '\n---> Too many results, Daily API limit might exceed\n' + style.END, 'red') for x in xrange(1,more_iters + 1): google_search(domain, (x*10)+1) print "\n\n-----------------------------------------n" else: print colored(style.BOLD + '\n[-] google_cse_key and google_cse_cx not configured. Skipping paste(s) search.\nPlease refer to http://datasploit.readthedocs.io/en/latest/apiGeneration/.\n' + style.END, 'red') ''' if cfg.google_cse_key != "" and cfg.google_cse_key != "XYZ" and cfg.google_cse_cx != "" and cfg.google_cse_cx != "XYZ": total_results, results = google_search(domain, 1) if not total_results == 0: lhtml = """<div style='overflow-x: scroll;'><table class='bordered highlight'> <thead> <tr> <th>Title</th> <th>URL</th> <th>Snippet</th> </tr> </thead> <tbody> """ for x in results['items']: lhtml += "<tr>" lhtml += "<td>%s</td><td>%s</td><td>%s</td>" % (cgi.escape(x['title']), cgi.escape(x['link']), cgi.escape(x['snippet'])) lhtml += "</tr>" if (total_results != 0 and total_results > 10): more_iters = (total_results / 10) if more_iters >= 10: print colored(style.BOLD + '\n---> Too many results, Daily API limit might exceed\n' + style.END, 'red') for x in xrange(1,more_iters + 1): total_results, results = google_search(domain, (x*10)+1) if results: for y in results['items']: lhtml += "<tr>" lhtml += "<td>%s</td><td>%s</td><td>%s</td>" % (cgi.escape(y['title']), cgi.escape(y['link']), cgi.escape(y['snippet'])) lhtml += "</tr>" print "\n\n-----------------------------\n" lhtml += "</tbody></table></div>" html, scroll = make_html(html, "Google Search Pastes", "google-search-pastes", lhtml, scroll) else: print colored(style.BOLD + '\n[-] google_cse_key and google_cse_cx not configured. Skipping paste(s) search.\nPlease refer to http://datasploit.readthedocs.io/en/latest/apiGeneration/.\n' + style.END, 'red') #convert domain to reverse_domain for passing to checkpunkspider() reversed_domain = "" for x in reversed(domain.split(".")): reversed_domain = reversed_domain + "." + x reversed_domain = reversed_domain[1:] res = checkpunkspider(reversed_domain) if 'data' in res.keys() and len(res['data']) >= 1: dict_to_apend['punkspider'] = res['data'] print colored("[+] Few vulnerabilities found at Punkspider", 'green' ) lhtml = "<div style='overflow-x: scroll;'><table class='bordered highlight'><thead>" lhtml += "<tr>" lhtml += "<th>Bug Type</th>" lhtml += "<th>Method</th>" lhtml += "<th>URL</th>" lhtml += "<th>Parameter</th>" lhtml += "</tr></thead><tbody>" for x in res['data']: lhtml += "<tr>" print "==> ", x['bugType'] lhtml += "<td>%s</td>" % x['bugType'] print "Method:", x['verb'].upper() lhtml += "<td>%s</td>" % x['verb'].upper() print "URL:\n" + x['vulnerabilityUrl'] lhtml += "<td>%s</td>" % x['vulnerabilityUrl'] print "Param:", x['parameter'] lhtml += "<td>%s</td>" % x['parameter'] lhtml += "</tr>" lhtml += "</tbody></table></div>" html, scroll = make_html(html, "Punkspider Vulnerabilities", "punkspider-vulns", lhtml, scroll) else: print colored("[-] No Vulnerabilities found on PunkSpider", 'red') print colored(style.BOLD + '\n---> Wapplyzing web page of base domain:\n' + style.END, 'blue') wappalyze_results = {} #make proper URL with domain. Check on ssl as well as 80. print "Hitting HTTP:\n", lhtml = "" try: targeturl = "http://" + domain list_of_techs = wappalyzeit(targeturl) wappalyze_results['http'] = list_of_techs except: print "[-] HTTP connection was unavailable" wappalyze_results['http'] = [] if wappalyze_results['http']: lhtml += "<ul class='collection'><li class='collection-item'>HTTP<ul class='collection'>" for i in wappalyze_results['http']: lhtml += "<li class='collection-item'>%s</li>" % i lhtml += "</ul></li></ul>" print "\nHitting HTTPS:\n", try: targeturl = "https://" + domain list_of_techs = wappalyzeit(targeturl) wappalyze_results['https'] = list_of_techs except: print "[-] HTTPS connection was unavailable" wappalyze_results['https'] = [] if wappalyze_results['https']: lhtml += "<ul class='collection'><li class='collection-item'>HTTPS<ul class='collection'>" for i in wappalyze_results['https']: lhtml += "<li class='collection-item'>%s</li>" % i lhtml += "</ul></li></ul>" html, scroll = make_html(html, "WappAlyzer Report", "wappalyzer-report", lhtml, scroll) if len(wappalyze_results.keys()) >= 1: dict_to_apend['wappalyzer'] = wappalyze_results #make Search github code for the given domain. git_results = github_search(domain, 'Code') if git_results is not None: html, scroll = make_html(html, "Github Report", "github-report", git_results, scroll) else: print colored("Sad! Nothing found on github", 'red') #collecting emails for the domain and adding information in master email list. if cfg.emailhunter != "": emails = emailhunter(domain) if len(collected_emails) >= 1: lhtml = "<ul class='collection'>" for x in collected_emails: lhtml += "<li class='collection-item'>%s</li>" % str(x) print str(x) lhtml += "</ul>" html, scroll = make_html(html, "Email Harvesting", "email-harvesting", lhtml, scroll) dict_to_apend['email_ids'] = collected_emails ''' ##### code for automated osint on enumerated email email_ids while True: a = raw_input(colored("\n\nDo you want to launch osint check for these emails? [(Y)es/(N)o/(S)pecificEmail]: ", 'red')) if a.lower() =="yes" or a.lower() == "y": for x in collected_emails: print "Checking for %s" % x print_emailosint(x) break elif a.lower() =="no" or a.lower() == "n": break elif a.lower() =="s": while True: b = raw_input("Please Enter the EmailId you want to tun OSINT.) [(C)ancel?]: ") if b.lower() =="c": break else: print_emailosint(b) break break else: print("[-] Wrong choice. Please enter Yes or No [Y/N]: \n") #print emailOsint.username_list ''' dns_ip_history = netcraft_domain_history(domain) if len(dns_ip_history.keys()) >= 1: lhtml = "<div style='overflow-x: scroll;'><table class='bordered highlight'><thead>" lhtml += "<tr><th>IP</th><th>Description</th></tr></thead><tbody>" for x in dns_ip_history.keys(): lhtml += "<tr><td>%s</td><td>%s</td></tr>" % (dns_ip_history[x], x) print "%s: %s" % (dns_ip_history[x], x) lhtml += "</tbody></table></div>" dict_to_apend['domain_ip_history'] = dns_ip_history html, scroll = make_html(html, "DNS IP History", "dns-ip-history", lhtml, scroll) #subdomains [to be called before pagelinks so as to avoid repititions.] subdomains(domain) ##print "---> Check_subdomains from wolframalpha" ##find_subdomains_from_wolfram(domain) #domain pagelinks links=pagelinks(domain) links = list(set(links)) links.sort() if len(links) >= 1: lhtml = "<ul class='collection'>" for x in links: print x lhtml += "<li class='collection-item'>%s</li>" % x lhtml += "</ul>" dict_to_apend['pagelinks'] = links html, scroll = make_html(html, "Page Links", "page-links", lhtml, scroll) #calling and printing subdomains after pagelinks. subdomains_from_netcraft(domain) print colored(style.BOLD + '---> Finding subdomains: \n' + style.END, 'blue') time.sleep(0.9) if len(subdomain_list) >= 1: lhtml = "<ul class='collection'>" for sub in subdomain_list: lhtml += "<li class='collection-item'>%s</li>" % sub print sub lhtml += "</ul>" dict_to_apend['subdomains'] = subdomain_list html, scroll = make_html(html, "Subdomains", "subdomains", lhtml, scroll) #wikileaks leaklinks=wikileaks(domain) lhtml = "<div style='overflow-x: scroll;'><table class='bordered highlight'><thead><tr><th>URL</th><th>Description</th></tr></thead><tbody>" for tl,lnk in leaklinks.items(): lhtml += "<tr><td>%s</td><td>%s</td></tr>" % (lnk, tl) print "%s (%s)" % (lnk, tl) lhtml += "</tbody></table></div>" html, scroll = make_html(html, "WikiLeaks", "wikileaks", lhtml, scroll) if len(leaklinks.keys()) >= 1: dict_to_apend['wikileaks'] = leaklinks print "For all results, visit: "+ 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date'%(domain) links_brd =boardsearch_forumsearch(domain) if links_brd: lhtml = "<ul class='collection'>" for tl,lnk in links_brd.items(): lhtml += "<li class='collection-item'>%s (%s)</li>" % (lnk, tl) print "%s (%s)" % (lnk, tl) lhtml += "</ul>" html, scroll = make_html(html, "Forum Search", "forum-search", lhtml, scroll) if len(links_brd.keys()) >= 1: dict_to_apend['forum_links'] = links_brd if cfg.zoomeyeuser != "" and cfg.zoomeyepass != "": temp_list =[] zoomeye_results = search_zoomeye(domain) dict_zoomeye_results = json.loads(zoomeye_results) if 'matches' in dict_zoomeye_results.keys(): lhtml = "" print len(dict_zoomeye_results['matches']) for x in dict_zoomeye_results['matches']: if x['site'].split('.')[-2] == domain.split('.')[-2]: temp_list.append(x) if 'title' in x.keys(): zoom_data = "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (x['ip'], x['site'], x['title'], x['headers'].replace("\n",""), x['geoinfo']) lhtml += zoom_data print zoom_data else: for val in x.keys(): zoom_data = "%s: %s" % (val, x[val]) lhtml += zoom_data print zoom_data html, scroll = make_html(html, "Zoomeye Search", "zoomeye-search", "<pre>%s</pre>" % lhtml, scroll) if len(temp_list) >= 1: dict_to_apend['zoomeye'] = temp_list if cfg.censysio_id != "" and cfg.censysio_secret != "": print colored(style.BOLD + '\n---> Kicking off Censys Search. This may take a while..\n' + style.END, 'blue') censys_search(domain) if len(censys_list) >= 1: dict_to_apend['censys'] = censys_list lhtml = "<ul class='collection'>" for x in censys_list: if not str(x) == 'None': lhtml += "<li class='collection-item'><pre>%s</pre></li>" % cgi.escape(str(x)) print x lhtml += "</ul>" html, scroll = make_html(html, "Censys Search", "censys-search", lhtml, scroll) if cfg.shodan_api != "": res_from_shodan = json.loads(shodandomainsearch(domain)) if 'matches' in res_from_shodan.keys(): dict_to_apend['shodan'] = res_from_shodan['matches'] lhtml = "<ul class='collection'>" for x in res_from_shodan['matches']: shod_data = "IP: %s\nHosts: %s\nDomain: %s\nPort: %s\nData: %s\nLocation: %s\n" % (x['ip_str'], x['hostnames'], x['domains'], x['port'], x['data'].replace("\n",""), x['location']) lhtml += "<li class='collection-item'><pre>%s</pre></li>" % shod_data print shod_data lhtml += "</ul>" html, scroll = make_html(html, "Shodan Search", "shodan-search", lhtml, scroll) if output: scroll_html = "" for i in scroll: scroll_html += "<li><a href='#%s'>%s</a></li>" % i fh = open("base.html", "r") main_html = fh.read() fh.close() if not os.path.exists("reports/%s" % domain): os.mkdir("reports/%s" % domain) now_time = datetime.now().strftime("%Y-%m-%d_%H-%M-%S") fh = open("reports/%s/%s_%s.html" % (domain, domain, now_time), "w") main_html = main_html.replace("{DOMAIN_NAME}", domain) main_html = main_html.replace("{MAIN_HTML}", html) main_html = main_html.replace("{scrollROLL_SECTION}", scroll_html) fh.write(main_html) fh.close() fh = open("reports/%s/%s_%s.json" % (domain, domain, now_time), "w") fh.write(str(dict_to_apend)) fh.close() fh = open("reports/%s/%s_%s.subdomains.txt" % (domain, domain, now_time), "w") for x in subdomain_list: fh.write("\n%s" % x) fh.close() fh = open("reports/%s/%s_%s.emails.txt" % (domain, domain, now_time), "w") for x in collected_emails: fh.write("\n%s" % x) fh.close() print colored(style.BOLD + '---> Saving work... \n' + style.END, 'green') print "[+] HTML Report saved to: reports/%s/%s_%s.html" % (domain, domain, now_time) print "[+] JSON Report saved to: reports/%s/%s_%s.json" % (domain, domain, now_time) print "[+] List of Emails saved to: reports/%s/%s_%s.emails.txt" % (domain, domain, now_time) print "[+] List of Emails saved to: reports/%s/%s_%s.subdomains.txt" % (domain, domain, now_time) print "\n" if active == 'true': print colored(style.BOLD + '---> Running Active Scans: \n' + style.END, 'green') run_active(domain, subdomain_list, collected_emails) '''
def do_everything(domain): dict_to_apend['targetname'] = domain API_URL = "https://www.censys.io/api/v1" #print cfg.zoomeyeuser #print WhoIs information whoisdata = whoisnew(domain) print whoisdata dict_to_apend['whois'] = whoisdata #print DNS Information dns_records = parse_dns_records(domain) #dict_to_apend['dns_records'] = dns_records > not working #bson.errors.InvalidDocument: Cannot encode object: <DNS IN A rdata: 54.208.84.166> for x in dns_records.keys(): print x if "No" in dns_records[x] and "Found" in dns_records[x]: print "\t%s" % (dns_records[x]) else: for y in dns_records[x]: print "\t%s" % (y) #print type(dns_records[x]) #convert domain to reverse_domain for passing to checkpunkspider() reversed_domain = "" for x in reversed(domain.split(".")): reversed_domain = reversed_domain + "." + x reversed_domain = reversed_domain[1:] res = checkpunkspider(reversed_domain) if 'data' in res.keys() and len(res['data']) >= 1: dict_to_apend['punkspider'] = res['data'] print colored("[+] Few vulnerabilities found at Punkspider", 'green' ) for x in res['data']: print "==> ", x['bugType'] print "Method:", x['verb'].upper() print "URL:\n" + x['vulnerabilityUrl'] print "Param:", x['parameter'] else: print colored("[-] No Vulnerabilities found on PunkSpider", 'red') #make proper URL with domain. Check on ssl as well as 80. #print "---> Wapplyzing " + domain print colored(style.BOLD + '\n---> Wappalyzing web pages\n' + style.END, 'blue') time.sleep(0.3) print colored("->Trying Wapalyzer on HTTP: ", 'blue') wappalyze_results = {} try: targeturl = "http://" + domain list_of_techs = wappalyzeit(targeturl) if len(list_of_techs) >= 1: wappalyze_results['http'] = list_of_techs for abc in list_of_techs: print abc, else: pass print "\n" except: print colored("[-] HTTP connection was unavailable", 'red') print colored("->Trying Wapalyzer on HTTPS: ", 'blue') try: targeturl = "https://" + domain list_of_techs = wappalyzeit(targeturl) if len(list_of_techs) >= 1: wappalyze_results['https'] = list_of_techs for abc in list_of_techs: print abc, else: pass print "\n" except: print colored("[-] HTTP connection was unavailable", 'red') if len(wappalyze_results.keys()) >= 1: dict_to_apend['wappalyzer'] = wappalyze_results #make Search github code for the given domain. git_results = github_search(domain, 'Code') if git_results is not None: print git_results else: print colored("Sad! Nothing found on github", 'red') #collecting emails for the domain and adding information in master email list. if cfg.emailhunter != "": emails = emailhunter(domain) if len(collected_emails) >= 1: for x in collected_emails: print str(x) + ", ", dict_to_apend['email_ids'] = collected_emails while True: a = raw_input(colored("\n\nDo you want to launch osint check for these emails? [(Y)es/(N)o/(S)pecificEmail]: ", 'red')) if a.lower() =="yes" or a.lower() == "y": for x in collected_emails: print "Checking for %s" % x print_emailosint(x) break elif a.lower() =="no" or a.lower() == "n": break elif a.lower() =="s": while True: b = raw_input("Please Enter the EmailId you want to tun OSINT.) [(C)ancel?]: ") if b.lower() =="c": break else: print_emailosint(b) break break else: print("[-] Wrong choice. Please enter Yes or No [Y/N]: \n") #print emailOsint.username_list dns_ip_history = netcraft_domain_history(domain) if len(dns_ip_history.keys()) >= 1: for x in dns_ip_history.keys(): print "%s: %s" % (dns_ip_history[x], x) dict_to_apend['domain_ip_history'] = dns_ip_history #subdomains [to be called before pagelinks so as to avoid repititions.] subdomains(domain) ##print "---> Check_subdomains from wolframalpha" ##find_subdomains_from_wolfram(domain) #domain pagelinks links=pagelinks(domain) if len(links) >= 1: for x in links: print x dict_to_apend['pagelinks'] = links #calling and printing subdomains after pagelinks. subdomains_from_netcraft(domain) print colored(style.BOLD + '---> Finding subdomains: \n' + style.END, 'blue') time.sleep(0.9) if len(subdomain_list) >= 1: for sub in subdomain_list: print sub dict_to_apend['subdomains'] = subdomain_list #wikileaks leaklinks=wikileaks(domain) for tl,lnk in leaklinks.items(): print "%s (%s)" % (lnk, tl) if len(leaklinks.keys()) >= 1: dict_to_apend['wikileaks'] = leaklinks print "For all results, visit: "+ 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date'%(domain) links_brd =boardsearch_forumsearch(domain) for tl,lnk in links_brd.items(): print "%s (%s)" % (lnk, tl) if len(links_brd.keys()) >= 1: dict_to_apend['forum_links'] = links_brd print colored(style.BOLD + '\n---> Performing passive SSL Scan\n' + style.END, 'blue') results = check_ssl_htbsecurity(domain) htb_res_dict = {} htb_res_lists = [] if 'ERROR' in results.keys(): print results['ERROR'] elif 'TOKEN' in results.keys(): if 'MULTIPLE_IPS' in results.keys(): print colored('Picking up One IP from bunch of IPs returned: %s', 'green') % results['MULTIPLE_IPS'][0] results_new = check_ssl_htbsecurity(results['MULTIPLE_IPS'][0]) print "OverAll Rating: %s" % results_new['GRADE'] htb_res_dict['rating'] = results_new['GRADE'] print 'Check https://www.htbridge.com/ssl/ for more information' for x in results_new['VALUE'].keys(): if str("[5]") in str(results_new['VALUE'][x]) or str("[3]") in str(results_new['VALUE'][x]): if x == 'httpHeaders': pass else: print results_new['VALUE'][x] htb_res_lists.append(results_new['VALUE'][x]) else: print "OverAll Rating: %s" % results['GRADE'] htb_res_dict['rating'] = results['GRADE'] for x in results['VALUE'].keys(): if str("[5]") in str(results['VALUE'][x]) or str("[3]") in str(results['VALUE'][x]): if x == 'httpHeaders': pass else: print results['VALUE'][x] htb_res_lists.append(results['VALUE'][x]) htb_res_dict['issues'] = htb_res_lists dict_to_apend['ssl_scan'] = htb_res_dict #checks results from zoomeye #filters need to be applied #zoomeye_results = search_zoomeye(domain) #dict_zoomeye_results = json.loads(zoomeye_results) #if 'matches' in dict_zoomeye_results.keys(): # for x in dict_zoomeye_results['matches']: # if x['site'].split('.')[-2] == domain.split('.')[-2]: # print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (x['ip'], x['site'], x['title'], x['headers'].replace("\n",""), x['geoinfo']) #print "\n-----------------------------\n" if cfg.zoomeyeuser != "" and cfg.zoomeyepass != "": temp_list =[] zoomeye_results = search_zoomeye(domain) dict_zoomeye_results = json.loads(zoomeye_results) if 'matches' in dict_zoomeye_results.keys(): print len(dict_zoomeye_results['matches']) for x in dict_zoomeye_results['matches']: if x['site'].split('.')[-2] == domain.split('.')[-2]: temp_list.append(x) if 'title' in x.keys() : print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (x['ip'], x['site'], x['title'], x['headers'].replace("\n",""), x['geoinfo']) else: for val in x.keys(): print "%s: %s" % (val, x[val]) if len(temp_list) >= 1: dict_to_apend['zoomeye'] = temp_list if cfg.censysio_id != "" and cfg.censysio_secret != "": print colored(style.BOLD + '\n---> Kicking off Censys Search. This may take a while..\n' + style.END, 'blue') censys_search(domain) if len(censys_list) >= 1: dict_to_apend['censys'] = censys_list for x in censys_list: if x is not None and x != 'None': print x #Code for shodan Ip search. now we are doing Hostname search. #converts domain to IP. Prints a statement for the same. #ip_addr = socket.gethostbyname(domain) #checks for information at shodan, and comes back with whatever available. ## need to apply filter here (lot of noise coming in) #res_from_shodan = json.loads(shodansearch(ip_addr)) #print res_from_shodan #for iterate_shodan_list in res_from_shodan['data']: # print "ISP: %s \n Hosts: %s \n IP: %s \n Data: %s\n" % (iterate_shodan_list['isp'], iterate_shodan_list['hostnames'], iterate_shodan_list['ip_str'], iterate_shodan_list['data'].strip("\n")) #print "\n-----------------------------\n" if cfg.shodan_api != "": res_from_shodan = json.loads(shodandomainsearch(domain)) if 'matches' in res_from_shodan.keys(): dict_to_apend['shodan'] = res_from_shodan['matches'] for x in res_from_shodan['matches']: print "IP: %s\nHosts: %s\nDomain: %s\nPort: %s\nData: %s\nLocation: %s\n" % (x['ip_str'], x['hostnames'], x['domains'], x['port'], x['data'].replace("\n",""), x['location']) #insert data into mongodb instance try: result = db.domaindata.insert(dict_to_apend, check_keys=False) print 'output saved to MongoDb' except: print "More data than I can handle, hence not saved in MongoDb. Apologies."