Esempio n. 1
0
def domain_page_links(domain):
    try:
        pagelinks_records = pagelinks(domain)
        pagelinks_records = set(pagelinks_records)
        for x in pagelinks_records:
            print x
            save_data = domain_pagelinks_db(domain_pagelink=(x))
            save_data.save()
    except Exception as error:
        print error
Esempio n. 2
0
def main():
    domain = sys.argv[1]
    #subdomains [to be called before pagelinks so as to avoid repititions.]
    print "\t\t\t[+] Finding Subdomains and appending\n"
    subdomains(domain)
    ##print "\t\t\t[+] Check_subdomains from wolframalpha"
    ##find_subdomains_from_wolfram(domain)
    pagelinks_list = pagelinks(domain)

    subdomains_from_netcraft(domain)

    #printing all subdomains
    print "\n\t\t\t[+] List of subdomains found\n"
    for sub in subdomain_list:
        print sub
Esempio n. 3
0
def main():
	domain = sys.argv[1]
	#subdomains [to be called before pagelinks so as to avoid repititions.]
	print "\t\t\t[+] Finding Subdomains and appending\n"
	subdomains(domain)
	##print "\t\t\t[+] Check_subdomains from wolframalpha"
	##find_subdomains_from_wolfram(domain)
	pagelinks_list = pagelinks(domain)

	subdomains_from_netcraft(domain)

	#printing all subdomains
	print "\n\t\t\t[+] List of subdomains found\n"
	for sub in subdomain_list:
		print sub
Esempio n. 4
0
def main():
    domain = sys.argv[1]

    API_URL = "https://www.censys.io/api/v1"
    #print cfg.zoomeyeuser

    #print WhoIs information
    print whoisnew(domain)
    print "\n-----------------------------\n"

    #print DNS Information
    dns_records = parse_dns_records(domain)
    for x in dns_records.keys():
        print x
        if "No" in dns_records[x] and "Found" in dns_records[x]:
            print "\t%s" % (dns_records[x])
        else:
            for y in dns_records[x]:
                print "\t%s" % (y)
            #print type(dns_records[x])
    print "\n-----------------------------\n"

    #convert domain to reverse_domain for passing to checkpunkspider()
    reversed_domain = ""
    for x in reversed(domain.split(".")):
        reversed_domain = reversed_domain + "." + x
    reversed_domain = reversed_domain[1:]
    checkpunkspider(reversed_domain)
    print "\n-----------------------------\n"

    #make proper URL with domain. Check on ssl as well as 80.
    print "\t\t\t[+] Wapplyzing " + domain
    print "Hitting HTTP:\n",
    try:
        targeturl = "http://" + domain
        wappalyzeit(targeturl)
    except:
        print "[-] HTTP connection was unavailable"
    print "Hitting HTTPS:\n",
    try:
        targeturl = "https://" + domain
        wappalyzeit(targeturl)
    except:
        print "[-] HTTPS connection was unavailable"
    print "\n-----------------------------\n"

    #make Search github code for the given domain.
    print github_search(domain, 'Code')
    print "\n-----------------------------\n"

    #collecting emails for the domain and adding information in master email list.
    emailhunter(domain)
    print "\t\t\t[+] Finding Email Ids\n"
    for x in collected_emails:
        print str(x) + ", ",
    print "\n-----------------------------\n"

    dns_history = netcraft_domain_history(domain)
    for x in dns_history.keys():
        print "%s: %s" % (dns_history[x], x)
    print "\n-----------------------------\n"

    #subdomains [to be called before pagelinks so as to avoid repititions.]
    print "\t\t\t[+] Finding Subdomains and appending\n"
    subdomains(domain)
    ##print "\t\t\t[+] Check_subdomains from wolframalpha"
    ##find_subdomains_from_wolfram(domain)
    print "\n-----------------------------\n"

    #domain pagelinks
    print "\t\t\t[+] Pagelinks\n"
    links = pagelinks(domain)
    for x in links:
        print x
    print "\n-----------------------------\n"

    #calling and printing subdomains after pagelinks.

    subdomains_from_netcraft(domain)
    print "\n\t\t\t[+] List of subdomains found\n"
    for sub in subdomain_list:
        print sub

    #wikileaks
    print "\t\t\t[+] Associated WikiLeaks\n"
    leaklinks = wikileaks(domain)
    for tl, lnk in leaklinks.items():
        print "%s (%s)" % (lnk, tl)
    print "For all results, visit: " + 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date' % (
        domain)
    print "\n-----------------------------\n"

    print "\t\t\t[+] Associated Forum Links\n"
    links = boardsearch_forumsearch(domain)
    for tl, lnk in links.items():
        print "%s (%s)" % (lnk, tl)
    print "\n-----------------------------\n"

    results = check_ssl_htbsecurity(domain)
    if 'ERROR' in results.keys():
        print results['ERROR']
    elif 'TOKEN' in results.keys():
        print 'Picking up One IP from bunch of IPs returned: %s' % results[
            'MULTIPLE_IPS'][0]
        results_new = check_ssl_htbsecurity(results['MULTIPLE_IPS'][0])
        print "OverAll Rating: %s" % results_new['GRADE']
        print 'Check https://www.htbridge.com/ssl/ for more information'
        for x in results_new['VALUE'].keys():
            if str("[5]") in str(results_new['VALUE'][x]) or str("[3]") in str(
                    results_new['VALUE'][x]):
                if x == 'httpHeaders':
                    pass
                else:
                    print results_new['VALUE'][x]
    else:
        print "OverAll Rating: %s" % results['GRADE']
        for x in results['VALUE'].keys():
            if str("[5]") in str(results['VALUE'][x]) or str("[3]") in str(
                    results['VALUE'][x]):
                if x == 'httpHeaders':
                    pass
                else:
                    print results['VALUE'][x]
    print "\n-----------------------------\n"

    #checks results from zoomeye
    #filters need to be applied
    if cfg.zoomeyeuser != "" and cfg.zoomeyepass != "":
        zoomeye_results = search_zoomeye(domain)
        dict_zoomeye_results = json.loads(zoomeye_results)
        if 'matches' in dict_zoomeye_results.keys():
            for x in dict_zoomeye_results['matches']:
                if x['site'].split('.')[-2] == domain.split('.')[-2]:
                    print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (
                        x['ip'], x['site'], x['title'], x['headers'].replace(
                            "\n", ""), x['geoinfo'])
        print "\n-----------------------------\n"

    print "[+]\t Kicking off Censys Search. This may take a while.."
    censys_search(domain)
    for x in censys_list:
        print x
    print "\n-----------------------------\n"

    #checks for host ip and other details in shodan.
    '''
	#Code for shodan Ip search. now we are doing Hostname search.

	#converts domain to IP. Prints a statement for the same.
	ip_addr = socket.gethostbyname(domain)

	#checks for information at shodan, and comes back with whatever available.
	## need to apply filter here (lot of noise coming in)
	res_from_shodan = json.loads(shodansearch(ip_addr))
	#print res_from_shodan
	for iterate_shodan_list in res_from_shodan['data']:
		print "ISP: %s \n Hosts: %s \n IP: %s \n Data: %s\n" % (iterate_shodan_list['isp'], iterate_shodan_list['hostnames'], iterate_shodan_list['ip_str'], iterate_shodan_list['data'].strip("\n"))
	print "\n-----------------------------\n"
	'''

    res_from_shodan = json.loads(shodandomainsearch(domain))
    if 'matches' in res_from_shodan.keys():
        for x in res_from_shodan['matches']:
            print "IP: %s\nHosts: %s\nDomain: %s\nPort: %s\nData: %s\nLocation: %s\n" % (
                x['ip_str'], x['hostnames'], x['domains'], x['port'],
                x['data'].replace("\n", ""), x['location'])
    print "-----------------------------\n"
Esempio n. 5
0
def main(): 
	domain = sys.argv[1]
	
	API_URL = "https://www.censys.io/api/v1"
	#print cfg.zoomeyeuser

	
	#print WhoIs information
	print whoisnew(domain)
	print "\n-----------------------------\n"


	#print DNS Information
	dns_records = parse_dns_records(domain)
	for x in dns_records.keys():
		print x
		if "No" in dns_records[x] and "Found" in dns_records[x]:
			print "\t%s" % (dns_records[x])
		else:
			for y in dns_records[x]:
				print "\t%s" % (y)
			#print type(dns_records[x])
	print "\n-----------------------------\n"

	#convert domain to reverse_domain for passing to checkpunkspider()
	reversed_domain = ""
	for x in reversed(domain.split(".")):
		reversed_domain = reversed_domain + "." + x
	reversed_domain = reversed_domain[1:]
	checkpunkspider(reversed_domain)
	print "\n-----------------------------\n"

	#make proper URL with domain. Check on ssl as well as 80.
	print "\t\t\t[+] Wapplyzing " + domain 
	print "Hitting HTTP:\n",
	try:
		targeturl = "http://" + domain
		wappalyzeit(targeturl)
	except:
		print "[-] HTTP connection was unavailable"
	print "Hitting HTTPS:\n",
	try:
		targeturl = "https://" + domain
		wappalyzeit(targeturl)
	except:
		print "[-] HTTPS connection was unavailable"
	print "\n-----------------------------\n"

	#make Search github code for the given domain.
	print github_search(domain, 'Code')
	print "\n-----------------------------\n"
	
	#collecting emails for the domain and adding information in master email list. 
	emailhunter(domain)
	print "\t\t\t[+] Finding Email Ids\n"
	for x in collected_emails:
		print str(x) + ", ",
	print "\n-----------------------------\n"


	dns_history = netcraft_domain_history(domain)
	for x in dns_history.keys():
		print "%s: %s" % (dns_history[x], x)
	print "\n-----------------------------\n"	

	
	#subdomains [to be called before pagelinks so as to avoid repititions.]
	print "\t\t\t[+] Finding Subdomains and appending\n"
	subdomains(domain)
	##print "\t\t\t[+] Check_subdomains from wolframalpha"
	##find_subdomains_from_wolfram(domain)
	print "\n-----------------------------\n"
	

	#domain pagelinks
	print "\t\t\t[+] Pagelinks\n"
	links=pagelinks(domain)	
	for x in links:
		print x
	print "\n-----------------------------\n"


	#calling and printing subdomains after pagelinks.

	subdomains_from_netcraft(domain)
	print "\n\t\t\t[+] List of subdomains found\n"
	for sub in subdomain_list:
		print sub
	
	#wikileaks
	print "\t\t\t[+] Associated WikiLeaks\n"
	leaklinks=wikileaks(domain)
	for tl,lnk in leaklinks.items():
		print "%s (%s)" % (lnk, tl)
	print "For all results, visit: "+ 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date'%(domain)
	print "\n-----------------------------\n"
	
	
	print "\t\t\t[+] Associated Forum Links\n"
	links=boardsearch_forumsearch(domain)
	for tl,lnk in links.items():
		print "%s (%s)" % (lnk, tl)
	print "\n-----------------------------\n"
	

	results = check_ssl_htbsecurity(domain)
	if 'ERROR' in results.keys():
		print results['ERROR']
	elif 'TOKEN' in results.keys():
		print 'Picking up One IP from bunch of IPs returned: %s' % results['MULTIPLE_IPS'][0]
		results_new = check_ssl_htbsecurity(results['MULTIPLE_IPS'][0])
		print "OverAll Rating: %s" % results_new['GRADE']
		print 'Check https://www.htbridge.com/ssl/ for more information'
		for x in results_new['VALUE'].keys():
			if str("[5]") in str(results_new['VALUE'][x]) or str("[3]") in str(results_new['VALUE'][x]):
				if x == 'httpHeaders':
					pass
				else:
					print results_new['VALUE'][x]
	else:
		print "OverAll Rating: %s" % results['GRADE']
		for x in results['VALUE'].keys():
			if str("[5]") in str(results['VALUE'][x]) or str("[3]") in str(results['VALUE'][x]):
				if x == 'httpHeaders':
					pass
				else:
					print results['VALUE'][x]
	print "\n-----------------------------\n"

	

	#checks results from zoomeye
	#filters need to be applied
	zoomeye_results = search_zoomeye(domain)
	dict_zoomeye_results = json.loads(zoomeye_results)
	if 'matches' in dict_zoomeye_results.keys():
		for x in dict_zoomeye_results['matches']:
			if x['site'].split('.')[-2] == domain.split('.')[-2]:
				print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (x['ip'], x['site'], x['title'], x['headers'].replace("\n",""), x['geoinfo'])
	print "\n-----------------------------\n"

	print "[+]\t Kicking off Censys Search. This may take a while.."
	censys_search(domain)
	for x in censys_list:
		print x
	print "\n-----------------------------\n"



	#checks for host ip and other details in shodan.

	'''
	#Code for shodan Ip search. now we are doing Hostname search.

	#converts domain to IP. Prints a statement for the same.
	ip_addr = socket.gethostbyname(domain)

	#checks for information at shodan, and comes back with whatever available.
	## need to apply filter here (lot of noise coming in)
	res_from_shodan = json.loads(shodansearch(ip_addr))
	#print res_from_shodan
	for iterate_shodan_list in res_from_shodan['data']:
		print "ISP: %s \n Hosts: %s \n IP: %s \n Data: %s\n" % (iterate_shodan_list['isp'], iterate_shodan_list['hostnames'], iterate_shodan_list['ip_str'], iterate_shodan_list['data'].strip("\n"))
	print "\n-----------------------------\n"
	'''

	res_from_shodan = json.loads(shodandomainsearch(domain))
	if 'matches' in res_from_shodan.keys():
		for x in res_from_shodan['matches']:
			print "IP: %s\nHosts: %s\nDomain: %s\nPort: %s\nData: %s\nLocation: %s\n" % (x['ip_str'], x['hostnames'], x['domains'], x['port'], x['data'].replace("\n",""), x['location'])
	print "-----------------------------\n"
Esempio n. 6
0
def main(): 
	domain = sys.argv[1]
	
	API_URL = "https://www.censys.io/api/v1"
	#print cfg.zoomeyeuser


	#print WhoIs information
	print whoisnew(domain)
	print "\n-----------------------------\n"


	#print DNS Information
	dns_records = parse_dns_records(domain)
	for x in dns_records.keys():
		print x
		if "No" in dns_records[x] and "Found" in dns_records[x]:
			print "\t%s" % (dns_records[x])
		else:
			for y in dns_records[x]:
				print "\t%s" % (y)
			#print type(dns_records[x])
	print "\n-----------------------------\n"

	#converts domain to IP. Prints a statement for the same.
	ip_addr = socket.gethostbyname(domain)

	#checks for information at shodan, and comes back with whatever available.
	## need to apply filter here (lot of noise coming in)
	res_from_shodan = json.loads(shodansearch(ip_addr))
	print res_from_shodan
	#for iterate_shodan_list in res_from_shodan['data']:
	#	print "ISP: %s \n Hosts: %s \n IP: %s \n Data: %s\n" % (iterate_shodan_list['isp'], iterate_shodan_list['hostnames'], iterate_shodan_list['ip_str'], iterate_shodan_list['data'].strip("\n"))
	print "\n-----------------------------\n"

	#checks results from zoomeye
	#filters need to be applied
	zoomeye_results = search_zoomeye(domain)
	dict_zoomeye_results = json.loads(zoomeye_results)
	print dict_zoomeye_results
	print "\n-----------------------------\n"

	#convert domain to reverse_domain for passing to checkpunkspider()
	reversed_domain = ""
	for x in reversed(domain.split(".")):
		reversed_domain = reversed_domain + "." + x
	reversed_domain = reversed_domain[1:]
	checkpunkspider(reversed_domain)
	print "\n-----------------------------\n"

	#make proper URL with domain. Check on ssl as well as 80.
	print "\t\t\t[+] Wapplyzing " + domain 
	print "Hitting HTTP:\n",
	try:
		targeturl = "http://" + domain
		wappalyzeit(targeturl)
	except:
		print "[-] HTTP connection was unavailable"
	print "Hitting HTTPS:\n",
	try:
		targeturl = "https://" + domain
		wappalyzeit(targeturl)
	except:
		print "[-] HTTPS connection was unavailable"
	print "\n-----------------------------\n"

	#make Search github code for the given domain.
	print github_search(domain, 'Code')
	print "\n-----------------------------\n"
	
	#collecting emails for the domain and adding information in master email list. 
	emailhunter(domain)
	print "\t\t\t[+] Finding Email Ids\n"
	for x in collected_emails:
		print str(x) + ", ",
	print "\n-----------------------------\n"


	dns_history = netcraft_domain_history(domain)
	for x in dns_history.keys():
		print "%s: %s" % (dns_history[x], x)
	print "\n-----------------------------\n"	

	#subdomains [to be called before pagelinks so as to avoid repititions.]
	print "\t\t\t[+] Finding Subdomains and appending\n"
	subdomains(domain)
	##print "\t\t\t[+] Check_subdomains from wolframalpha"
	##find_subdomains_from_wolfram(domain)
	print "\n-----------------------------\n"

	

	#domain pagelinks
	print "\t\t\t[+] Pagelinks\n"
	links=pagelinks(domain)	
	for x in links:
		print x
	print "\n-----------------------------\n"

		#wikileaks
	print "\t\t\t[+] Associated WikiLeaks\n"
	leaklinks=wikileaks(domain)
	for tl,lnk in leaklinks.items():
		print "%s (%s)" % (lnk, tl)
	print "For all results, visit: "+ 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date'%(domain)
	print "\n-----------------------------\n"
	
	
	print "\t\t\t[+] Associated Forum Links\n"
	links=boardsearch_forumsearch(domain)
	for tl,lnk in links.items():
		print "%s (%s)" % (lnk, tl)
	print "\n-----------------------------\n"
	
	censys_search(domain)
	for x in censys_list:
		print x


	subdomains_from_netcraft(domain)
	print "\n\t\t\t[+] List of subdomains found\n"
	for sub in subdomain_list:
		print sub
Esempio n. 7
0
def do_everything(domain):
    dict_to_apend['targetname'] = domain

    API_URL = "https://www.censys.io/api/v1"
    #print cfg.zoomeyeuser

    #print WhoIs information
    whoisdata = whoisnew(domain)
    print whoisdata
    dict_to_apend['whois'] = whoisdata

    #print DNS Information
    dns_records = parse_dns_records(domain)
    #dict_to_apend['dns_records'] = dns_records > not working
    #bson.errors.InvalidDocument: Cannot encode object: <DNS IN A rdata: 54.208.84.166>

    for x in dns_records.keys():
        print x
        if "No" in dns_records[x] and "Found" in dns_records[x]:
            print "\t%s" % (dns_records[x])
        else:
            for y in dns_records[x]:
                print "\t%s" % (y)
            #print type(dns_records[x])

    print colored(style.BOLD + '\n---> Finding Paste(s)..\n' + style.END,
                  'blue')
    if cfg.google_cse_key != "" and cfg.google_cse_key != "XYZ" and cfg.google_cse_cx != "" and cfg.google_cse_cx != "XYZ":
        total_results = google_search(domain, 1)
        if (total_results != 0 and total_results > 10):
            more_iters = (total_results / 10)
            if more_iters >= 10:
                print colored(
                    style.BOLD +
                    '\n---> Too many results, Daily API limit might exceed\n' +
                    style.END, 'red')
            for x in xrange(1, more_iters + 1):
                google_search(domain, (x * 10) + 1)
        print "\n\n-----------------------------\n"
    else:
        print colored(
            style.BOLD +
            '\n[-] google_cse_key and google_cse_cx not configured. Skipping paste(s) search.\nPlease refer to http://datasploit.readthedocs.io/en/latest/apiGeneration/.\n'
            + style.END, 'red')

    #convert domain to reverse_domain for passing to checkpunkspider()
    reversed_domain = ""
    for x in reversed(domain.split(".")):
        reversed_domain = reversed_domain + "." + x
    reversed_domain = reversed_domain[1:]
    res = checkpunkspider(reversed_domain)
    if 'data' in res.keys() and len(res['data']) >= 1:
        dict_to_apend['punkspider'] = res['data']
        print colored("[+] Few vulnerabilities found at Punkspider", 'green')
        for x in res['data']:
            print "==> ", x['bugType']
            print "Method:", x['verb'].upper()
            print "URL:\n" + x['vulnerabilityUrl']
            print "Param:", x['parameter']
    else:
        print colored("[-] No Vulnerabilities found on PunkSpider", 'red')

    print colored(
        style.BOLD + '\n---> Wapplyzing web page of base domain:\n' +
        style.END, 'blue')

    wappalyze_results = {}
    #make proper URL with domain. Check on ssl as well as 80.
    print "Hitting HTTP:\n",
    try:
        targeturl = "http://" + domain
        list_of_techs = wappalyzeit(targeturl)
        wappalyze_results['http'] = list_of_techs
    except:
        print "[-] HTTP connection was unavailable"
        wappalyze_results['http'] = []
    print "\nHitting HTTPS:\n",
    try:
        targeturl = "https://" + domain
        list_of_techs = wappalyzeit(targeturl)
        wappalyze_results['https'] = list_of_techs
    except:
        print "[-] HTTPS connection was unavailable"
        wappalyze_results['https'] = []

    if len(wappalyze_results.keys()) >= 1:
        dict_to_apend['wappalyzer'] = wappalyze_results

    #make Search github code for the given domain.

    git_results = github_search(domain, 'Code')
    if git_results is not None:
        print git_results
    else:
        print colored("Sad! Nothing found on github", 'red')

    #collecting emails for the domain and adding information in master email list.
    if cfg.emailhunter != "":
        emails = emailhunter(domain)
        if len(collected_emails) >= 1:
            for x in collected_emails:
                print str(x)
            dict_to_apend['email_ids'] = collected_emails
    '''
	##### code for automated osint on enumerated email email_ids

	while True:
		a = raw_input(colored("\n\nDo you want to launch osint check for these emails? [(Y)es/(N)o/(S)pecificEmail]: ", 'red'))
		if a.lower() =="yes" or a.lower() == "y":
			for x in collected_emails:
				print "Checking for %s" % x
				print_emailosint(x)
			break
		elif a.lower() =="no" or a.lower() == "n":
			break
		elif a.lower() =="s":
			while True:
				b = raw_input("Please Enter the EmailId you want to tun OSINT.) [(C)ancel?]: ")
				if b.lower() =="c":
					break
				else:
					print_emailosint(b)
					break
			break

		else:
			print("[-] Wrong choice. Please enter Yes or No  [Y/N]: \n")
		#print emailOsint.username_list
	'''

    dns_ip_history = netcraft_domain_history(domain)
    if len(dns_ip_history.keys()) >= 1:
        for x in dns_ip_history.keys():
            print "%s: %s" % (dns_ip_history[x], x)
        dict_to_apend['domain_ip_history'] = dns_ip_history

    #subdomains [to be called before pagelinks so as to avoid repititions.]
    subdomains(domain)
    ##print "---> Check_subdomains from wolframalpha"
    ##find_subdomains_from_wolfram(domain)

    #domain pagelinks
    links = pagelinks(domain)
    if len(links) >= 1:
        for x in links:
            print x
        dict_to_apend['pagelinks'] = links

    #calling and printing subdomains after pagelinks.

    subdomains_from_netcraft(domain)
    print colored(style.BOLD + '---> Finding subdomains: \n' + style.END,
                  'blue')
    time.sleep(0.9)
    if len(subdomain_list) >= 1:
        for sub in subdomain_list:
            print sub
        dict_to_apend['subdomains'] = subdomain_list

    #wikileaks
    leaklinks = wikileaks(domain)
    for tl, lnk in leaklinks.items():
        print "%s (%s)" % (lnk, tl)
    if len(leaklinks.keys()) >= 1:
        dict_to_apend['wikileaks'] = leaklinks
    print "For all results, visit: " + 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date' % (
        domain)

    links_brd = boardsearch_forumsearch(domain)
    for tl, lnk in links_brd.items():
        print "%s (%s)" % (lnk, tl)
    if len(links_brd.keys()) >= 1:
        dict_to_apend['forum_links'] = links_brd

    if cfg.zoomeyeuser != "" and cfg.zoomeyepass != "":
        temp_list = []
        zoomeye_results = search_zoomeye(domain)
        dict_zoomeye_results = json.loads(zoomeye_results)
        if 'matches' in dict_zoomeye_results.keys():
            print len(dict_zoomeye_results['matches'])
            for x in dict_zoomeye_results['matches']:
                if x['site'].split('.')[-2] == domain.split('.')[-2]:
                    temp_list.append(x)
                    if 'title' in x.keys():
                        print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (
                            x['ip'], x['site'], x['title'],
                            x['headers'].replace("\n", ""), x['geoinfo'])
                    else:
                        for val in x.keys():
                            print "%s: %s" % (val, x[val])
        if len(temp_list) >= 1:
            dict_to_apend['zoomeye'] = temp_list

    if cfg.censysio_id != "" and cfg.censysio_secret != "":
        print colored(
            style.BOLD +
            '\n---> Kicking off Censys Search. This may take a while..\n' +
            style.END, 'blue')
        censys_search(domain)
        if len(censys_list) >= 1:
            dict_to_apend['censys'] = censys_list
            for x in censys_list:
                if x is not None and x != 'None':
                    print x

    if cfg.shodan_api != "":
        res_from_shodan = json.loads(shodandomainsearch(domain))
        if 'matches' in res_from_shodan.keys():
            dict_to_apend['shodan'] = res_from_shodan['matches']
            for x in res_from_shodan['matches']:
                print "IP: %s\nHosts: %s\nDomain: %s\nPort: %s\nData: %s\nLocation: %s\n" % (
                    x['ip_str'], x['hostnames'], x['domains'], x['port'],
                    x['data'].replace("\n", ""), x['location'])
    '''
Esempio n. 8
0
def do_everything(domain):
	dict_to_apend['targetname'] = domain
	
	API_URL = "https://www.censys.io/api/v1"
	#print cfg.zoomeyeuser

	
	#print WhoIs information
	whoisdata = whoisnew(domain)
	print whoisdata
	dict_to_apend['whois'] = whoisdata


	
	#print DNS Information
	dns_records = parse_dns_records(domain)
	#dict_to_apend['dns_records'] = dns_records > not working 
	#bson.errors.InvalidDocument: Cannot encode object: <DNS IN A rdata: 54.208.84.166>
	
	for x in dns_records.keys():
		print x
		if "No" in dns_records[x] and "Found" in dns_records[x]:
			print "\t%s" % (dns_records[x])
		else:
			for y in dns_records[x]:
				print "\t%s" % (y)
			#print type(dns_records[x])
	
	print colored(style.BOLD + '\n---> Finding Paste(s)..\n' + style.END, 'blue')
	if cfg.google_cse_key != "" and cfg.google_cse_key != "XYZ" and cfg.google_cse_cx != "" and cfg.google_cse_cx != "XYZ":
		total_results = google_search(domain, 1)
		if (total_results != 0 and total_results > 10):
			more_iters = (total_results / 10)
			if more_iters >= 10:
					print colored(style.BOLD + '\n---> Too many results, Daily API limit might exceed\n' + style.END, 'red')
			for x in xrange(1,more_iters + 1):	
				google_search(domain, (x*10)+1)
		print "\n\n-----------------------------\n"
	else:
		print colored(style.BOLD + '\n[-] google_cse_key and google_cse_cx not configured. Skipping paste(s) search.\nPlease refer to http://datasploit.readthedocs.io/en/latest/apiGeneration/.\n' + style.END, 'red')


	#convert domain to reverse_domain for passing to checkpunkspider()
	reversed_domain = ""
	for x in reversed(domain.split(".")):
		reversed_domain = reversed_domain + "." + x
	reversed_domain = reversed_domain[1:]
	res = checkpunkspider(reversed_domain)
	if 'data' in res.keys() and len(res['data']) >= 1:
		dict_to_apend['punkspider'] = res['data']
		print colored("[+] Few vulnerabilities found at Punkspider", 'green'	)
		for x in res['data']:
			print "==> ", x['bugType']
			print "Method:", x['verb'].upper()
			print "URL:\n" + x['vulnerabilityUrl']
			print "Param:", x['parameter']
	else:
		print colored("[-] No Vulnerabilities found on PunkSpider", 'red')



	print colored(style.BOLD + '\n---> Wapplyzing web page of base domain:\n' + style.END, 'blue')


	wappalyze_results = {}
	#make proper URL with domain. Check on ssl as well as 80.
	print "Hitting HTTP:\n",
	try:
		targeturl = "http://" + domain
		list_of_techs = wappalyzeit(targeturl)
		wappalyze_results['http'] = list_of_techs
	except:
		print "[-] HTTP connection was unavailable"
		wappalyze_results['http'] = []
	print "\nHitting HTTPS:\n",
	try:
		targeturl = "https://" + domain
		list_of_techs = wappalyzeit(targeturl)
		wappalyze_results['https'] = list_of_techs
	except:
		print "[-] HTTPS connection was unavailable"
		wappalyze_results['https'] = []
	

	if len(wappalyze_results.keys()) >= 1:
		dict_to_apend['wappalyzer'] = wappalyze_results

	
	#make Search github code for the given domain.
	
	git_results = github_search(domain, 'Code')
	if git_results is not None:
		print git_results
	else:
		print colored("Sad! Nothing found on github", 'red')
	
	#collecting emails for the domain and adding information in master email list. 
	if cfg.emailhunter != "":
		emails = emailhunter(domain)
		if len(collected_emails) >= 1:
			for x in collected_emails:
				print str(x)	
			dict_to_apend['email_ids'] = collected_emails


	'''
	##### code for automated osint on enumerated email email_ids

	while True:
		a = raw_input(colored("\n\nDo you want to launch osint check for these emails? [(Y)es/(N)o/(S)pecificEmail]: ", 'red'))
		if a.lower() =="yes" or a.lower() == "y":	
			for x in collected_emails:
				print "Checking for %s" % x
				print_emailosint(x)
			break
		elif a.lower() =="no" or a.lower() == "n":
			break
		elif a.lower() =="s":
			while True:
				b = raw_input("Please Enter the EmailId you want to tun OSINT.) [(C)ancel?]: ")
				if b.lower() =="c":
					break
				else:
					print_emailosint(b)
					break
			break

		else:
			print("[-] Wrong choice. Please enter Yes or No  [Y/N]: \n")
		#print emailOsint.username_list
	'''


	
	dns_ip_history = netcraft_domain_history(domain)
	if len(dns_ip_history.keys()) >= 1:
		for x in dns_ip_history.keys():
			print "%s: %s" % (dns_ip_history[x], x)
		dict_to_apend['domain_ip_history'] = dns_ip_history


	#subdomains [to be called before pagelinks so as to avoid repititions.]
	subdomains(domain)
	##print "---> Check_subdomains from wolframalpha"
	##find_subdomains_from_wolfram(domain)

	

	#domain pagelinks
	links=pagelinks(domain)	
	if len(links) >= 1:
		for x in links:
			print x
		dict_to_apend['pagelinks'] = links

	
	#calling and printing subdomains after pagelinks.

	subdomains_from_netcraft(domain)
	print colored(style.BOLD + '---> Finding subdomains: \n' + style.END, 'blue')
	time.sleep(0.9)
	if len(subdomain_list) >= 1:
		for sub in subdomain_list:
			print sub
		dict_to_apend['subdomains'] = subdomain_list
	
	#wikileaks
	leaklinks=wikileaks(domain)
	for tl,lnk in leaklinks.items():
		print "%s (%s)" % (lnk, tl)
	if len(leaklinks.keys()) >= 1:
		dict_to_apend['wikileaks'] = leaklinks
	print "For all results, visit: "+ 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date'%(domain)
	
	

	links_brd =boardsearch_forumsearch(domain)
	for tl,lnk in links_brd.items():
		print "%s (%s)" % (lnk, tl)
	if len(links_brd.keys()) >= 1:
		dict_to_apend['forum_links'] = links_brd


	if cfg.zoomeyeuser != "" and cfg.zoomeyepass != "":
		temp_list =[]
		zoomeye_results = search_zoomeye(domain)
		dict_zoomeye_results = json.loads(zoomeye_results)
		if 'matches' in dict_zoomeye_results.keys():
			print len(dict_zoomeye_results['matches'])
			for x in dict_zoomeye_results['matches']:
				if x['site'].split('.')[-2] == domain.split('.')[-2]:
					temp_list.append(x)
					if 'title' in x.keys() :
						print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (x['ip'], x['site'], x['title'], x['headers'].replace("\n",""), x['geoinfo'])
					else:
						for val in x.keys():
							print "%s: %s" % (val, x[val])
		if len(temp_list) >= 1:
			dict_to_apend['zoomeye'] = temp_list


	if cfg.censysio_id != "" and cfg.censysio_secret != "":
		print colored(style.BOLD + '\n---> Kicking off Censys Search. This may take a while..\n' + style.END, 'blue')
		censys_search(domain)
		if len(censys_list) >= 1:
			dict_to_apend['censys'] = censys_list
			for x in censys_list:
				if x is not None and x != 'None':
					print x


	if cfg.shodan_api != "":
		res_from_shodan = json.loads(shodandomainsearch(domain))
		if 'matches' in res_from_shodan.keys():
			dict_to_apend['shodan'] = res_from_shodan['matches']
			for x in res_from_shodan['matches']:
				print "IP: %s\nHosts: %s\nDomain: %s\nPort: %s\nData: %s\nLocation: %s\n" % (x['ip_str'], x['hostnames'], x['domains'], x['port'], x['data'].replace("\n",""), x['location'])


	'''
Esempio n. 9
0
def do_everything(domain):
	dict_to_apend['targetname'] = domain
	
	API_URL = "https://www.censys.io/api/v1"
	#print cfg.zoomeyeuser

	
	#print WhoIs information
	whoisdata = whoisnew(domain)
	print whoisdata
	dict_to_apend['whois'] = whoisdata
	print "\n-----------------------------\n"


	
	#print DNS Information
	dns_records = parse_dns_records(domain)
	#dict_to_apend['dns_records'] = dns_records > not working 
	#bson.errors.InvalidDocument: Cannot encode object: <DNS IN A rdata: 54.208.84.166>
	
	for x in dns_records.keys():
		print x
		if "No" in dns_records[x] and "Found" in dns_records[x]:
			print "\t%s" % (dns_records[x])
		else:
			for y in dns_records[x]:
				print "\t%s" % (y)
			#print type(dns_records[x])
	print "\n-----------------------------\n"

	
	#convert domain to reverse_domain for passing to checkpunkspider()
	reversed_domain = ""
	for x in reversed(domain.split(".")):
		reversed_domain = reversed_domain + "." + x
	reversed_domain = reversed_domain[1:]
	res = checkpunkspider(reversed_domain)
	if 'data' in res.keys() and len(res['data']) >= 1:
		dict_to_apend['punkspider'] = res['data']
		print "[+] Few vulnerabilities found at Punkspider"	
		for x in res['data']:
			print "==> ", x['bugType']
			print "Method:", x['verb'].upper()
			print "URL:\n" + x['vulnerabilityUrl']
			print "Param:", x['parameter']
	else:
		print "[-] No Vulnerabilities found on PunkSpider"
	print "\n-----------------------------\n"



	#make proper URL with domain. Check on ssl as well as 80.
	#print "\t\t\t[+] Wapplyzing " + domain 
	print "Hitting HTTP: ",
	wappalyze_results = {}
	try:
		targeturl = "http://" + domain
		list_of_techs = wappalyzeit(targeturl)
		if len(list_of_techs) >= 1:
			wappalyze_results['http'] = list_of_techs
			for abc in list_of_techs:
				print abc,
		else:
			pass
		print "\n"
	except:	
		print "[-] HTTP connection was unavailable"
	
	print "Hitting HTTPS: ",
	try:
		targeturl = "https://" + domain
		list_of_techs = wappalyzeit(targeturl)
		if len(list_of_techs) >= 1:
			wappalyze_results['https'] = list_of_techs
			for abc in list_of_techs:
				print abc,
		else:
			pass
		print "\n"
	except:
		print "[-] HTTP connection was unavailable"
	if len(wappalyze_results.keys()) >= 1:
		dict_to_apend['wappalyzer'] = wappalyze_results




	
	#make Search github code for the given domain.
	git_results = github_search(domain, 'Code')
	if git_results is not None:
		print git_results
	else:
		print "Sad! Nothing found on github"
	print "\n-----------------------------\n"
	
	#collecting emails for the domain and adding information in master email list. 
	if cfg.emailhunter != "":
		emails = emailhunter(domain)
		print "\t\t\t[+] Finding Email Ids\n"
		if len(collected_emails) >= 1:
			for x in collected_emails:
				print str(x) + ", ",
			dict_to_apend['email_ids'] = collected_emails
		print "\n-----------------------------\n"


	
	dns_ip_history = netcraft_domain_history(domain)
	if len(dns_ip_history.keys()) >= 1:
		for x in dns_ip_history.keys():
			print "%s: %s" % (dns_ip_history[x], x)
		dict_to_apend['domain_ip_history'] = dns_ip_history
	print "\n-----------------------------\n"	

	
	
	#subdomains [to be called before pagelinks so as to avoid repititions.]
	print "\t\t\t[+] Finding Subdomains and appending\n"
	subdomains(domain)
	##print "\t\t\t[+] Check_subdomains from wolframalpha"
	##find_subdomains_from_wolfram(domain)
	print "\n-----------------------------\n"

	

	#domain pagelinks
	print "\t\t\t[+] Pagelinks\n"
	links=pagelinks(domain)	
	if len(links) >= 1:
		for x in links:
			print x
		dict_to_apend['pagelinks'] = links
	print "\n-----------------------------\n"

	
	#calling and printing subdomains after pagelinks.

	subdomains_from_netcraft(domain)
	print "\n\t\t\t[+] List of subdomains found\n"

	if len(subdomain_list) >= 1:
		for sub in subdomain_list:
			print sub
		dict_to_apend['subdomains'] = subdomain_list
	
	#wikileaks
	print "\t\t\t[+] Associated WikiLeaks\n"
	leaklinks=wikileaks(domain)
	for tl,lnk in leaklinks.items():
		print "%s (%s)" % (lnk, tl)
	if len(leaklinks.keys()) >= 1:
		dict_to_apend['wikileaks'] = leaklinks
	print "For all results, visit: "+ 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date'%(domain)
	print "\n-----------------------------\n"
	
	
	print "\t\t\t[+] Associated Forum Links\n"
	links_brd =boardsearch_forumsearch(domain)
	for tl,lnk in links_brd.items():
		print "%s (%s)" % (lnk, tl)
	if len(links_brd.keys()) >= 1:
		dict_to_apend['forum_links'] = links_brd
	print "\n-----------------------------\n"

	
	
	results = check_ssl_htbsecurity(domain)
	htb_res_dict = {}
	htb_res_lists = []
	if 'ERROR' in results.keys():
		print results['ERROR']
	elif 'TOKEN' in results.keys():
		if 'MULTIPLE_IPS' in results.keys():
			print 'Picking up One IP from bunch of IPs returned: %s' % results['MULTIPLE_IPS'][0]
			results_new = check_ssl_htbsecurity(results['MULTIPLE_IPS'][0])
			print "OverAll Rating: %s" % results_new['GRADE']
			htb_res_dict['rating'] = results_new['GRADE']
			print 'Check https://www.htbridge.com/ssl/ for more information'
			for x in results_new['VALUE'].keys():
				if str("[5]") in str(results_new['VALUE'][x]) or str("[3]") in str(results_new['VALUE'][x]):
					if x == 'httpHeaders':
						pass
					else:
						print results_new['VALUE'][x]
						htb_res_lists.append(results_new['VALUE'][x])
	else:
		print "OverAll Rating: %s" % results['GRADE']
		htb_res_dict['rating'] = results['GRADE']
		for x in results['VALUE'].keys():
			if str("[5]") in str(results['VALUE'][x]) or str("[3]") in str(results['VALUE'][x]):
				if x == 'httpHeaders':
					pass
				else:
					print results['VALUE'][x]
					htb_res_lists.append(results['VALUE'][x])
	htb_res_dict['issues'] = htb_res_lists
	dict_to_apend['ssl_scan'] = htb_res_dict
	print "\n-----------------------------\n"

	
	
	#checks results from zoomeye
	#filters need to be applied
	#zoomeye_results = search_zoomeye(domain)
	#dict_zoomeye_results = json.loads(zoomeye_results)
	#if 'matches' in dict_zoomeye_results.keys():
	#	for x in dict_zoomeye_results['matches']:
	#		if x['site'].split('.')[-2] == domain.split('.')[-2]:
	#			print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (x['ip'], x['site'], x['title'], x['headers'].replace("\n",""), x['geoinfo'])
	#print "\n-----------------------------\n"

	if cfg.zoomeyeuser != "" and cfg.zoomeyepass != "":
		temp_list =[]
		zoomeye_results = search_zoomeye(domain)
		dict_zoomeye_results = json.loads(zoomeye_results)
		if 'matches' in dict_zoomeye_results.keys():
			print len(dict_zoomeye_results['matches'])
			for x in dict_zoomeye_results['matches']:
				if x['site'].split('.')[-2] == domain.split('.')[-2]:
					temp_list.append(x)
					if 'title' in x.keys() :
						print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (x['ip'], x['site'], x['title'], x['headers'].replace("\n",""), x['geoinfo'])
					else:
						for val in x.keys():
							print "%s: %s" % (val, x[val])
		if len(temp_list) >= 1:
			dict_to_apend['zoomeye'] = temp_list
		print "\n-----------------------------\n"
		


	if cfg.censysio_id != "" and cfg.censysio_secret != "":
		print "[+]\t Kicking off Censys Search. This may take a while.."
		censys_search(domain)
		if len(censys_list) >= 1:
			dict_to_apend['censys'] = censys_list
			for x in censys_list:
				print x
		print "\n-----------------------------\n"
	

	
	#Code for shodan Ip search. now we are doing Hostname search.
		
		#converts domain to IP. Prints a statement for the same.
		#ip_addr = socket.gethostbyname(domain)

		#checks for information at shodan, and comes back with whatever available.
		## need to apply filter here (lot of noise coming in)
		#res_from_shodan = json.loads(shodansearch(ip_addr))
		#print res_from_shodan
		#for iterate_shodan_list in res_from_shodan['data']:
		#	print "ISP: %s \n Hosts: %s \n IP: %s \n Data: %s\n" % (iterate_shodan_list['isp'], iterate_shodan_list['hostnames'], iterate_shodan_list['ip_str'], iterate_shodan_list['data'].strip("\n"))
		#print "\n-----------------------------\n"
	

	
	if cfg.shodan_api != "":
		res_from_shodan = json.loads(shodandomainsearch(domain))
		if 'matches' in res_from_shodan.keys():
			dict_to_apend['shodan'] = res_from_shodan['matches']
			for x in res_from_shodan['matches']:
				print "IP: %s\nHosts: %s\nDomain: %s\nPort: %s\nData: %s\nLocation: %s\n" % (x['ip_str'], x['hostnames'], x['domains'], x['port'], x['data'].replace("\n",""), x['location'])
		print "-----------------------------\n"


	#insert data into mongodb instance

	result = db.domaindata.insert(dict_to_apend, check_keys=False)
	print 'done'
Esempio n. 10
0
def do_everything(domain, output, active):
	dict_to_apend['targetname'] = domain
	
	API_URL = "https://www.censys.io/api/v1"
	#print cfg.zoomeyeuser

	html = ""
	
	scroll = []

	#print WhoIs information
	whoisdata = whoisnew(domain)
	print whoisdata
	dict_to_apend['whois'] = whoisdata
	html, scroll = make_html(html, "WHOIS Information", "whois", "<pre>%s</pre>" % str(whoisdata), scroll)
	
	#print DNS Information
	dns_records = parse_dns_records(domain)
	#dict_to_apend['dns_records'] = dns_records > not working 
	#bson.errors.InvalidDocument: Cannot encode object: <DNS IN A rdata: 54.208.84.166>
	lhtml = "<ul class='collection'>"	
	for x in dns_records.keys():
		print x
		lhtml += "<li class='collection-item'>%s<ul class='collection'>" % x
		if "No" in dns_records[x] and "Found" in dns_records[x]:
			print "\t%s" % (dns_records[x])
			lhtml += "<li class='collection-item'>%s</li>" % dns_records[x]
		else:
			for y in dns_records[x]:
				if x == "TXT Records":
					lhtml += "<li class='collection-item'>%s</li>" % ", ".join(y.strings)
					print "\t%s" % ", ".join(y.strings)
				else:
					lhtml += "<li class='collection-item'>%s</li>" % str(y)
					print "\t%s" % (y)
			#print type(dns_records[x])
		lhtml += "</ul></li>"
	lhtml += "</ul>"
	html, scroll = make_html(html, "DNS Records", "dns-records", lhtml, scroll)
	

	
	print colored(style.BOLD + '\n---> Finding Paste(s)..\n' + style.END, 'blue')



	'''print colored(style.BOLD + '\n---> Finding Paste(s)..\n' + style.END, 'blue')
	if cfg.google_cse_key and cfg.google_cse_key != "XYZ" and cfg.google_cse_cx and cfg.google_cse_cx != "XYZ":
		total_results = google_search(domain, 1)
		if (total_results != 0 and total_results > 10):
			more_iters = (total_results / 10)
			if more_iters >= 10:
					print colored(style.BOLD + '\n---> Too many results, Daily API limit might exceed\n' + style.END, 'red')
			for x in xrange(1,more_iters + 1):	
				google_search(domain, (x*10)+1)
		print "\n\n-----------------------------------------n"
	else:
		print colored(style.BOLD + '\n[-] google_cse_key and google_cse_cx not configured. Skipping paste(s) search.\nPlease refer to http://datasploit.readthedocs.io/en/latest/apiGeneration/.\n' + style.END, 'red')
	'''


	if cfg.google_cse_key != "" and cfg.google_cse_key != "XYZ" and cfg.google_cse_cx != "" and cfg.google_cse_cx != "XYZ":
		total_results, results = google_search(domain, 1)
		if not total_results == 0:
			lhtml = """<div style='overflow-x: scroll;'><table class='bordered highlight'>
					<thead>
					<tr>
						<th>Title</th>
						<th>URL</th>
						<th>Snippet</th>
					</tr>
					</thead>
					<tbody>
			"""
			for x in results['items']:
				lhtml += "<tr>"
				lhtml += "<td>%s</td><td>%s</td><td>%s</td>" % (cgi.escape(x['title']), cgi.escape(x['link']), cgi.escape(x['snippet']))
				lhtml += "</tr>"
			if (total_results != 0 and total_results > 10):
				more_iters = (total_results / 10)
				if more_iters >= 10:
						print colored(style.BOLD + '\n---> Too many results, Daily API limit might exceed\n' + style.END, 'red')
				for x in xrange(1,more_iters + 1):	
					total_results, results = google_search(domain, (x*10)+1)
					if results:
						for y in results['items']:
							lhtml += "<tr>"
							lhtml += "<td>%s</td><td>%s</td><td>%s</td>" % (cgi.escape(y['title']), cgi.escape(y['link']), cgi.escape(y['snippet']))
							lhtml += "</tr>"
			print "\n\n-----------------------------\n"
			lhtml += "</tbody></table></div>"
			html, scroll = make_html(html, "Google Search Pastes", "google-search-pastes", lhtml, scroll)
	else:
		print colored(style.BOLD + '\n[-] google_cse_key and google_cse_cx not configured. Skipping paste(s) search.\nPlease refer to http://datasploit.readthedocs.io/en/latest/apiGeneration/.\n' + style.END, 'red')
	


	#convert domain to reverse_domain for passing to checkpunkspider()
	reversed_domain = ""
	for x in reversed(domain.split(".")):
		reversed_domain = reversed_domain + "." + x
	reversed_domain = reversed_domain[1:]
	res = checkpunkspider(reversed_domain)
	if 'data' in res.keys() and len(res['data']) >= 1:
		dict_to_apend['punkspider'] = res['data']
		print colored("[+] Few vulnerabilities found at Punkspider", 'green'	)
		lhtml = "<div style='overflow-x: scroll;'><table class='bordered highlight'><thead>"
		lhtml += "<tr>"
		lhtml += "<th>Bug Type</th>"
		lhtml += "<th>Method</th>"
		lhtml += "<th>URL</th>"
		lhtml += "<th>Parameter</th>"
		lhtml += "</tr></thead><tbody>"
		for x in res['data']:
			lhtml += "<tr>"
			print "==> ", x['bugType']
			lhtml += "<td>%s</td>" % x['bugType']
			print "Method:", x['verb'].upper()
			lhtml += "<td>%s</td>" % x['verb'].upper()
			print "URL:\n" + x['vulnerabilityUrl']
			lhtml += "<td>%s</td>" % x['vulnerabilityUrl']
			print "Param:", x['parameter']
			lhtml += "<td>%s</td>" % x['parameter']
			lhtml += "</tr>"
		lhtml += "</tbody></table></div>"
		html, scroll = make_html(html, "Punkspider Vulnerabilities", "punkspider-vulns", lhtml, scroll)
	else:
		print colored("[-] No Vulnerabilities found on PunkSpider", 'red')

	print colored(style.BOLD + '\n---> Wapplyzing web page of base domain:\n' + style.END, 'blue')


	wappalyze_results = {}
	#make proper URL with domain. Check on ssl as well as 80.
	print "Hitting HTTP:\n",
	lhtml = ""
	try:
		targeturl = "http://" + domain
		list_of_techs = wappalyzeit(targeturl)
		wappalyze_results['http'] = list_of_techs
	except:
		print "[-] HTTP connection was unavailable"
		wappalyze_results['http'] = []
	if wappalyze_results['http']:
		lhtml += "<ul class='collection'><li class='collection-item'>HTTP<ul class='collection'>"
		for i in wappalyze_results['http']:
			lhtml += "<li class='collection-item'>%s</li>" % i
		lhtml += "</ul></li></ul>"
	print "\nHitting HTTPS:\n",
	try:
		targeturl = "https://" + domain
		list_of_techs = wappalyzeit(targeturl)
		wappalyze_results['https'] = list_of_techs
	except:
		print "[-] HTTPS connection was unavailable"
		wappalyze_results['https'] = []
	if wappalyze_results['https']:
		lhtml += "<ul class='collection'><li class='collection-item'>HTTPS<ul class='collection'>"
		for i in wappalyze_results['https']:
			lhtml += "<li class='collection-item'>%s</li>" % i
		lhtml += "</ul></li></ul>"
	html, scroll = make_html(html, "WappAlyzer Report", "wappalyzer-report", lhtml, scroll)

	if len(wappalyze_results.keys()) >= 1:
		dict_to_apend['wappalyzer'] = wappalyze_results
	
	#make Search github code for the given domain.
	
	git_results = github_search(domain, 'Code')
	if git_results is not None:
		html, scroll = make_html(html, "Github Report", "github-report", git_results, scroll)
	else:
		print colored("Sad! Nothing found on github", 'red')
	
	#collecting emails for the domain and adding information in master email list. 
	if cfg.emailhunter != "":
		emails = emailhunter(domain)
		if len(collected_emails) >= 1:
			lhtml = "<ul class='collection'>"
			for x in collected_emails:
				lhtml += "<li class='collection-item'>%s</li>" % str(x)
				print str(x)	
			lhtml += "</ul>"
			html, scroll = make_html(html, "Email Harvesting", "email-harvesting", lhtml, scroll)
			dict_to_apend['email_ids'] = collected_emails

	'''
	##### code for automated osint on enumerated email email_ids

	while True:
		a = raw_input(colored("\n\nDo you want to launch osint check for these emails? [(Y)es/(N)o/(S)pecificEmail]: ", 'red'))
		if a.lower() =="yes" or a.lower() == "y":	
			for x in collected_emails:
				print "Checking for %s" % x
				print_emailosint(x)
			break
		elif a.lower() =="no" or a.lower() == "n":
			break
		elif a.lower() =="s":
			while True:
				b = raw_input("Please Enter the EmailId you want to tun OSINT.) [(C)ancel?]: ")
				if b.lower() =="c":
					break
				else:
					print_emailosint(b)
					break
			break

		else:
			print("[-] Wrong choice. Please enter Yes or No  [Y/N]: \n")
		#print emailOsint.username_list
	'''
	
	dns_ip_history = netcraft_domain_history(domain)
	if len(dns_ip_history.keys()) >= 1:
		lhtml = "<div style='overflow-x: scroll;'><table class='bordered highlight'><thead>"
		lhtml += "<tr><th>IP</th><th>Description</th></tr></thead><tbody>"
		for x in dns_ip_history.keys():
			lhtml += "<tr><td>%s</td><td>%s</td></tr>" % (dns_ip_history[x], x)
			print "%s: %s" % (dns_ip_history[x], x)
		lhtml += "</tbody></table></div>"
		dict_to_apend['domain_ip_history'] = dns_ip_history
	
	html, scroll = make_html(html, "DNS IP History", "dns-ip-history", lhtml, scroll)

	#subdomains [to be called before pagelinks so as to avoid repititions.]
	subdomains(domain)
	##print "---> Check_subdomains from wolframalpha"
	##find_subdomains_from_wolfram(domain)

	

	#domain pagelinks
	links=pagelinks(domain)	
	links = list(set(links))
	links.sort()
	if len(links) >= 1:
		lhtml = "<ul class='collection'>"
		for x in links:
			print x
			lhtml += "<li class='collection-item'>%s</li>" % x
		lhtml += "</ul>"
		dict_to_apend['pagelinks'] = links
		html, scroll = make_html(html, "Page Links", "page-links", lhtml, scroll)

	
	#calling and printing subdomains after pagelinks.

	subdomains_from_netcraft(domain)
	print colored(style.BOLD + '---> Finding subdomains: \n' + style.END, 'blue')
	time.sleep(0.9)
	if len(subdomain_list) >= 1:
		lhtml = "<ul class='collection'>"
		for sub in subdomain_list:
			lhtml += "<li class='collection-item'>%s</li>" % sub
			print sub
		lhtml += "</ul>"
		dict_to_apend['subdomains'] = subdomain_list
		html, scroll = make_html(html, "Subdomains", "subdomains", lhtml, scroll)

	#wikileaks
	leaklinks=wikileaks(domain)
	lhtml = "<div style='overflow-x: scroll;'><table class='bordered highlight'><thead><tr><th>URL</th><th>Description</th></tr></thead><tbody>"
	for tl,lnk in leaklinks.items():
		lhtml += "<tr><td>%s</td><td>%s</td></tr>" % (lnk, tl)
		print "%s (%s)" % (lnk, tl)
	lhtml += "</tbody></table></div>"
	html, scroll = make_html(html, "WikiLeaks", "wikileaks", lhtml, scroll)

	if len(leaklinks.keys()) >= 1:
		dict_to_apend['wikileaks'] = leaklinks
	print "For all results, visit: "+ 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date'%(domain)

	links_brd =boardsearch_forumsearch(domain)
	if links_brd:
		lhtml = "<ul class='collection'>" 
		for tl,lnk in links_brd.items():
			lhtml += "<li class='collection-item'>%s (%s)</li>" % (lnk, tl)
			print "%s (%s)" % (lnk, tl)
		lhtml += "</ul>"
		html, scroll = make_html(html, "Forum Search", "forum-search", lhtml, scroll)
	if len(links_brd.keys()) >= 1:
		dict_to_apend['forum_links'] = links_brd

	if cfg.zoomeyeuser != "" and cfg.zoomeyepass != "":
		temp_list =[]
		zoomeye_results = search_zoomeye(domain)
		dict_zoomeye_results = json.loads(zoomeye_results)
		if 'matches' in dict_zoomeye_results.keys():
			lhtml = ""
			print len(dict_zoomeye_results['matches'])
			for x in dict_zoomeye_results['matches']:
				if x['site'].split('.')[-2] == domain.split('.')[-2]:
					temp_list.append(x)
					if 'title' in x.keys():
						zoom_data = "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (x['ip'], x['site'], x['title'], x['headers'].replace("\n",""), x['geoinfo'])
						lhtml += zoom_data
						print zoom_data
					else:
						for val in x.keys():
							zoom_data = "%s: %s" % (val, x[val])
							lhtml += zoom_data
	                                                print zoom_data
			html, scroll = make_html(html, "Zoomeye Search", "zoomeye-search", "<pre>%s</pre>" % lhtml, scroll)

		if len(temp_list) >= 1:
			dict_to_apend['zoomeye'] = temp_list


	
	if cfg.censysio_id != "" and cfg.censysio_secret != "":
		print colored(style.BOLD + '\n---> Kicking off Censys Search. This may take a while..\n' + style.END, 'blue')
		censys_search(domain)
		if len(censys_list) >= 1:
			dict_to_apend['censys'] = censys_list
			lhtml = "<ul class='collection'>"
			for x in censys_list:
				if not str(x) == 'None':
					lhtml += "<li class='collection-item'><pre>%s</pre></li>" % cgi.escape(str(x))
					print x
			lhtml += "</ul>"
			html, scroll = make_html(html, "Censys Search", "censys-search", lhtml, scroll)
	

	if cfg.shodan_api != "":
		res_from_shodan = json.loads(shodandomainsearch(domain))
		if 'matches' in res_from_shodan.keys():
			dict_to_apend['shodan'] = res_from_shodan['matches']
			lhtml = "<ul class='collection'>"
			for x in res_from_shodan['matches']:
				shod_data = "IP: %s\nHosts: %s\nDomain: %s\nPort: %s\nData: %s\nLocation: %s\n" % (x['ip_str'], x['hostnames'], x['domains'], x['port'], x['data'].replace("\n",""), x['location'])
				lhtml += "<li class='collection-item'><pre>%s</pre></li>" % shod_data
				print shod_data
			lhtml += "</ul>"
			html, scroll = make_html(html, "Shodan Search", "shodan-search", lhtml, scroll)


	if output:
		scroll_html = ""
		for i in scroll:
			scroll_html += "<li><a href='#%s'>%s</a></li>" % i
		
		fh = open("base.html", "r")
		main_html = fh.read()
		fh.close()
		
		if not os.path.exists("reports/%s" % domain):
			os.mkdir("reports/%s" % domain)

		now_time = datetime.now().strftime("%Y-%m-%d_%H-%M-%S")	

		fh = open("reports/%s/%s_%s.html" % (domain, domain, now_time), "w")
		main_html = main_html.replace("{DOMAIN_NAME}", domain)
		main_html = main_html.replace("{MAIN_HTML}", html)
		main_html = main_html.replace("{scrollROLL_SECTION}", scroll_html)
		fh.write(main_html)
		fh.close()


		fh = open("reports/%s/%s_%s.json" % (domain, domain, now_time), "w")
		fh.write(str(dict_to_apend))
		fh.close()


		fh = open("reports/%s/%s_%s.subdomains.txt" % (domain, domain, now_time), "w")
		for x in subdomain_list:
			fh.write("\n%s" % x)
		fh.close()

		fh = open("reports/%s/%s_%s.emails.txt" % (domain, domain, now_time), "w")
		for x in collected_emails:
			fh.write("\n%s" % x)
		fh.close()

		print colored(style.BOLD + '---> Saving work... \n' + style.END, 'green')

		print "[+] HTML Report saved to: reports/%s/%s_%s.html" % (domain, domain, now_time)
		print "[+] JSON Report saved to: reports/%s/%s_%s.json" % (domain, domain, now_time)
		print "[+] List of Emails saved to: reports/%s/%s_%s.emails.txt" % (domain, domain, now_time)
		print "[+] List of Emails saved to: reports/%s/%s_%s.subdomains.txt" % (domain, domain, now_time)
		print "\n"

	if active == 'true':
		print colored(style.BOLD + '---> Running Active Scans: \n' + style.END, 'green')
		run_active(domain, subdomain_list, collected_emails)


	'''
Esempio n. 11
0
def do_everything(domain):
	dict_to_apend['targetname'] = domain
	
	API_URL = "https://www.censys.io/api/v1"
	#print cfg.zoomeyeuser

	
	#print WhoIs information
	whoisdata = whoisnew(domain)
	print whoisdata
	dict_to_apend['whois'] = whoisdata


	
	#print DNS Information
	dns_records = parse_dns_records(domain)
	#dict_to_apend['dns_records'] = dns_records > not working 
	#bson.errors.InvalidDocument: Cannot encode object: <DNS IN A rdata: 54.208.84.166>
	
	for x in dns_records.keys():
		print x
		if "No" in dns_records[x] and "Found" in dns_records[x]:
			print "\t%s" % (dns_records[x])
		else:
			for y in dns_records[x]:
				print "\t%s" % (y)
			#print type(dns_records[x])

	
	#convert domain to reverse_domain for passing to checkpunkspider()
	reversed_domain = ""
	for x in reversed(domain.split(".")):
		reversed_domain = reversed_domain + "." + x
	reversed_domain = reversed_domain[1:]
	res = checkpunkspider(reversed_domain)
	if 'data' in res.keys() and len(res['data']) >= 1:
		dict_to_apend['punkspider'] = res['data']
		print colored("[+] Few vulnerabilities found at Punkspider", 'green'	)
		for x in res['data']:
			print "==> ", x['bugType']
			print "Method:", x['verb'].upper()
			print "URL:\n" + x['vulnerabilityUrl']
			print "Param:", x['parameter']
	else:
		print colored("[-] No Vulnerabilities found on PunkSpider", 'red')



	#make proper URL with domain. Check on ssl as well as 80.
	#print "---> Wapplyzing " + domain 
	print colored(style.BOLD + '\n---> Wappalyzing web pages\n' + style.END, 'blue')
	time.sleep(0.3)
	print colored("->Trying Wapalyzer on HTTP: ", 'blue')
	wappalyze_results = {}
	try:
		targeturl = "http://" + domain
		list_of_techs = wappalyzeit(targeturl)
		if len(list_of_techs) >= 1:
			wappalyze_results['http'] = list_of_techs
			for abc in list_of_techs:
				print abc,
		else:
			pass
		print "\n"
	except:	
		print colored("[-] HTTP connection was unavailable", 'red')
	
	print colored("->Trying Wapalyzer on HTTPS: ", 'blue')
	try:
		targeturl = "https://" + domain
		list_of_techs = wappalyzeit(targeturl)
		if len(list_of_techs) >= 1:
			wappalyze_results['https'] = list_of_techs
			for abc in list_of_techs:
				print abc,
		else:
			pass
		print "\n"
	except:
		print colored("[-] HTTP connection was unavailable", 'red')
	if len(wappalyze_results.keys()) >= 1:
		dict_to_apend['wappalyzer'] = wappalyze_results




	
	#make Search github code for the given domain.
	
	git_results = github_search(domain, 'Code')
	if git_results is not None:
		print git_results
	else:
		print colored("Sad! Nothing found on github", 'red')
	
	#collecting emails for the domain and adding information in master email list. 
	if cfg.emailhunter != "":
		emails = emailhunter(domain)
		if len(collected_emails) >= 1:
			for x in collected_emails:
				print str(x) + ", ",
			dict_to_apend['email_ids'] = collected_emails

	while True:
		a = raw_input(colored("\n\nDo you want to launch osint check for these emails? [(Y)es/(N)o/(S)pecificEmail]: ", 'red'))
		if a.lower() =="yes" or a.lower() == "y":	
			for x in collected_emails:
				print "Checking for %s" % x
				print_emailosint(x)
			break
		elif a.lower() =="no" or a.lower() == "n":
			break
		elif a.lower() =="s":
			while True:
				b = raw_input("Please Enter the EmailId you want to tun OSINT.) [(C)ancel?]: ")
				if b.lower() =="c":
					break
				else:
					print_emailosint(b)
					break
			break

		else:
			print("[-] Wrong choice. Please enter Yes or No  [Y/N]: \n")
		#print emailOsint.username_list

	
	dns_ip_history = netcraft_domain_history(domain)
	if len(dns_ip_history.keys()) >= 1:
		for x in dns_ip_history.keys():
			print "%s: %s" % (dns_ip_history[x], x)
		dict_to_apend['domain_ip_history'] = dns_ip_history


	#subdomains [to be called before pagelinks so as to avoid repititions.]
	subdomains(domain)
	##print "---> Check_subdomains from wolframalpha"
	##find_subdomains_from_wolfram(domain)

	

	#domain pagelinks
	links=pagelinks(domain)	
	if len(links) >= 1:
		for x in links:
			print x
		dict_to_apend['pagelinks'] = links

	
	#calling and printing subdomains after pagelinks.

	subdomains_from_netcraft(domain)
	print colored(style.BOLD + '---> Finding subdomains: \n' + style.END, 'blue')
	time.sleep(0.9)
	if len(subdomain_list) >= 1:
		for sub in subdomain_list:
			print sub
		dict_to_apend['subdomains'] = subdomain_list
	
	#wikileaks
	leaklinks=wikileaks(domain)
	for tl,lnk in leaklinks.items():
		print "%s (%s)" % (lnk, tl)
	if len(leaklinks.keys()) >= 1:
		dict_to_apend['wikileaks'] = leaklinks
	print "For all results, visit: "+ 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date'%(domain)
	
	

	links_brd =boardsearch_forumsearch(domain)
	for tl,lnk in links_brd.items():
		print "%s (%s)" % (lnk, tl)
	if len(links_brd.keys()) >= 1:
		dict_to_apend['forum_links'] = links_brd

	
	print colored(style.BOLD + '\n---> Performing passive SSL Scan\n' + style.END, 'blue')
	results = check_ssl_htbsecurity(domain)
	htb_res_dict = {}
	htb_res_lists = []
	if 'ERROR' in results.keys():
		print results['ERROR']
	elif 'TOKEN' in results.keys():
		if 'MULTIPLE_IPS' in results.keys():
			print colored('Picking up One IP from bunch of IPs returned: %s', 'green') % results['MULTIPLE_IPS'][0]
			results_new = check_ssl_htbsecurity(results['MULTIPLE_IPS'][0])
			print "OverAll Rating: %s" % results_new['GRADE']
			htb_res_dict['rating'] = results_new['GRADE']
			print 'Check https://www.htbridge.com/ssl/ for more information'
			for x in results_new['VALUE'].keys():
				if str("[5]") in str(results_new['VALUE'][x]) or str("[3]") in str(results_new['VALUE'][x]):
					if x == 'httpHeaders':
						pass
					else:
						print results_new['VALUE'][x]
						htb_res_lists.append(results_new['VALUE'][x])
	else:
		print "OverAll Rating: %s" % results['GRADE']
		htb_res_dict['rating'] = results['GRADE']
		for x in results['VALUE'].keys():
			if str("[5]") in str(results['VALUE'][x]) or str("[3]") in str(results['VALUE'][x]):
				if x == 'httpHeaders':
					pass
				else:
					print results['VALUE'][x]
					htb_res_lists.append(results['VALUE'][x])
	htb_res_dict['issues'] = htb_res_lists
	dict_to_apend['ssl_scan'] = htb_res_dict

	
	
	#checks results from zoomeye
	#filters need to be applied
	#zoomeye_results = search_zoomeye(domain)
	#dict_zoomeye_results = json.loads(zoomeye_results)
	#if 'matches' in dict_zoomeye_results.keys():
	#	for x in dict_zoomeye_results['matches']:
	#		if x['site'].split('.')[-2] == domain.split('.')[-2]:
	#			print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (x['ip'], x['site'], x['title'], x['headers'].replace("\n",""), x['geoinfo'])
	#print "\n-----------------------------\n"

	if cfg.zoomeyeuser != "" and cfg.zoomeyepass != "":
		temp_list =[]
		zoomeye_results = search_zoomeye(domain)
		dict_zoomeye_results = json.loads(zoomeye_results)
		if 'matches' in dict_zoomeye_results.keys():
			print len(dict_zoomeye_results['matches'])
			for x in dict_zoomeye_results['matches']:
				if x['site'].split('.')[-2] == domain.split('.')[-2]:
					temp_list.append(x)
					if 'title' in x.keys() :
						print "IP: %s\nSite: %s\nTitle: %s\nHeaders: %s\nLocation: %s\n" % (x['ip'], x['site'], x['title'], x['headers'].replace("\n",""), x['geoinfo'])
					else:
						for val in x.keys():
							print "%s: %s" % (val, x[val])
		if len(temp_list) >= 1:
			dict_to_apend['zoomeye'] = temp_list
		



	if cfg.censysio_id != "" and cfg.censysio_secret != "":
		print colored(style.BOLD + '\n---> Kicking off Censys Search. This may take a while..\n' + style.END, 'blue')
		censys_search(domain)
		if len(censys_list) >= 1:
			dict_to_apend['censys'] = censys_list
			for x in censys_list:
				if x is not None and x != 'None':
					print x


	
	#Code for shodan Ip search. now we are doing Hostname search.
		
		#converts domain to IP. Prints a statement for the same.
		#ip_addr = socket.gethostbyname(domain)

		#checks for information at shodan, and comes back with whatever available.
		## need to apply filter here (lot of noise coming in)
		#res_from_shodan = json.loads(shodansearch(ip_addr))
		#print res_from_shodan
		#for iterate_shodan_list in res_from_shodan['data']:
		#	print "ISP: %s \n Hosts: %s \n IP: %s \n Data: %s\n" % (iterate_shodan_list['isp'], iterate_shodan_list['hostnames'], iterate_shodan_list['ip_str'], iterate_shodan_list['data'].strip("\n"))
		#print "\n-----------------------------\n"
	

	
	if cfg.shodan_api != "":
		res_from_shodan = json.loads(shodandomainsearch(domain))
		if 'matches' in res_from_shodan.keys():
			dict_to_apend['shodan'] = res_from_shodan['matches']
			for x in res_from_shodan['matches']:
				print "IP: %s\nHosts: %s\nDomain: %s\nPort: %s\nData: %s\nLocation: %s\n" % (x['ip_str'], x['hostnames'], x['domains'], x['port'], x['data'].replace("\n",""), x['location'])


	#insert data into mongodb instance
	try:
		result = db.domaindata.insert(dict_to_apend, check_keys=False)
		print 'output saved to MongoDb'
	except:
		print "More data than I can handle, hence not saved in MongoDb. Apologies."
Esempio n. 12
0
def main():
    domain = sys.argv[1]

    API_URL = "https://www.censys.io/api/v1"
    #print cfg.zoomeyeuser

    #print WhoIs information
    print whoisnew(domain)
    print "\n-----------------------------\n"

    #print DNS Information
    dns_records = parse_dns_records(domain)
    for x in dns_records.keys():
        print x
        if "No" in dns_records[x] and "Found" in dns_records[x]:
            print "\t%s" % (dns_records[x])
        else:
            for y in dns_records[x]:
                print "\t%s" % (y)
            #print type(dns_records[x])
    print "\n-----------------------------\n"

    #converts domain to IP. Prints a statement for the same.
    ip_addr = socket.gethostbyname(domain)

    #checks for information at shodan, and comes back with whatever available.
    ## need to apply filter here (lot of noise coming in)
    res_from_shodan = json.loads(shodansearch(ip_addr))
    print res_from_shodan
    #for iterate_shodan_list in res_from_shodan['data']:
    #	print "ISP: %s \n Hosts: %s \n IP: %s \n Data: %s\n" % (iterate_shodan_list['isp'], iterate_shodan_list['hostnames'], iterate_shodan_list['ip_str'], iterate_shodan_list['data'].strip("\n"))
    print "\n-----------------------------\n"

    #checks results from zoomeye
    #filters need to be applied
    zoomeye_results = search_zoomeye(domain)
    dict_zoomeye_results = json.loads(zoomeye_results)
    print dict_zoomeye_results
    print "\n-----------------------------\n"

    #convert domain to reverse_domain for passing to checkpunkspider()
    reversed_domain = ""
    for x in reversed(domain.split(".")):
        reversed_domain = reversed_domain + "." + x
    reversed_domain = reversed_domain[1:]
    checkpunkspider(reversed_domain)
    print "\n-----------------------------\n"

    #make proper URL with domain. Check on ssl as well as 80.
    print "\t\t\t[+] Wapplyzing " + domain
    print "Hitting HTTP:\n",
    try:
        targeturl = "http://" + domain
        wappalyzeit(targeturl)
    except:
        print "[-] HTTP connection was unavailable"
    print "Hitting HTTPS:\n",
    try:
        targeturl = "https://" + domain
        wappalyzeit(targeturl)
    except:
        print "[-] HTTPS connection was unavailable"
    print "\n-----------------------------\n"

    #make Search github code for the given domain.
    print github_search(domain, 'Code')
    print "\n-----------------------------\n"

    #collecting emails for the domain and adding information in master email list.
    emailhunter(domain)
    print "\t\t\t[+] Finding Email Ids\n"
    for x in collected_emails:
        print str(x) + ", ",
    print "\n-----------------------------\n"

    dns_history = netcraft_domain_history(domain)
    for x in dns_history.keys():
        print "%s: %s" % (dns_history[x], x)
    print "\n-----------------------------\n"

    #subdomains [to be called before pagelinks so as to avoid repititions.]
    print "\t\t\t[+] Finding Subdomains and appending\n"
    subdomains(domain)
    ##print "\t\t\t[+] Check_subdomains from wolframalpha"
    ##find_subdomains_from_wolfram(domain)
    print "\n-----------------------------\n"

    #domain pagelinks
    print "\t\t\t[+] Pagelinks\n"
    links = pagelinks(domain)
    for x in links:
        print x
    print "\n-----------------------------\n"

    #wikileaks
    print "\t\t\t[+] Associated WikiLeaks\n"
    leaklinks = wikileaks(domain)
    for tl, lnk in leaklinks.items():
        print "%s (%s)" % (lnk, tl)
    print "For all results, visit: " + 'https://search.wikileaks.org/?query=&exact_phrase=%s&include_external_sources=True&order_by=newest_document_date' % (
        domain)
    print "\n-----------------------------\n"

    print "\t\t\t[+] Associated Forum Links\n"
    links = boardsearch_forumsearch(domain)
    for tl, lnk in links.items():
        print "%s (%s)" % (lnk, tl)
    print "\n-----------------------------\n"

    censys_search(domain)
    for x in censys_list:
        print x

    subdomains_from_netcraft(domain)
    print "\n\t\t\t[+] List of subdomains found\n"
    for sub in subdomain_list:
        print sub