Ejemplo n.º 1
0
def connect_and_pull_info():
    """ Connect to the proxy source and pull the proxies in JSON form """
    results = {}
    count = 0
    data = json.loads(urllib2.urlopen(PROXY_URL).read())
    for i in range(0, 60):
        try:
            results[count] = data[i]
            count += 1
        except IndexError:
            pass
    amount = len(results)
    if amount != 0:
        LOGGER.info("Found {} possible proxies, moving to connection attempts..".format(len(results)))
    else:
        LOGGER.warning("No usable proxies discovered")
    return results
Ejemplo n.º 2
0
def verify_xss_vulnerable(context, scripted_url):
    """ Verify if there's a vulnerability in the URL by checking
        if the alert exists as is in the HTML, if it does exist that
        means that the URL does not convert the alert
    >>> verify_xss_vulnerable("<HTML>" ,"https://www.google.com/webhp?safe=1")
    False """
    script = scripted_url.split("=")[1]
    if script in context:
        return True
    else:
        LOGGER.warning("Basic tests failed, moving to tampered data..")
        data = tamper_payload(scripted_url)
        url_data_tampered = get_context(data)
        script = data.split("=")[1]
        if script in url_data_tampered:
            return True
        else:
            return False
Ejemplo n.º 3
0
def run_hash_verification(hash_to_verify, hash_ver_file=None):
    """ Pointer to run the Hash Verification system"""
    if hash_ver_file is not None and hash_to_verify is None:
        try:
            total = len(open(hash_ver_file).readlines())
            LOGGER.info("Found a total of {} hashes in file..".format(total))
        except IOError:
            LOGGER.critical(
                "That file does not exist, check path and try again.")

        with open(hash_ver_file, "r+") as hashes:
            for h in hashes.readlines():
                question = prompt("Attempt to verify '{}'[y/N]: ".format(
                    h.strip()))
                if question.startswith("y"):
                    LOGGER.info("Analyzing hash: '{}'".format(h.strip()))
                    HashChecker(h.strip()).obtain_hash_type()
                    print("\n")
                else:
                    LOGGER.warning("Skipping '{}'..".format(h.strip()))
    else:
        LOGGER.info("Analyzing hash: '{}'".format(hash_to_verify))
        HashChecker(hash_to_verify).obtain_hash_type()
Ejemplo n.º 4
0
def run_dork_checker(dork, dork_file=None, proxy=None):
    """ Pointer to run a Dork Check on a given Google Dork """
    if dork is not None:
        LOGGER.info("Starting dork scan, using query: '{}'..".format(dork))
        try:
            LOGGER.info(
                DorkScanner(dork, dork_file=dork_file,
                            proxy=proxy).check_urls_for_queries())
        except HTTPError:
            LOGGER.fatal(GoogleBlockException(GOOGLE_TEMP_BLOCK_ERROR_MESSAGE))
    elif dork is None and dork_file is not None:
        if proxy is None:
            proxy_warn = "It is advised to use proxies while running "
            proxy_warn += "a dork list due to the temporary Google "
            proxy_warn += "bans.."
            LOGGER.warning(proxy_warn)
            question = prompt(
                "Would you like to find proxies with the built in finder first[y/N]: "
            )
            if question.upper().startswith("Y"):
                subprocess.call(["python", "pybelt.py", "-f"])
            else:
                pass
        try:
            with open("{}".format(dork_file)) as dork_list:
                for dork in dork_list.readlines():
                    LOGGER.info("Starting dork scan on {}..".format(
                        dork.strip()))
                    LOGGER.info(
                        DorkScanner(dork, dork_file=dork_file,
                                    proxy=proxy).check_urls_for_queries())
        except HTTPError:
            LOGGER.fatal(GoogleBlockException(GOOGLE_TEMP_BLOCK_ERROR_MESSAGE))
        except IOError:
            LOGGER.fatal(
                "The filename {} does not exist, please verify path and try again"
                .format(dork_file))
Ejemplo n.º 5
0
def run_xss_scan(url, url_file=None, proxy=None, user_agent=False):
    """ Pointer to run a XSS Scan on a given URL """
    proxy = proxy if proxy is not None else None
    header = RANDOM_USER_AGENT if user_agent is not False else None
    if proxy is not None:
        LOGGER.info("Proxy configured, running through: {}".format(proxy))
    if user_agent is True:
        LOGGER.info("Grabbed random user agent: {}".format(header))

    if url_file is not None:  # Scan a given file full of URLS
        file_path = url_file
        done = 0
        try:
            total = len(open(url_file).readlines())
            LOGGER.info("Found a total of {} URLS to scan..".format(total))
            with open(file_path) as urls:
                for url in urls.readlines():
                    if QUERY_REGEX.match(url.strip()):
                        question = prompt(
                            "Would you like to scan '{}' for XSS vulnerabilities[y/N]: "
                            .format(url.strip()))
                        if question.lower().startswith("y"):
                            done += 1

                            try:
                                if not xss.main(url.strip(),
                                                proxy=proxy,
                                                headers=header):
                                    LOGGER.info(
                                        "URL '{}' does not appear to be vulnerable to XSS"
                                        .format(url.strip()))
                                else:
                                    LOGGER.info(
                                        "URL '{}' appears to be vulnerable to XSS"
                                        .format(url.strip()))
                            except ConnectionError:
                                LOGGER.warning(
                                    "{} failed to respond, skipping..".format(
                                        url.strip()))

                            LOGGER.info(
                                "URLS scanned: {}, URLS left: {}".format(
                                    done, total - done))
                        else:
                            done += 1
                            pass
                    else:
                        done += 1
                        LOGGER.warn(
                            "URL '{}' does not contain a query (GET) parameter, skipping.."
                            .format(url.strip()))
            LOGGER.info("All URLS in file have been scanned, shutting down..")
        except IOError:
            LOGGER.fatal(
                "That file does not exist, verify path and try again.")

    else:  # Scan a single URL
        if QUERY_REGEX.match(url):
            LOGGER.info("Searching: {} for XSS vulnerabilities..".format(
                url, proxy=proxy, headers=header))
            if not xss.main(url, proxy=proxy, headers=header):
                LOGGER.error(
                    "{} does not appear to be vulnerable to XSS".format(url))
            else:
                LOGGER.info("{} seems to be vulnerable to XSS.".format(url))
        else:
            error_message = "The URL you provided does not contain a query "
            error_message += "(GET) parameter. In order for this scan you run "
            error_message += "successfully you will need to provide a URL with "
            error_message += "A query (GET) parameter example: http://127.0.0.1/php?id=2"
            LOGGER.fatal(error_message)
Ejemplo n.º 6
0
def run_sqli_scan(url,
                  url_file=None,
                  proxy=None,
                  user_agent=False,
                  tamper=None):
    """ Pointer to run a SQLi Scan on a given URL """
    error_message = "URL: '{}' threw an exception {} "
    error_message += "and Pybelt is unable to resolve the URL, "
    error_message += "this could mean that the URL is not allowing connections "
    error_message += "or that the URL is bad. Attempt to connect "
    error_message += "to the URL manually, if a connection occurs "
    error_message += "make an issue."
    if url_file is not None:  # Run through a file list
        file_path = url_file
        done = 0
        try:
            total = len(open(file_path).readlines())
            LOGGER.info("Found a total of {} urls in file {}..".format(
                total, file_path))
            with open(file_path) as urls:
                for url in urls.readlines():
                    if QUERY_REGEX.match(url.strip()):
                        question = prompt(
                            "Would you like to scan '{}' for SQLi vulnerabilities[y/N]: "
                            .format(url.strip()))
                        if question.lower().startswith("y"):
                            LOGGER.info("Starting scan on url: '{}'".format(
                                url.strip()))
                            try:
                                LOGGER.info(
                                    SQLiScanner(url.strip()).sqli_search())
                                done += 1
                                LOGGER.info(
                                    "URLS scanned: {}, URLS left: {}".format(
                                        done, total - done))
                            except urllib2.URLError:
                                done += 1
                                LOGGER.warning(
                                    "{} did not respond, skipping..".format(
                                        url.strip()))
                        else:
                            done += 1
                            pass
                    else:
                        done += 1
                        LOGGER.warn(
                            "URL '{}' does not contain a query (GET) parameter, skipping.."
                            .format(url.strip()))
                        pass
            LOGGER.info("No more URLS found in file, shutting down..")
        except HTTPError as e:
            LOGGER.fatal(error_message.format(url.strip(), e))
        except IOError as e:
            print e
            LOGGER.fatal(
                "That file does not exist, verify path and try again.")

    else:  # Run a single URL
        try:
            if QUERY_REGEX.match(url):
                LOGGER.info("Starting SQLi scan on '{}'..".format(url))
                LOGGER.info(SQLiScanner(url).sqli_search())
            else:
                LOGGER.error(
                    "URL does not contain a query (GET) parameter. Example: http://example.com/php?id=2"
                )
        except HTTPError as e:
            LOGGER.fatal(error_message.format(url, e))