コード例 #1
0
ファイル: measure_perf.py プロジェクト: BarriP/pyston-perf
 def save_report_callback(exe, benchmark, elapsed, size):
     old_val = model.get_result(args.save_report, benchmark)
     model.save_result(args.save_report, benchmark, elapsed, size)
     if old_val is not None and args.take_min:
         print "(prev min: %.2fs / %2.1fMB)" % (old_val[0], old_val[1]),
コード例 #2
0
ファイル: measure_perf.py プロジェクト: BarriP/pyston-perf
 def save_callback(exe, benchmark, elapsed, size):
     report_name = report_name_for_exe(exe)
     model.save_result(report_name, benchmark, elapsed, size)
コード例 #3
0
ファイル: measure_perf.py プロジェクト: BarriP/pyston-perf
def main():
    parser = argparse.ArgumentParser()
    parser.add_argument("--pyston_dir",
                        dest="pyston_dir",
                        action="store",
                        default=None)
    parser.add_argument("--submit", dest="submit", action="store_true")
    parser.add_argument("--no-run-pyston",
                        dest="run_pyston",
                        action="store_false",
                        default=True)
    parser.add_argument("--run-pyston-interponly",
                        dest="run_pyston_interponly",
                        action="store_true",
                        default=False)
    parser.add_argument("--run-pyston-nocache",
                        dest="run_pyston_nocache",
                        action="store_true",
                        default=False)
    parser.add_argument("--run-cpython",
                        action="store",
                        nargs="?",
                        default=None,
                        const="python")
    parser.add_argument("--run-pypy",
                        action="store",
                        nargs="?",
                        default=None,
                        const="pypy")
    parser.add_argument("--save",
                        dest="save_report",
                        action="store",
                        nargs="?",
                        default=None,
                        const="tmp")
    parser.add_argument("--compare",
                        dest="compare_to",
                        action="append",
                        nargs="?",
                        default=None,
                        const="tmp")
    parser.add_argument("--clear",
                        dest="clear",
                        action="store",
                        nargs="?",
                        default=None,
                        const="tmp")
    parser.add_argument("--use-previous", action="store_true")
    parser.add_argument("--save-by-commit",
                        dest="save_by_commit",
                        action="store_true")
    parser.add_argument("--view",
                        dest="view",
                        action="store",
                        nargs="?",
                        default=None,
                        const="last")
    parser.add_argument("--allow-dirty",
                        dest="allow_dirty",
                        action="store_true")
    parser.add_argument("--list-reports",
                        dest="list_reports",
                        action="store_true")
    parser.add_argument("--pyston-executables-subdir",
                        dest="pyston_executables_subdir",
                        action="store",
                        default=".")
    parser.add_argument("--pyston-executable",
                        dest="pyston_executable",
                        action="store")
    parser.add_argument("--pyston-executable-name", action="store")
    parser.add_argument("--run-times",
                        dest="run_times",
                        action="store",
                        default='1')
    parser.add_argument("--extra-jit-args",
                        dest="extra_jit_args",
                        action="append")
    parser.add_argument("--take-min", action="store_true")
    parser.add_argument("--benchmark-filter",
                        "--filter",
                        dest="benchmark_filter",
                        action="append")
    parser.add_argument("--all-benchmarks", action="store_true")
    args = parser.parse_args()

    if args.list_reports:
        for report_name in model.list_reports():
            print report_name
        return

    if args.clear:
        model.clear_report(args.clear)
        return

    executables = []

    callbacks = []
    filters = []

    if args.pyston_dir is None:
        args.pyston_dir = os.path.join(os.path.dirname(__file__),
                                       "../../pyston")

    extra_jit_args = args.extra_jit_args or []

    pyston_executable = args.pyston_executable
    if not pyston_executable:
        pyston_executable = os.path.join(
            args.pyston_dir,
            os.path.join(args.pyston_executables_subdir, "pyston_release"))

    if not args.view:
        assert os.path.exists(pyston_executable), pyston_executable

    pyston_executable_name = args.pyston_executable_name
    if pyston_executable and not pyston_executable_name:
        pyston_executable_name = os.path.basename(pyston_executable)
        if pyston_executable_name == "pyston_release":
            pyston_executable_name = "pyston"

    global_opts = {}
    global_opts['take_min'] = args.take_min
    global_opts['run_times'] = int(args.run_times)

    if args.run_pyston:
        executables.append(
            Executable([pyston_executable] + extra_jit_args,
                       pyston_executable_name, global_opts))

    if args.run_cpython:
        python_executable = args.run_cpython
        python_name = commands.getoutput(
            python_executable +
            " -c 'import sys; print \"cpython %d.%d\" % (sys.version_info.major, sys.version_info.minor)'"
        )
        executables.append(
            Executable([python_executable], python_name, global_opts))

    if args.run_pypy:
        pypy_executable = args.run_pypy
        pypy_build = commands.getoutput(
            pypy_executable +
            """ -c 'import sys; print "%s.%s.%s" % sys.pypy_version_info[:3]'"""
        )
        pypy_name = "pypy %s" % pypy_build
        executables.append(
            Executable([pypy_executable], pypy_name, global_opts))

    main_benchmarks = [
        "django_template3_10x.py",
        "pyxl_bench_10x.py",
        "sqlalchemy_imperative2_10x.py",
    ]

    perf_tracking_benchmarks = [
        "django_migrate.py",
        "virtualenv_bench.py",
        "interp2.py",
        "raytrace.py",
        "nbody.py",
        "fannkuch.py",
        "chaos.py",
        "fasta.py",
        "pidigits.py",
        "richards.py",
        "deltablue.py",
        "django_template2.py",
        "django_template.py",
    ]

    unaveraged_benchmarks = [
        "django_template3.py",
        "pyxl_bench.py",
        "pyxl_bench2.py",
        "sqlalchemy_imperative2.py",
        "pyxl_bench2_10x.py",
    ]

    compare_to_interp_benchmarks = [
        "django_migrate.py",
        "sre_parse_parse.py",
        "raytrace_small.py",
        "deltablue.py",
        "richards.py",
    ]

    if args.run_pyston_nocache:
        opts = dict(global_opts)
        opts['clear_cache'] = True
        executables.append(
            Executable([pyston_executable] + extra_jit_args, "pyston_nocache",
                       opts))

    if args.run_pyston_interponly:
        executables.append(
            Executable([pyston_executable, "-I"] + extra_jit_args,
                       "pyston_interponly", global_opts))
        unaveraged_benchmarks += set(compare_to_interp_benchmarks).difference(
            main_benchmarks)

        def interponly_filter(exe, benchmark):
            if exe.name != "pyston_interponly":
                return False
            return benchmark not in compare_to_interp_benchmarks

        filters.append(interponly_filter)

    if args.benchmark_filter:

        def benchmark_filter(exe, benchmark):
            return not any(
                [re.search(p, benchmark) for p in args.benchmark_filter])

        filters.append(benchmark_filter)

    benchmarks = ([Benchmark("(calibration)", False)] +
                  [Benchmark(b, True) for b in main_benchmarks] +
                  [Benchmark(b, False) for b in unaveraged_benchmarks])

    if args.all_benchmarks:
        benchmarks += [Benchmark(b, False) for b in perf_tracking_benchmarks]

    benchmark_dir = os.path.join(os.path.dirname(__file__), "benchmark_suite")

    git_rev = None

    if args.view:

        def view_filter(exe, benchmark):
            v = model.get_result(args.view, benchmark)
            if v is not None:
                return v
            return True

        filters.append(view_filter)

    if args.submit:

        def submit_callback(exe, benchmark, elapsed, size):
            benchmark = os.path.basename(benchmark)

            if benchmark.endswith(".py"):
                benchmark = benchmark[:-3]
            else:
                assert benchmark == "(calibration)" or benchmark.startswith(
                    "(geomean")

            if "cpython" in exe.name.lower():
                commitid = "default"
            elif "pypy" in exe.name.lower():
                commitid = "default"
            else:
                commitid = get_git_rev(args.pyston_dir, args.allow_dirty)
            codespeed_submit.submit(commitid=commitid,
                                    benchmark=benchmark,
                                    executable=exe.name,
                                    value=elapsed)
            codespeed_submit.submit(commitid=commitid,
                                    benchmark=(benchmark + "_maxrss"),
                                    executable=exe.name,
                                    value=size)

        callbacks.append(submit_callback)

    def report_name_for_exe(exe):
        if "cpython" in exe.name.lower():
            report_name = "cpython"
        elif "pypy" in exe.name.lower():
            report_name = exe.name.lower().replace(' ', '_')
        else:
            assert 'pyston' in exe.name.lower()
            report_name = "%s_%s" % (exe.name, git_rev)
        return report_name

    if args.save_by_commit:
        git_rev = git_rev or get_git_rev(args.pyston_dir, args.allow_dirty)

        def save_callback(exe, benchmark, elapsed, size):
            report_name = report_name_for_exe(exe)
            model.save_result(report_name, benchmark, elapsed, size)

        callbacks.append(save_callback)

    if args.compare_to:
        print "Comparing to '%s'" % args.compare_to

        def compare_callback(exe, benchmark, elapsed, size):
            for report_name in args.compare_to:
                v = model.get_result(report_name, benchmark)
                if v is None:
                    print "(no %s)" % report_name,
                else:
                    print "%s: %.2fs (%s%%)" % (
                        report_name, v[0], "{:5.1f}".format(
                            (elapsed - v[0]) / v[0] * 100)),

        callbacks.append(compare_callback)

    if args.save_report:
        assert len(
            executables) == 1, "Can't save a run on multiple executables"

        if not args.use_previous and args.save_report != args.view:
            model.clear_report(args.save_report)
        print "Saving results as '%s'" % args.save_report

        def save_report_callback(exe, benchmark, elapsed, size):
            old_val = model.get_result(args.save_report, benchmark)
            model.save_result(args.save_report, benchmark, elapsed, size)
            if old_val is not None and args.take_min:
                print "(prev min: %.2fs / %2.1fMB)" % (old_val[0], old_val[1]),

        callbacks.append(save_report_callback)

    tmp_results = []

    def save_last_callback(exe, benchmark, elapsed, size):
        tmp_results.append((exe, benchmark, elapsed, size))

    callbacks.append(save_last_callback)

    if args.use_previous:
        if args.save_report:
            skip_report_name = lambda exe: args.save_report
        else:
            git_rev = git_rev or get_git_rev(args.pyston_dir, args.allow_dirty)
            skip_report_name = report_name_for_exe

        def repeated_filter(exe, benchmark):
            v = model.get_result(skip_report_name(exe), benchmark)
            if v:
                return v
            return False

        filters.append(repeated_filter)

    try:
        run_tests(executables, benchmarks, filters, callbacks, benchmark_dir)
    # except KeyboardInterrupt:
    # print "Interrupted"
    # sys.exit(1)
    finally:
        model.clear_report("last")
        print "Saving results to 'last'"
        for (exe, benchmark, elapsed, size) in tmp_results:
            model.save_result("last", benchmark, elapsed, size)
コード例 #4
0
 def save_report_callback(exe, benchmark, elapsed):
     old_val = model.get_result(args.save_report, benchmark)
     model.save_result(args.save_report, benchmark, elapsed)
     if old_val is not None and args.take_min:
         print "(prev min: %.2fs)" % (old_val,),
コード例 #5
0
ファイル: measure_perf.py プロジェクト: rudi-c/pyston-perf
 def save_callback(exe, benchmark, elapsed):
     report_name = report_name_for_exe(exe)
     model.save_result(report_name, benchmark, elapsed)
コード例 #6
0
ファイル: measure_perf.py プロジェクト: rudi-c/pyston-perf
def main():
    parser = argparse.ArgumentParser()
    parser.add_argument("--pyston_dir", dest="pyston_dir", action="store", default=None)
    parser.add_argument("--submit", dest="submit", action="store_true")
    parser.add_argument("--no-run-pyston", dest="run_pyston", action="store_false", default=True)
    parser.add_argument("--run-pyston-interponly", dest="run_pyston_interponly", action="store_true", default=False)
    parser.add_argument("--run-pyston-nocache", dest="run_pyston_nocache", action="store_true", default=False)
    parser.add_argument("--run-cpython", dest="run_cpython", action="store_true")
    parser.add_argument("--run-pypy", action="store_true")
    parser.add_argument("--save", dest="save_report", action="store", nargs="?", default=None, const="tmp")
    parser.add_argument("--compare", dest="compare_to", action="append", nargs="?", default=None, const="tmp")
    parser.add_argument("--clear", dest="clear", action="store", nargs="?", default=None, const="tmp")
    parser.add_argument("--use-previous", action="store_true")
    parser.add_argument("--save-by-commit", dest="save_by_commit", action="store_true")
    parser.add_argument("--view", dest="view", action="store", nargs="?", default=None, const="last")
    parser.add_argument("--allow-dirty", dest="allow_dirty", action="store_true")
    parser.add_argument("--list-reports", dest="list_reports", action="store_true")
    parser.add_argument("--pyston-executables-subdir", dest="pyston_executables_subdir", action="store", default=".")
    parser.add_argument("--pyston-executable", dest="pyston_executable", action="store")
    parser.add_argument("--run-times", dest="run_times", action="store", default='1')
    parser.add_argument("--extra-jit-args", dest="extra_jit_args", action="append")
    parser.add_argument("--take-min", action="store_true")
    parser.add_argument("--benchmark-filter", "--filter", dest="benchmark_filter", action="append")
    args = parser.parse_args()

    if args.list_reports:
        for report_name in model.list_reports():
            print report_name
        return

    if args.clear:
        model.clear_report(args.clear)
        return

    executables = []

    callbacks = []
    filters = []

    if args.pyston_dir is None:
        args.pyston_dir = os.path.join(os.path.dirname(__file__), "../../pyston")

    extra_jit_args = args.extra_jit_args or []

    pyston_executable = args.pyston_executable
    if not pyston_executable:
        pyston_executable = os.path.join(args.pyston_dir, os.path.join(args.pyston_executables_subdir, "pyston_release"))

    if not args.view:
        assert os.path.exists(pyston_executable), pyston_executable

    global_opts = {}
    global_opts['take_min'] = args.take_min

    if args.run_pyston:
        opts = dict(global_opts)
        opts['run_times'] = int(args.run_times)
        executables.append(Executable([pyston_executable] + extra_jit_args, "pyston", opts))

    if args.run_cpython:
        python_executable = "python"
        python_name = commands.getoutput(python_executable +
                " -c 'import sys; print \"cpython %d.%d\" % (sys.version_info.major, sys.version_info.minor)'")
        executables.append(Executable([python_executable], python_name, global_opts))

    if args.run_pypy:
        pypy_executable = "pypy"
        pypy_build = commands.getoutput(pypy_executable +
                " -c 'import platform; print platform.python_build()[0]'")
        pypy_name = "pypy %s" % pypy_build.split('+')[0]
        executables.append(Executable([pypy_executable], pypy_name, global_opts))

    only_pyston = args.run_pyston and len(executables) == 1

    averaged_benchmarks = [
        "django_template.py",
        "pyxl_bench.py",
        "sqlalchemy_imperative2.py",
        "django_migrate.py",
        "virtualenv_bench.py",
        "interp2.py",
        "raytrace.py",
        "nbody.py",
        "fannkuch.py",
        "chaos.py",
        "fasta.py",
        "pidigits.py",
        "richards.py",
        "deltablue.py",
        ]

    unaveraged_benchmarks = [
            ]

    compare_to_interp_benchmarks = [
            "django_migrate.py",
            "sre_parse_parse.py",
            "raytrace_small.py",
            "deltablue.py",
            "richards.py",
            ]

    if args.run_pyston_nocache:
        opts = dict(global_opts)
        opts['clear_cache'] = True
        executables.append(Executable([pyston_executable] + extra_jit_args, "pyston_nocache", opts))

    if args.run_pyston_interponly:
        executables.append(Executable([pyston_executable, "-I"] + extra_jit_args, "pyston_interponly", global_opts))
        unaveraged_benchmarks += set(compare_to_interp_benchmarks).difference(averaged_benchmarks)

        def interponly_filter(exe, benchmark):
            if exe.name != "pyston_interponly":
                return False
            return benchmark not in compare_to_interp_benchmarks
        filters.append(interponly_filter)

    if args.benchmark_filter:
        def benchmark_filter(exe, benchmark):
            return not any([re.search(p, benchmark) for p in args.benchmark_filter])
        filters.append(benchmark_filter)

    benchmarks = ([Benchmark("(calibration)", False)] +
            [Benchmark(b, True) for b in averaged_benchmarks] +
            [Benchmark(b, False) for b in unaveraged_benchmarks])

    benchmark_dir = os.path.join(os.path.dirname(__file__), "benchmark_suite")

    git_rev = None

    if args.view:
        def view_filter(exe, benchmark):
            v = model.get_result(args.view, benchmark)
            if v is not None:
                return v
            return True
        filters.append(view_filter)

    if args.submit:
        def submit_callback(exe, benchmark, elapsed):
            benchmark = os.path.basename(benchmark)

            if benchmark.endswith(".py"):
                benchmark = benchmark[:-3]
            else:
                assert benchmark == "(calibration)" or benchmark.startswith("(geomean")

            if "cpython" in exe.name.lower():
                commitid = "default"
            elif "pypy" in exe.name.lower():
                commitid = pypy_build
            else:
                commitid = get_git_rev(args.pyston_dir, args.allow_dirty)
            codespeed_submit.submit(commitid=commitid, benchmark=benchmark, executable=exe.name, value=elapsed)
        callbacks.append(submit_callback)

    def report_name_for_exe(exe):
        if "cpython" in exe.name.lower():
            report_name = "cpython"
        elif "pypy" in exe.name.lower():
            report_name = exe.name.lower().replace(' ', '_')
        else:
            assert 'pyston' in exe.name.lower()
            report_name = "%s_%s" % (exe.name, git_rev)
        return report_name

    if args.save_by_commit:
        git_rev = git_rev or get_git_rev(args.pyston_dir, args.allow_dirty)
        def save_callback(exe, benchmark, elapsed):
            report_name = report_name_for_exe(exe)
            model.save_result(report_name, benchmark, elapsed)
        callbacks.append(save_callback)

    if args.compare_to:
        print "Comparing to '%s'" % args.compare_to
        def compare_callback(exe, benchmark, elapsed):
            for report_name in args.compare_to:
                v = model.get_result(report_name, benchmark)
                if v is None:
                    print "(no %s)" % report_name,
                else:
                    print "%s: %.2f (%+0.1f%%)" % (report_name, v, (elapsed - v) / v * 100),
        callbacks.append(compare_callback)

    if args.save_report:
        assert len(executables) == 1, "Can't save a run on multiple executables"

        if not args.use_previous and args.save_report != args.view:
            model.clear_report(args.save_report)
        print "Saving results as '%s'" % args.save_report
        def save_report_callback(exe, benchmark, elapsed):
            old_val = model.get_result(args.save_report, benchmark)
            model.save_result(args.save_report, benchmark, elapsed)
            if old_val is not None and args.take_min:
                print "(prev min: %.2fs)" % (old_val,),
        callbacks.append(save_report_callback)

    tmp_results = []
    def save_last_callback(exe, benchmark, elapsed):
        tmp_results.append((exe, benchmark, elapsed))
    callbacks.append(save_last_callback)

    if args.use_previous:
        if args.save_report:
            skip_report_name = lambda exe: args.save_report
        else:
            git_rev = git_rev or get_git_rev(args.pyston_dir, args.allow_dirty)
            skip_report_name = report_name_for_exe
        def repeated_filter(exe, benchmark):
            v = model.get_result(skip_report_name(exe), benchmark)
            if v:
                return v
            return False
        filters.append(repeated_filter)

    try:
        run_tests(executables, benchmarks, filters, callbacks, benchmark_dir)
    except KeyboardInterrupt:
        print "Interrupted"
        sys.exit(1)
    finally:
        model.clear_report("last")
        print "Saving results to 'last'"
        for (exe, benchmark, elapsed) in tmp_results:
            model.save_result("last", benchmark, elapsed)