def testRunPerformanceTestsTelemetryCommandGenerator_StorySelectionConfig_Abridged(self): options = run_performance_tests.parse_arguments([ '../../tools/perf/run_benchmark', '--browser=release_x64', '--run-ref-build', r'--isolated-script-test-output=c:\a\b\c\output.json', ]) story_selection_config = { 'abridged': True, } command = run_performance_tests.TelemetryCommandGenerator( 'fake_benchmark_name', options, story_selection_config).generate( 'fake_output_dir') self.assertIn('--run-abridged-story-set', command)
def testRunPerformanceTestsTelemetryCommandGenerator_ReferenceBrowserComeLast(self): """This tests for crbug.com/928928.""" options = run_performance_tests.parse_arguments([ '../../tools/perf/run_benchmark', '--browser=release_x64', '--run-ref-build', '--test-shard-map-filename=win-10-perf_map.json', r'--isolated-script-test-output=c:\a\b\c\output.json', ]) self.assertIn('--browser=release_x64', options.passthrough_args) command = run_performance_tests.TelemetryCommandGenerator( 'fake_benchmark_name', options, is_reference=True).generate( 'fake_output_dir') original_browser_arg_index = command.index('--browser=release_x64') reference_browser_arg_index = command.index('--browser=reference') self.assertTrue(reference_browser_arg_index > original_browser_arg_index)