Ejemplo n.º 1
0
def main():
    parser = argparse.ArgumentParser()

    logging.getLogger().setLevel(logging.INFO)
    parser.add_argument(
        '--command',
        required=True,
        help='FFX command to run. Runtime arguments are handled '
        'using the %%args%% placeholder.')
    parser.add_argument('child_args',
                        nargs='*',
                        help='Arguments for the command.')
    AddCommonArgs(parser)
    AddTargetSpecificArgs(parser)
    args = parser.parse_args()

    # Prepare the arglist for "ffx". %args% is replaced with all positional
    # arguments given to the script.
    ffx_args = shlex.split(args.command)
    # replace %args% in the command with the given arguments.
    try:
        args_index = ffx_args.index('%args%')
        ffx_args[args_index:args_index + 1] = args.child_args
    except ValueError:
        # %args% is not present; use the command as-is.
        pass

    with GetDeploymentTargetForArgs(args) as target:
        target.Start()
        target.StartSystemLog(args.package)

        # Extend the lifetime of |pkg_repo| beyond InstallPackage so that the
        # package can be instantiated after resolution.
        with target.GetPkgRepo() as pkg_repo:
            target.InstallPackage(args.package)
            process = target.RunFFXCommand(ffx_args)

            # It's possible that components installed by this script may be
            # instantiated at arbitrary points in the future.
            # This script (specifically |pkg_repo|) must be kept alive until it
            # is explicitly terminated by the user, otherwise pkgsvr will
            # throw an error when launching components.
            logging.info('Command is now running. Press CTRL-C to exit.')
            try:
                while True:
                    time.sleep(1)
            except KeyboardInterrupt:
                pass

    return 0
Ejemplo n.º 2
0
def main():
  parser = argparse.ArgumentParser()

  logging.getLogger().setLevel(logging.INFO)
  parser.add_argument('--command',
                      required=True,
                      help='FFX command to run. Runtime arguments are handled '
                      'using the %%args%% placeholder.')
  AddCommonArgs(parser)
  AddTargetSpecificArgs(parser)
  args, runtime_args = parser.parse_known_args()

  command_substituted = [
      chunk.replace('%args%', ' '.join(runtime_args))
      for chunk in shlex.split(args.command)
  ]

  with GetDeploymentTargetForArgs(args) as target:
    target.Start()
    target.StartSystemLog(args.package)

    # Extend the lifetime of |pkg_repo| beyond InstallPackage so that the
    # package can be instantiated after resolution.
    with target.GetPkgRepo() as pkg_repo:
      target.InstallPackage(args.package)
      process = target.RunFFXCommand(command_substituted)

      # It's possible that components installed by this script may be
      # instantiated at arbitrary points in the future.
      # This script (specifically |pkg_repo|) must be kept alive until it
      # is explicitly terminated by the user, otherwise pkgsvr will
      # throw an error when launching components.
      logging.info('Command is now running. Press CTRL-C to exit.')
      try:
        while True:
          time.sleep(1)
      except KeyboardInterrupt:
        pass

  return 0
Ejemplo n.º 3
0
def RunTestOnFuchsiaDevice(script_cmd):
  """Preps Fuchsia device with pave and package update, then runs script."""

  parser = argparse.ArgumentParser()
  AddCommonArgs(parser)
  AddTargetSpecificArgs(parser)
  runner_script_args, test_args = parser.parse_known_args()
  ConfigureLogging(runner_script_args)

  # If out_dir is not set, assume the script is being launched
  # from the output directory.
  if not runner_script_args.out_dir:
    runner_script_args.out_dir = os.getcwd()

  # Create a temporary log file that Telemetry will look to use to build
  # an artifact when tests fail.
  temp_log_file = False
  if not runner_script_args.system_log_file:
    runner_script_args.system_log_file = os.path.join(tempfile.mkdtemp(),
                                                      'system-log')
    temp_log_file = True

  package_names = ['web_engine_with_webui', 'web_engine_shell']
  web_engine_dir = os.path.join(runner_script_args.out_dir, 'gen', 'fuchsia',
                                'engine')

  # Pass all other arguments to the gpu integration tests.
  script_cmd.extend(test_args)
  listener_process = None
  symbolizer_process = None
  try:
    with GetDeploymentTargetForArgs(runner_script_args) as target:
      target.Start()
      fuchsia_device_address, fuchsia_ssh_port = target._GetEndpoint()
      script_cmd.extend(
          ['--chromium-output-directory', runner_script_args.out_dir])
      script_cmd.extend(['--fuchsia-device-address', fuchsia_device_address])
      script_cmd.extend(['--fuchsia-ssh-config', target._GetSshConfigPath()])
      if fuchsia_ssh_port:
        script_cmd.extend(['--fuchsia-ssh-port', str(fuchsia_ssh_port)])
      script_cmd.extend(
          ['--fuchsia-system-log-file', runner_script_args.system_log_file])
      # Add to the script
      if runner_script_args.verbose:
        script_cmd.append('-v')

      # Set up logging of WebEngine
      listener_process = target.RunCommandPiped(['log_listener'],
                                                stdout=subprocess.PIPE,
                                                stderr=subprocess.STDOUT)
      build_ids_paths = map(
          lambda package_name: os.path.join(web_engine_dir, package_name,
                                            'ids.txt'), package_names)
      symbolizer_process = RunSymbolizer(
          listener_process.stdout, open(runner_script_args.system_log_file,
                                        'w'), build_ids_paths)

      # Keep the Amber repository live while the test runs.
      with target.GetAmberRepo():
        # Install necessary packages on the device.
        far_files = map(
            lambda package_name: os.path.join(web_engine_dir, package_name,
                                              package_name + '.far'),
            package_names)
        target.InstallPackage(far_files)
        return subprocess.call(script_cmd)
  finally:
    if temp_log_file:
      shutil.rmtree(os.path.dirname(runner_script_args.system_log_file))
    if listener_process:
      listener_process.kill()
    if symbolizer_process:
      symbolizer_process.kill()
Ejemplo n.º 4
0
def main():
  parser = argparse.ArgumentParser()
  AddTestExecutionArgs(parser)
  AddCommonArgs(parser)
  AddTargetSpecificArgs(parser)
  args = parser.parse_args()

  # Flag out_dir is required for tests launched with this script.
  if not args.out_dir:
    raise ValueError("out-dir must be specified.")

  # Code coverage uses runtests, which calls run_test_component.
  if args.code_coverage:
    args.use_run_test_component = True

  ConfigureLogging(args)

  child_args = []
  if args.test_launcher_shard_index != None:
    child_args.append(
        '--test-launcher-shard-index=%d' % args.test_launcher_shard_index)
  if args.test_launcher_total_shards != None:
    child_args.append(
        '--test-launcher-total-shards=%d' % args.test_launcher_total_shards)
  if args.single_process_tests:
    child_args.append('--single-process-tests')
  if args.test_launcher_bot_mode:
    child_args.append('--test-launcher-bot-mode')
  if args.test_launcher_batch_limit:
    child_args.append('--test-launcher-batch-limit=%d' %
                       args.test_launcher_batch_limit)

  # Only set --test-launcher-jobs if the caller specifies it, in general.
  # If the caller enables the test-server then we need to launch the right
  # number of instances to match the maximum number of parallel test jobs, so
  # in that case we set --test-launcher-jobs based on the number of CPU cores
  # specified for the emulator to use.
  test_concurrency = None
  if args.test_launcher_jobs:
    test_concurrency = args.test_launcher_jobs
  elif args.enable_test_server:
    if args.device == 'device':
      test_concurrency = DEFAULT_TEST_SERVER_CONCURRENCY
    else:
      test_concurrency = args.cpu_cores
  if test_concurrency:
    child_args.append('--test-launcher-jobs=%d' % test_concurrency)

  if args.gtest_filter:
    child_args.append('--gtest_filter=' + args.gtest_filter)
  if args.gtest_repeat:
    child_args.append('--gtest_repeat=' + args.gtest_repeat)
    child_args.append('--test-launcher-timeout=-1')
  if args.test_launcher_retry_limit:
    child_args.append(
        '--test-launcher-retry-limit=' + args.test_launcher_retry_limit)
  if args.gtest_break_on_failure:
    child_args.append('--gtest_break_on_failure')
  if args.test_launcher_summary_output:
    child_args.append('--test-launcher-summary-output=' + TEST_RESULT_PATH)
  if args.isolated_script_test_output:
    child_args.append('--isolated-script-test-output=' + TEST_RESULT_PATH)
  if args.isolated_script_test_perf_output:
    child_args.append('--isolated-script-test-perf-output=' +
                      TEST_PERF_RESULT_PATH)

  if args.child_arg:
    child_args.extend(args.child_arg)
  if args.child_args:
    child_args.extend(args.child_args)

  test_realms = []
  if args.use_run_test_component:
    test_realms = [TEST_REALM_NAME]

  try:
    with GetDeploymentTargetForArgs(args) as target, \
         SystemLogReader() as system_logger, \
         RunnerLogManager(args.runner_logs_dir, BuildIdsPaths(args.package)):
      target.Start()

      if args.system_log_file and args.system_log_file != '-':
        system_logger.Start(target, args.package, args.system_log_file)

      if args.test_launcher_filter_file:
        target.PutFile(args.test_launcher_filter_file,
                       TEST_FILTER_PATH,
                       for_package=args.package_name,
                       for_realms=test_realms)
        child_args.append('--test-launcher-filter-file=' + TEST_FILTER_PATH)

      test_server = None
      if args.enable_test_server:
        assert test_concurrency
        test_server = SetupTestServer(target, test_concurrency,
                                      args.package_name, test_realms)

      run_package_args = RunTestPackageArgs.FromCommonArgs(args)
      if args.use_run_test_component:
        run_package_args.test_realm_label = TEST_REALM_NAME
        run_package_args.use_run_test_component = True
      returncode = RunTestPackage(args.out_dir, target, args.package,
                                  args.package_name, child_args,
                                  run_package_args)

      if test_server:
        test_server.Stop()

      if args.code_coverage:
        # Copy all the files in the profile directory. /* is used instead
        # of recursively copying due to permission issues for the latter.
        target.GetFile(TEST_LLVM_PROFILE_PATH + '/*', args.code_coverage_dir)

      if args.test_launcher_summary_output:
        target.GetFile(TEST_RESULT_PATH,
                       args.test_launcher_summary_output,
                       for_package=args.package_name,
                       for_realms=test_realms)

      if args.isolated_script_test_output:
        target.GetFile(TEST_RESULT_PATH,
                       args.isolated_script_test_output,
                       for_package=args.package_name,
                       for_realms=test_realms)

      if args.isolated_script_test_perf_output:
        target.GetFile(TEST_PERF_RESULT_PATH,
                       args.isolated_script_test_perf_output,
                       for_package=args.package_name,
                       for_realms=test_realms)

      return returncode

  except:
    return HandleExceptionAndReturnExitCode()
Ejemplo n.º 5
0
def main():
    parser = argparse.ArgumentParser()
    AddTestExecutionArgs(parser)
    AddCommonArgs(parser)
    AddTargetSpecificArgs(parser)
    args = parser.parse_args()

    # Flag out_dir is required for tests launched with this script.
    if not args.out_dir:
        raise ValueError("out-dir must be specified.")

    if args.component_version == "2":
        args.use_run_test_component = False

    if (args.code_coverage and args.component_version != "2"
            and not args.use_run_test_component):
        if args.enable_test_server:
            # TODO(1254563): Tests that need access to the test server cannot be run
            # as test component under CFv1. Because code coverage requires it, force
            # the test to run as a test component. It is expected that test that tries
            # to use the external test server will fail.
            args.use_run_test_component = True
        else:
            raise ValueError('Collecting code coverage info requires using '
                             'run-test-component.')

    ConfigureLogging(args)

    child_args = []
    if args.test_launcher_shard_index != None:
        child_args.append('--test-launcher-shard-index=%d' %
                          args.test_launcher_shard_index)
    if args.test_launcher_total_shards != None:
        child_args.append('--test-launcher-total-shards=%d' %
                          args.test_launcher_total_shards)
    if args.single_process_tests:
        child_args.append('--single-process-tests')
    if args.test_launcher_bot_mode:
        child_args.append('--test-launcher-bot-mode')
    if args.test_launcher_batch_limit:
        child_args.append('--test-launcher-batch-limit=%d' %
                          args.test_launcher_batch_limit)

    # Only set --test-launcher-jobs if the caller specifies it, in general.
    # If the caller enables the test-server then we need to launch the right
    # number of instances to match the maximum number of parallel test jobs, so
    # in that case we set --test-launcher-jobs based on the number of CPU cores
    # specified for the emulator to use.
    test_concurrency = None
    if args.test_launcher_jobs:
        test_concurrency = args.test_launcher_jobs
    elif args.enable_test_server:
        if args.device == 'device':
            test_concurrency = DEFAULT_TEST_SERVER_CONCURRENCY
        else:
            test_concurrency = args.cpu_cores
    if test_concurrency:
        child_args.append('--test-launcher-jobs=%d' % test_concurrency)
    if args.test_launcher_print_test_stdio:
        child_args.append('--test-launcher-print-test-stdio=%s' %
                          args.test_launcher_print_test_stdio)

    if args.gtest_filter:
        child_args.append('--gtest_filter=' + args.gtest_filter)
    if args.gtest_repeat:
        child_args.append('--gtest_repeat=' + args.gtest_repeat)
        child_args.append('--test-launcher-timeout=-1')
    if args.test_launcher_retry_limit:
        child_args.append('--test-launcher-retry-limit=' +
                          args.test_launcher_retry_limit)
    if args.gtest_break_on_failure:
        child_args.append('--gtest_break_on_failure')
    if args.gtest_also_run_disabled_tests:
        child_args.append('--gtest_also_run_disabled_tests')

    if args.child_arg:
        child_args.extend(args.child_arg)
    if args.child_args:
        child_args.extend(args.child_args)

    test_realms = []
    if args.use_run_test_component:
        test_realms = [TEST_REALM_NAME]

    try:
        with GetDeploymentTargetForArgs(args) as target, \
             MakeTestOutputs(args.component_version,
                             target,
                             args.package_name,
                             test_realms) as test_outputs:
            if args.test_launcher_summary_output:
                child_args.append('--test-launcher-summary-output=' +
                                  test_outputs.GetDevicePath(TEST_RESULT_FILE))
            if args.isolated_script_test_output:
                child_args.append('--isolated-script-test-output=' +
                                  test_outputs.GetDevicePath(TEST_RESULT_FILE))
            if args.isolated_script_test_perf_output:
                child_args.append(
                    '--isolated-script-test-perf-output=' +
                    test_outputs.GetDevicePath(TEST_PERF_RESULT_FILE))

            target.Start()
            target.StartSystemLog(args.package)

            if args.test_launcher_filter_file:
                if args.component_version == "2":
                    # TODO(crbug.com/1279803): Until one can send file to the device when
                    # running a test, filter files must be read from the test package.
                    test_launcher_filter_files = map(
                        MapFilterFileToPackageFile,
                        args.test_launcher_filter_file.split(';'))
                    child_args.append('--test-launcher-filter-file=' +
                                      ';'.join(test_launcher_filter_files))
                else:
                    test_launcher_filter_files = args.test_launcher_filter_file.split(
                        ';')
                    with tempfile.NamedTemporaryFile(
                            'a+b') as combined_filter_file:
                        for filter_file in test_launcher_filter_files:
                            with open(filter_file, 'rb') as f:
                                combined_filter_file.write(f.read())
                        combined_filter_file.seek(0)
                        target.PutFile(combined_filter_file.name,
                                       TEST_FILTER_PATH,
                                       for_package=args.package_name,
                                       for_realms=test_realms)
                        child_args.append('--test-launcher-filter-file=' +
                                          TEST_FILTER_PATH)

            test_server = None
            if args.enable_test_server:
                assert test_concurrency
                test_server = SetupTestServer(target, test_concurrency,
                                              args.package_name, test_realms)

            run_package_args = RunTestPackageArgs.FromCommonArgs(args)
            if args.use_run_test_component:
                run_package_args.test_realm_label = TEST_REALM_NAME
                run_package_args.use_run_test_component = True
            if args.component_version == "2":
                run_package_args.output_directory = test_outputs.GetOutputDirectory(
                )
            returncode = RunTestPackage(target, test_outputs.GetFfxSession(),
                                        args.package, args.package_name,
                                        args.component_version, child_args,
                                        run_package_args)

            if test_server:
                test_server.Stop()

            if args.code_coverage:
                test_outputs.GetCoverageProfiles(args.code_coverage_dir)

            if args.test_launcher_summary_output:
                test_outputs.GetFile(TEST_RESULT_FILE,
                                     args.test_launcher_summary_output)

            if args.isolated_script_test_output:
                test_outputs.GetFile(TEST_RESULT_FILE,
                                     args.isolated_script_test_output)

            if args.isolated_script_test_perf_output:
                test_outputs.GetFile(TEST_PERF_RESULT_FILE,
                                     args.isolated_script_test_perf_output)

            return returncode

    except:
        return HandleExceptionAndReturnExitCode()
Ejemplo n.º 6
0
def RunTestOnFuchsiaDevice(script_cmd):
    """Preps Fuchsia device with pave and package update, then runs script."""

    parser = argparse.ArgumentParser()
    AddCommonArgs(parser)
    AddTargetSpecificArgs(parser)
    runner_script_args, test_args = parser.parse_known_args()
    ConfigureLogging(runner_script_args)

    # If out_dir is not set, assume the script is being launched
    # from the output directory.
    if not runner_script_args.out_dir:
        runner_script_args.out_dir = os.getcwd()

    # Create a temporary log file that Telemetry will look to use to build
    # an artifact when tests fail.
    clean_up_logs_on_exit = False
    if not runner_script_args.logs_dir:
        runner_script_args.logs_dir = tempfile.mkdtemp()

    package_names = ['web_engine_with_webui', 'web_engine_shell']
    web_engine_dir = os.path.join(runner_script_args.out_dir, 'gen', 'fuchsia',
                                  'engine')
    package_paths = map(
        lambda package_name: os.path.join(web_engine_dir, package_name),
        package_names)

    # Pass all other arguments to the gpu integration tests.
    script_cmd.extend(test_args)
    try:
        with GetDeploymentTargetForArgs(runner_script_args) as target:
            target.Start()
            target.StartSystemLog(package_paths)
            fuchsia_device_address, fuchsia_ssh_port = target._GetEndpoint()
            script_cmd.extend(
                ['--chromium-output-directory', runner_script_args.out_dir])
            script_cmd.extend(
                ['--fuchsia-device-address', fuchsia_device_address])
            script_cmd.extend(
                ['--fuchsia-ssh-config',
                 target._GetSshConfigPath()])
            if fuchsia_ssh_port:
                script_cmd.extend(
                    ['--fuchsia-ssh-port',
                     str(fuchsia_ssh_port)])
            script_cmd.extend([
                '--fuchsia-system-log-file',
                os.path.join(runner_script_args.logs_dir, 'system_log')
            ])
            # Add to the script
            if runner_script_args.verbose:
                script_cmd.append('-v')

            # Keep the package repository live while the test runs.
            with target.GetPkgRepo():
                # Install necessary packages on the device.
                far_files = map(
                    lambda package_name: os.path.join(
                        web_engine_dir, package_name, package_name + '.far'),
                    package_names)
                target.InstallPackage(far_files)
                return subprocess.call(script_cmd)
    finally:
        if clean_up_logs_on_exit:
            shutil.rmtree(runner_script_args.logs_dir)