Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(109)

Unified Diff: build/android/test_runner.py

Issue 2198613002: [Android] Get rid of old perf test runner. (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@master
Patch Set: get rid of test_env Created 4 years, 5 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « build/android/pylib/utils/test_environment.py ('k') | build/android/test_runner.pydeps » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: build/android/test_runner.py
diff --git a/build/android/test_runner.py b/build/android/test_runner.py
index 0fb133cc5ddb82d7de0cdab3285c8144087907b4..362c0afad3d9ed1bfc542d7d824a2795d3754f49 100755
--- a/build/android/test_runner.py
+++ b/build/android/test_runner.py
@@ -38,9 +38,6 @@ from pylib.junit import setup as junit_setup
from pylib.junit import test_dispatcher as junit_dispatcher
from pylib.monkey import setup as monkey_setup
from pylib.monkey import test_options as monkey_test_options
-from pylib.perf import setup as perf_setup
-from pylib.perf import test_options as perf_test_options
-from pylib.perf import test_runner as perf_test_runner
from pylib.results import json_results
from pylib.results import report_results
@@ -612,29 +609,6 @@ def AddPerfTestOptions(parser):
AddDeviceOptions(parser)
-def ProcessPerfTestOptions(args):
- """Processes all perf test options.
-
- Args:
- args: argparse.Namespace object.
-
- Returns:
- A PerfOptions named tuple which contains all options relevant to
- perf tests.
- """
- # TODO(jbudorick): Move single_step handling down into the perf tests.
- if args.single_step:
- args.single_step = ' '.join(args.single_step_command)
- # TODO(jbudorick): Get rid of PerfOptions.
- return perf_test_options.PerfOptions(
- args.steps, args.flaky_steps, args.output_json_list,
- args.print_step, args.no_timeout, args.test_filter,
- args.dry_run, args.single_step, args.collect_chartjson_data,
- args.output_chartjson_data, args.get_output_dir_archive,
- args.max_battery_temp, args.min_battery_level,
- args.known_devices_file)
-
-
def AddPythonTestOptions(parser):
group = parser.add_argument_group('Python Test Options')
group.add_argument(
@@ -700,50 +674,6 @@ def _RunMonkeyTests(args, devices):
return exit_code
-def _RunPerfTests(args, active_devices):
- """Subcommand of RunTestsCommands which runs perf tests."""
- perf_options = ProcessPerfTestOptions(args)
-
- # Just save a simple json with a list of test names.
- if perf_options.output_json_list:
- return perf_test_runner.OutputJsonList(
- perf_options.steps, perf_options.output_json_list)
-
- # Just print the results from a single previously executed step.
- if perf_options.print_step:
- return perf_test_runner.PrintTestOutput(
- perf_options.print_step, perf_options.output_chartjson_data,
- perf_options.get_output_dir_archive)
-
- runner_factory, tests, devices = perf_setup.Setup(
- perf_options, active_devices)
-
- # shard=False means that each device will get the full list of tests
- # and then each one will decide their own affinity.
- # shard=True means each device will pop the next test available from a queue,
- # which increases throughput but have no affinity.
- results, _ = test_dispatcher.RunTests(
- tests, runner_factory, devices, shard=False, test_timeout=None,
- num_retries=args.num_retries)
-
- report_results.LogFull(
- results=results,
- test_type='Perf',
- test_package='Perf')
-
- if args.json_results_file:
- json_results.GenerateJsonResultsFile([results], args.json_results_file)
-
- if perf_options.single_step:
- return perf_test_runner.PrintTestOutput('single_step')
-
- perf_test_runner.PrintSummary(tests)
-
- # Always return 0 on the sharding stage. Individual tests exit_code
- # will be returned on the print_step stage.
- return 0
-
-
def _RunPythonTests(args):
"""Subcommand of RunTestsCommand which runs python unit tests."""
suite_vars = constants.PYTHON_UNIT_TEST_SUITES[args.suite_name]
@@ -835,8 +765,6 @@ def RunTestsCommand(args): # pylint: disable=too-many-return-statements
return _RunJUnitTests(args)
elif command == 'monkey':
return _RunMonkeyTests(args, get_devices())
- elif command == 'perf':
- return _RunPerfTests(args, get_devices())
elif command == 'python':
return _RunPythonTests(args)
else:
« no previous file with comments | « build/android/pylib/utils/test_environment.py ('k') | build/android/test_runner.pydeps » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698