| Index: build/android/pylib/local/device/local_device_test_run.py
|
| diff --git a/build/android/pylib/local/device/local_device_test_run.py b/build/android/pylib/local/device/local_device_test_run.py
|
| deleted file mode 100644
|
| index fa24eb10e26df4eb3a537d457d23c0146da32283..0000000000000000000000000000000000000000
|
| --- a/build/android/pylib/local/device/local_device_test_run.py
|
| +++ /dev/null
|
| @@ -1,99 +0,0 @@
|
| -# Copyright 2014 The Chromium Authors. All rights reserved.
|
| -# Use of this source code is governed by a BSD-style license that can be
|
| -# found in the LICENSE file.
|
| -
|
| -import logging
|
| -
|
| -from pylib import valgrind_tools
|
| -from pylib.base import base_test_result
|
| -from pylib.base import test_run
|
| -from pylib.base import test_collection
|
| -
|
| -
|
| -class LocalDeviceTestRun(test_run.TestRun):
|
| -
|
| - def __init__(self, env, test_instance):
|
| - super(LocalDeviceTestRun, self).__init__(env, test_instance)
|
| - self._tools = {}
|
| -
|
| - #override
|
| - def RunTests(self):
|
| - tests = self._GetTests()
|
| -
|
| - def run_tests_on_device(dev, tests):
|
| - r = base_test_result.TestRunResults()
|
| - for test in tests:
|
| - result = self._RunTest(dev, test)
|
| - if isinstance(result, base_test_result.BaseTestResult):
|
| - r.AddResult(result)
|
| - elif isinstance(result, list):
|
| - r.AddResults(result)
|
| - else:
|
| - raise Exception('Unexpected result type: %s' % type(result).__name__)
|
| - if isinstance(tests, test_collection.TestCollection):
|
| - tests.test_completed()
|
| - return r
|
| -
|
| - tries = 0
|
| - results = base_test_result.TestRunResults()
|
| - all_fail_results = {}
|
| - while tries < self._env.max_tries and tests:
|
| - logging.debug('try %d, will run %d tests:', tries, len(tests))
|
| - for t in tests:
|
| - logging.debug(' %s', t)
|
| -
|
| - if self._ShouldShard():
|
| - tc = test_collection.TestCollection(self._CreateShards(tests))
|
| - try_results = self._env.parallel_devices.pMap(
|
| - run_tests_on_device, tc).pGet(None)
|
| - else:
|
| - try_results = self._env.parallel_devices.pMap(
|
| - run_tests_on_device, tests).pGet(None)
|
| - for try_result in try_results:
|
| - for result in try_result.GetAll():
|
| - if result.GetType() in (base_test_result.ResultType.PASS,
|
| - base_test_result.ResultType.SKIP):
|
| - results.AddResult(result)
|
| - else:
|
| - all_fail_results[result.GetName()] = result
|
| -
|
| - results_names = set(r.GetName() for r in results.GetAll())
|
| - tests = [t for t in tests if self._GetTestName(t) not in results_names]
|
| - tries += 1
|
| -
|
| - all_unknown_test_names = set(self._GetTestName(t) for t in tests)
|
| - all_failed_test_names = set(all_fail_results.iterkeys())
|
| -
|
| - unknown_tests = all_unknown_test_names.difference(all_failed_test_names)
|
| - failed_tests = all_failed_test_names.intersection(all_unknown_test_names)
|
| -
|
| - if unknown_tests:
|
| - results.AddResults(
|
| - base_test_result.BaseTestResult(
|
| - u, base_test_result.ResultType.UNKNOWN)
|
| - for u in unknown_tests)
|
| - if failed_tests:
|
| - results.AddResults(all_fail_results[f] for f in failed_tests)
|
| -
|
| - return results
|
| -
|
| - def GetTool(self, device):
|
| - if not str(device) in self._tools:
|
| - self._tools[str(device)] = valgrind_tools.CreateTool(
|
| - self._env.tool, device)
|
| - return self._tools[str(device)]
|
| -
|
| - def _CreateShards(self, tests):
|
| - raise NotImplementedError
|
| -
|
| - def _GetTestName(self, test):
|
| - return test
|
| -
|
| - def _GetTests(self):
|
| - raise NotImplementedError
|
| -
|
| - def _RunTest(self, device, test):
|
| - raise NotImplementedError
|
| -
|
| - def _ShouldShard(self):
|
| - raise NotImplementedError
|
|
|