| Index: build/android/pylib/gtest/test_runner.py
|
| diff --git a/build/android/pylib/gtest/test_runner.py b/build/android/pylib/gtest/test_runner.py
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..a48f18a70f669cc27aae4d62f9f577c7081807ae
|
| --- /dev/null
|
| +++ b/build/android/pylib/gtest/test_runner.py
|
| @@ -0,0 +1,217 @@
|
| +# Copyright (c) 2012 The Chromium Authors. All rights reserved.
|
| +# Use of this source code is governed by a BSD-style license that can be
|
| +# found in the LICENSE file.
|
| +
|
| +import logging
|
| +import os
|
| +import re
|
| +import tempfile
|
| +
|
| +from pylib import pexpect
|
| +from pylib import ports
|
| +from pylib.base import base_test_result
|
| +from pylib.base import base_test_runner
|
| +from pylib.device import device_errors
|
| +from pylib.gtest import gtest_test_instance
|
| +from pylib.local import local_test_server_spawner
|
| +from pylib.perf import perf_control
|
| +
|
| +# Test case statuses.
|
| +RE_RUN = re.compile('\\[ RUN \\] ?(.*)\r\n')
|
| +RE_FAIL = re.compile('\\[ FAILED \\] ?(.*?)( \\((\\d+) ms\\))?\r\r\n')
|
| +RE_OK = re.compile('\\[ OK \\] ?(.*?)( \\((\\d+) ms\\))?\r\r\n')
|
| +
|
| +# Test run statuses.
|
| +RE_PASSED = re.compile('\\[ PASSED \\] ?(.*)\r\n')
|
| +RE_RUNNER_FAIL = re.compile('\\[ RUNNER_FAILED \\] ?(.*)\r\n')
|
| +# Signal handlers are installed before starting tests
|
| +# to output the CRASHED marker when a crash happens.
|
| +RE_CRASH = re.compile('\\[ CRASHED \\](.*)\r\n')
|
| +
|
| +# Bots that don't output anything for 20 minutes get timed out, so that's our
|
| +# hard cap.
|
| +_INFRA_STDOUT_TIMEOUT = 20 * 60
|
| +
|
| +
|
| +def _TestSuiteRequiresMockTestServer(suite_name):
|
| + """Returns True if the test suite requires mock test server."""
|
| + tests_require_net_test_server = ['unit_tests', 'net_unittests',
|
| + 'components_browsertests',
|
| + 'content_unittests',
|
| + 'content_browsertests']
|
| + return (suite_name in
|
| + tests_require_net_test_server)
|
| +
|
| +def _TestSuiteRequiresHighPerfMode(suite_name):
|
| + """Returns True if the test suite requires high performance mode."""
|
| + return 'perftests' in suite_name
|
| +
|
| +class TestRunner(base_test_runner.BaseTestRunner):
|
| + def __init__(self, test_options, device, test_package):
|
| + """Single test suite attached to a single device.
|
| +
|
| + Args:
|
| + test_options: A GTestOptions object.
|
| + device: Device to run the tests.
|
| + test_package: An instance of TestPackage class.
|
| + """
|
| +
|
| + super(TestRunner, self).__init__(device, test_options.tool)
|
| +
|
| + self.test_package = test_package
|
| + self.test_package.tool = self.tool
|
| + self._test_arguments = test_options.test_arguments
|
| +
|
| + timeout = test_options.timeout
|
| + if timeout == 0:
|
| + timeout = 60
|
| + # On a VM (e.g. chromium buildbots), this timeout is way too small.
|
| + if os.environ.get('BUILDBOT_SLAVENAME'):
|
| + timeout = timeout * 2
|
| +
|
| + self._timeout = min(timeout * self.tool.GetTimeoutScale(),
|
| + _INFRA_STDOUT_TIMEOUT)
|
| + if _TestSuiteRequiresHighPerfMode(self.test_package.suite_name):
|
| + self._perf_controller = perf_control.PerfControl(self.device)
|
| +
|
| + if _TestSuiteRequiresMockTestServer(self.test_package.suite_name):
|
| + self._servers = [
|
| + local_test_server_spawner.LocalTestServerSpawner(
|
| + ports.AllocateTestServerPort(), self.device, self.tool)]
|
| + else:
|
| + self._servers = []
|
| +
|
| + if test_options.app_data_files:
|
| + self._app_data_files = test_options.app_data_files
|
| + if test_options.app_data_file_dir:
|
| + self._app_data_file_dir = test_options.app_data_file_dir
|
| + else:
|
| + self._app_data_file_dir = tempfile.mkdtemp()
|
| + logging.critical('Saving app files to %s', self._app_data_file_dir)
|
| + else:
|
| + self._app_data_files = None
|
| + self._app_data_file_dir = None
|
| +
|
| + #override
|
| + def InstallTestPackage(self):
|
| + self.test_package.Install(self.device)
|
| +
|
| + def _ParseTestOutput(self, p):
|
| + """Process the test output.
|
| +
|
| + Args:
|
| + p: An instance of pexpect spawn class.
|
| +
|
| + Returns:
|
| + A TestRunResults object.
|
| + """
|
| + results = base_test_result.TestRunResults()
|
| +
|
| + log = ''
|
| + try:
|
| + while True:
|
| + full_test_name = None
|
| +
|
| + found = p.expect([RE_RUN, RE_PASSED, RE_RUNNER_FAIL],
|
| + timeout=self._timeout)
|
| + if found == 1: # RE_PASSED
|
| + break
|
| + elif found == 2: # RE_RUNNER_FAIL
|
| + break
|
| + else: # RE_RUN
|
| + full_test_name = p.match.group(1).replace('\r', '')
|
| + found = p.expect([RE_OK, RE_FAIL, RE_CRASH], timeout=self._timeout)
|
| + log = p.before.replace('\r', '')
|
| + if found == 0: # RE_OK
|
| + if full_test_name == p.match.group(1).replace('\r', ''):
|
| + duration_ms = int(p.match.group(3)) if p.match.group(3) else 0
|
| + results.AddResult(base_test_result.BaseTestResult(
|
| + full_test_name, base_test_result.ResultType.PASS,
|
| + duration=duration_ms, log=log))
|
| + elif found == 2: # RE_CRASH
|
| + results.AddResult(base_test_result.BaseTestResult(
|
| + full_test_name, base_test_result.ResultType.CRASH,
|
| + log=log))
|
| + break
|
| + else: # RE_FAIL
|
| + duration_ms = int(p.match.group(3)) if p.match.group(3) else 0
|
| + results.AddResult(base_test_result.BaseTestResult(
|
| + full_test_name, base_test_result.ResultType.FAIL,
|
| + duration=duration_ms, log=log))
|
| + except pexpect.EOF:
|
| + logging.error('Test terminated - EOF')
|
| + # We're here because either the device went offline, or the test harness
|
| + # crashed without outputting the CRASHED marker (crbug.com/175538).
|
| + if not self.device.IsOnline():
|
| + raise device_errors.DeviceUnreachableError(
|
| + 'Device %s went offline.' % str(self.device))
|
| + if full_test_name:
|
| + results.AddResult(base_test_result.BaseTestResult(
|
| + full_test_name, base_test_result.ResultType.CRASH,
|
| + log=p.before.replace('\r', '')))
|
| + except pexpect.TIMEOUT:
|
| + logging.error('Test terminated after %d second timeout.',
|
| + self._timeout)
|
| + if full_test_name:
|
| + results.AddResult(base_test_result.BaseTestResult(
|
| + full_test_name, base_test_result.ResultType.TIMEOUT,
|
| + log=p.before.replace('\r', '')))
|
| + finally:
|
| + p.close()
|
| +
|
| + ret_code = self.test_package.GetGTestReturnCode(self.device)
|
| + if ret_code:
|
| + logging.critical(
|
| + 'gtest exit code: %d\npexpect.before: %s\npexpect.after: %s',
|
| + ret_code, p.before, p.after)
|
| +
|
| + return results
|
| +
|
| + #override
|
| + def RunTest(self, test):
|
| + test_results = base_test_result.TestRunResults()
|
| + if not test:
|
| + return test_results, None
|
| +
|
| + try:
|
| + self.test_package.ClearApplicationState(self.device)
|
| + self.test_package.CreateCommandLineFileOnDevice(
|
| + self.device, test, self._test_arguments)
|
| + test_results = self._ParseTestOutput(
|
| + self.test_package.SpawnTestProcess(self.device))
|
| + if self._app_data_files:
|
| + self.test_package.PullAppFiles(self.device, self._app_data_files,
|
| + self._app_data_file_dir)
|
| + finally:
|
| + for s in self._servers:
|
| + s.Reset()
|
| + # Calculate unknown test results.
|
| + all_tests = set(test.split(':'))
|
| + all_tests_ran = set([t.GetName() for t in test_results.GetAll()])
|
| + unknown_tests = all_tests - all_tests_ran
|
| + test_results.AddResults(
|
| + [base_test_result.BaseTestResult(t, base_test_result.ResultType.UNKNOWN)
|
| + for t in unknown_tests])
|
| + retry = ':'.join([t.GetName() for t in test_results.GetNotPass()])
|
| + return test_results, retry
|
| +
|
| + #override
|
| + def SetUp(self):
|
| + """Sets up necessary test enviroment for the test suite."""
|
| + super(TestRunner, self).SetUp()
|
| + for s in self._servers:
|
| + s.SetUp()
|
| + if _TestSuiteRequiresHighPerfMode(self.test_package.suite_name):
|
| + self._perf_controller.SetHighPerfMode()
|
| + self.tool.SetupEnvironment()
|
| +
|
| + #override
|
| + def TearDown(self):
|
| + """Cleans up the test enviroment for the test suite."""
|
| + for s in self._servers:
|
| + s.TearDown()
|
| + if _TestSuiteRequiresHighPerfMode(self.test_package.suite_name):
|
| + self._perf_controller.SetDefaultPerfMode()
|
| + self.test_package.ClearApplicationState(self.device)
|
| + self.tool.CleanUpEnvironment()
|
| + super(TestRunner, self).TearDown()
|
|
|