| Index: tools/isolate/tests/run_test_cases_smoke_test.py
|
| diff --git a/tools/isolate/tests/run_test_cases_smoke_test.py b/tools/isolate/tests/run_test_cases_smoke_test.py
|
| deleted file mode 100755
|
| index 0adb764224ebd1ea8fc4fd1562ba8480e757f29d..0000000000000000000000000000000000000000
|
| --- a/tools/isolate/tests/run_test_cases_smoke_test.py
|
| +++ /dev/null
|
| @@ -1,172 +0,0 @@
|
| -#!/usr/bin/env python
|
| -# Copyright (c) 2012 The Chromium Authors. All rights reserved.
|
| -# Use of this source code is governed by a BSD-style license that can be
|
| -# found in the LICENSE file.
|
| -
|
| -import json
|
| -import logging
|
| -import os
|
| -import re
|
| -import subprocess
|
| -import sys
|
| -import unittest
|
| -
|
| -ROOT_DIR = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))
|
| -sys.path.insert(0, ROOT_DIR)
|
| -sys.path.append(os.path.join(ROOT_DIR, 'tests', 'gtest_fake'))
|
| -
|
| -import gtest_fake_base
|
| -
|
| -
|
| -def RunTest(test_file, extra_flags):
|
| - target = os.path.join(ROOT_DIR, 'tests', 'gtest_fake', test_file)
|
| - cmd = [
|
| - sys.executable,
|
| - os.path.join(ROOT_DIR, 'run_test_cases.py'),
|
| - ] + extra_flags
|
| -
|
| - cmd.append(target)
|
| - logging.debug(' '.join(cmd))
|
| - proc = subprocess.Popen(
|
| - cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
|
| - # pylint is confused.
|
| - out, err = proc.communicate() or ('', '')
|
| -
|
| - return (out, err, proc.returncode)
|
| -
|
| -
|
| -class TraceTestCases(unittest.TestCase):
|
| - def setUp(self):
|
| - # Make sure there's no environment variable that could do side effects.
|
| - os.environ.pop('GTEST_SHARD_INDEX', '')
|
| - os.environ.pop('GTEST_TOTAL_SHARDS', '')
|
| -
|
| - self.filename = 'test.results'
|
| -
|
| - def tearDown(self):
|
| - if os.path.exists(self.filename):
|
| - os.remove(self.filename)
|
| -
|
| - def _check_results(self, expected_out_re, out, err):
|
| - if sys.platform == 'win32':
|
| - out = out.replace('\r\n', '\n')
|
| - lines = out.splitlines()
|
| -
|
| - for index in range(len(expected_out_re)):
|
| - line = lines.pop(0)
|
| - self.assertTrue(
|
| - re.match('^%s$' % expected_out_re[index], line),
|
| - (index, expected_out_re[index], repr(line)))
|
| - self.assertEqual([], lines)
|
| - self.assertEqual('', err)
|
| -
|
| - def _check_results_file(self, expected_file_contents_entries):
|
| - self.assertTrue(os.path.exists(self.filename))
|
| -
|
| - with open(self.filename) as f:
|
| - file_contents = json.load(f)
|
| -
|
| - self.assertEqual(len(expected_file_contents_entries), len(file_contents))
|
| - for (entry_name, entry_count) in expected_file_contents_entries:
|
| - self.assertTrue(entry_name in file_contents)
|
| - self.assertEqual(entry_count, len(file_contents[entry_name]))
|
| -
|
| - def test_simple_pass(self):
|
| - out, err, return_code = RunTest(
|
| - 'gtest_fake_pass.py', ['--result', self.filename])
|
| -
|
| - self.assertEqual(0, return_code)
|
| -
|
| - expected_out_re = [
|
| - r'\[\d/\d\] \d\.\d\ds .+',
|
| - r'\[\d/\d\] \d\.\d\ds .+',
|
| - r'\[\d/\d\] \d\.\d\ds .+',
|
| - re.escape('Summary:'),
|
| - re.escape('Success: 3 100.00%'),
|
| - re.escape('Flaky: 0 0.00%'),
|
| - re.escape('Fail: 0 0.00%'),
|
| - r'\d+\.\ds Done running 3 tests with 3 executions. \d+\.\d test/s',
|
| - ]
|
| - self._check_results(expected_out_re, out, err)
|
| -
|
| - expected_result_file_entries = [
|
| - ('Foo.Bar1', 1),
|
| - ('Foo.Bar2', 1),
|
| - ('Foo.Bar3', 1)
|
| - ]
|
| - self._check_results_file(expected_result_file_entries)
|
| -
|
| - def test_simple_fail(self):
|
| - out, err, return_code = RunTest(
|
| - 'gtest_fake_fail.py', ['--result', self.filename])
|
| -
|
| - self.assertEqual(1, return_code)
|
| -
|
| - expected_out_re = [
|
| - r'\[\d/\d\] \d\.\d\ds .+',
|
| - r'\[\d/\d\] \d\.\d\ds .+',
|
| - r'\[\d/\d\] \d\.\d\ds .+',
|
| - r'\[\d/\d\] \d\.\d\ds .+',
|
| - r'\[\d/\d\] \d\.\d\ds .+',
|
| - r'\[\d/\d\] \d\.\d\ds .+',
|
| - re.escape('Note: Google Test filter = Baz.Fail'),
|
| - r'',
|
| - ] + [
|
| - re.escape(l) for l in
|
| - gtest_fake_base.get_test_output('Baz.Fail').splitlines()
|
| - ] + [
|
| - '',
|
| - ] + [
|
| - re.escape(l) for l in gtest_fake_base.get_footer(1, 1).splitlines()
|
| - ] + [
|
| - '',
|
| - re.escape('Summary:'),
|
| - re.escape('Baz.Fail failed'),
|
| - re.escape('Success: 3 75.00%'),
|
| - re.escape('Flaky: 0 0.00%'),
|
| - re.escape('Fail: 1 25.00%'),
|
| - r'\d+\.\ds Done running 4 tests with 6 executions. \d+\.\d test/s',
|
| - ]
|
| - self._check_results(expected_out_re, out, err)
|
| -
|
| - expected_result_file_entries = [
|
| - ('Foo.Bar1', 1),
|
| - ('Foo.Bar2', 1),
|
| - ('Foo.Bar3', 1),
|
| - ('Baz.Fail', 3)
|
| - ]
|
| - self._check_results_file(expected_result_file_entries)
|
| -
|
| - def test_simple_gtest_list_error(self):
|
| - out, err, return_code = RunTest(
|
| - 'gtest_fake_error.py', ['--no-dump'])
|
| -
|
| - expected_out_re = [
|
| - 'Failed to run %s %s --gtest_list_tests' % (
|
| - sys.executable,
|
| - os.path.join(ROOT_DIR, 'tests', 'gtest_fake', 'gtest_fake_error.py')),
|
| - 'stdout:',
|
| - '',
|
| - 'stderr:',
|
| - 'Unable to list tests'
|
| - ]
|
| -
|
| - self.assertEqual(1, return_code)
|
| - self._check_results(expected_out_re, out, err)
|
| -
|
| - def test_gtest_list_tests(self):
|
| - out, err, return_code = RunTest(
|
| - 'gtest_fake_fail.py', ['--gtest_list_tests'])
|
| -
|
| - expected_out = (
|
| - 'Foo.\n Bar1\n Bar2\n Bar3\nBaz.\n Fail\n'
|
| - ' YOU HAVE 2 tests with ignored failures (FAILS prefix)\n\n')
|
| - self.assertEqual(0, return_code)
|
| - self.assertEqual(expected_out, out)
|
| - self.assertEqual('', err)
|
| -
|
| -
|
| -if __name__ == '__main__':
|
| - VERBOSE = '-v' in sys.argv
|
| - logging.basicConfig(level=logging.DEBUG if VERBOSE else logging.ERROR)
|
| - unittest.main()
|
|
|