| Index: tools/unittests/run_perf_test.py | 
| diff --git a/tools/unittests/run_benchmarks_test.py b/tools/unittests/run_perf_test.py | 
| similarity index 91% | 
| copy from tools/unittests/run_benchmarks_test.py | 
| copy to tools/unittests/run_perf_test.py | 
| index d17025232c98356bbc1a69449f914ffb16c705c1..86c81993c7cfd7f8d71849feef243bc4a705787e 100644 | 
| --- a/tools/unittests/run_benchmarks_test.py | 
| +++ b/tools/unittests/run_perf_test.py | 
| @@ -17,7 +17,7 @@ import unittest | 
| # Requires python-coverage and python-mock. Native python coverage | 
| # version >= 3.7.1 should be installed to get the best speed. | 
|  | 
| -TEST_WORKSPACE = path.join(tempfile.gettempdir(), "test-v8-run-benchmarks") | 
| +TEST_WORKSPACE = path.join(tempfile.gettempdir(), "test-v8-run-perf") | 
|  | 
| V8_JSON = { | 
| "path": ["."], | 
| @@ -26,7 +26,7 @@ V8_JSON = { | 
| "main": "run.js", | 
| "run_count": 1, | 
| "results_regexp": "^%s: (.+)$", | 
| -  "benchmarks": [ | 
| +  "tests": [ | 
| {"name": "Richards"}, | 
| {"name": "DeltaBlue"}, | 
| ] | 
| @@ -37,7 +37,7 @@ V8_NESTED_SUITES_JSON = { | 
| "flags": ["--flag"], | 
| "run_count": 1, | 
| "units": "score", | 
| -  "benchmarks": [ | 
| +  "tests": [ | 
| {"name": "Richards", | 
| "path": ["richards"], | 
| "binary": "d7", | 
| @@ -47,7 +47,7 @@ V8_NESTED_SUITES_JSON = { | 
| "results_regexp": "^Richards: (.+)$"}, | 
| {"name": "Sub", | 
| "path": ["sub"], | 
| -     "benchmarks": [ | 
| +     "tests": [ | 
| {"name": "Leaf", | 
| "path": ["leaf"], | 
| "run_count_x64": 3, | 
| @@ -79,18 +79,18 @@ V8_GENERIC_JSON = { | 
|  | 
| Output = namedtuple("Output", "stdout, stderr") | 
|  | 
| -class BenchmarksTest(unittest.TestCase): | 
| +class PerfTest(unittest.TestCase): | 
| @classmethod | 
| def setUpClass(cls): | 
| cls.base = path.dirname(path.dirname(path.abspath(__file__))) | 
| sys.path.append(cls.base) | 
| cls._cov = coverage.coverage( | 
| -        include=([os.path.join(cls.base, "run_benchmarks.py")])) | 
| +        include=([os.path.join(cls.base, "run_perf.py")])) | 
| cls._cov.start() | 
| -    import run_benchmarks | 
| +    import run_perf | 
| from testrunner.local import commands | 
| global commands | 
| -    global run_benchmarks | 
| +    global run_perf | 
|  | 
| @classmethod | 
| def tearDownClass(cls): | 
| @@ -114,13 +114,13 @@ class BenchmarksTest(unittest.TestCase): | 
| f.write(json.dumps(json_content)) | 
|  | 
| def _MockCommand(self, *args): | 
| -    # Fake output for each benchmark run. | 
| -    benchmark_outputs = [Output(stdout=arg, stderr=None) for arg in args[1]] | 
| +    # Fake output for each test run. | 
| +    test_outputs = [Output(stdout=arg, stderr=None) for arg in args[1]] | 
| def execute(*args, **kwargs): | 
| -      return benchmark_outputs.pop() | 
| +      return test_outputs.pop() | 
| commands.Execute = MagicMock(side_effect=execute) | 
|  | 
| -    # Check that d8 is called from the correct cwd for each benchmark run. | 
| +    # Check that d8 is called from the correct cwd for each test run. | 
| dirs = [path.join(TEST_WORKSPACE, arg) for arg in args[0]] | 
| def chdir(*args, **kwargs): | 
| self.assertEquals(dirs.pop(), args[0]) | 
| @@ -134,7 +134,7 @@ class BenchmarksTest(unittest.TestCase): | 
| self._test_input, | 
| ] | 
| all_args += args | 
| -    return run_benchmarks.Main(all_args) | 
| +    return run_perf.Main(all_args) | 
|  | 
| def _LoadResults(self): | 
| with open(self._test_output) as f: | 
| @@ -196,8 +196,8 @@ class BenchmarksTest(unittest.TestCase): | 
| test_input = dict(V8_JSON) | 
| test_input["run_count"] = 2 | 
| del test_input["results_regexp"] | 
| -    test_input["benchmarks"][0]["results_regexp"] = "^Richards: (.+)$" | 
| -    test_input["benchmarks"][1]["results_regexp"] = "^DeltaBlue: (.+)$" | 
| +    test_input["tests"][0]["results_regexp"] = "^Richards: (.+)$" | 
| +    test_input["tests"][1]["results_regexp"] = "^DeltaBlue: (.+)$" | 
| self._WriteTestInput(test_input) | 
| self._MockCommand([".", "."], | 
| ["Richards: 100\nDeltaBlue: 200\n", | 
| @@ -274,11 +274,11 @@ class BenchmarksTest(unittest.TestCase): | 
| {"name": "DeltaBlue", "results": ["5", "6"], "stddev": "0.8"}, | 
| ]) | 
| self._VerifyErrors( | 
| -        ["Benchmark Richards should only run once since a stddev is provided " | 
| -         "by the benchmark.", | 
| -         "Benchmark DeltaBlue should only run once since a stddev is provided " | 
| -         "by the benchmark.", | 
| -         "Regexp \"^DeltaBlue\-stddev: (.+)$\" didn't match for benchmark " | 
| +        ["Test Richards should only run once since a stddev is provided " | 
| +         "by the test.", | 
| +         "Test DeltaBlue should only run once since a stddev is provided " | 
| +         "by the test.", | 
| +         "Regexp \"^DeltaBlue\-stddev: (.+)$\" didn't match for test " | 
| "DeltaBlue."]) | 
| self._VerifyMock(path.join("out", "x64.release", "d7"), "--flag", "run.js") | 
|  | 
| @@ -318,7 +318,7 @@ class BenchmarksTest(unittest.TestCase): | 
| {"name": "DeltaBlue", "results": ["10657567"], "stddev": ""}, | 
| ]) | 
| self._VerifyErrors( | 
| -        ["Regexp \"^Richards: (.+)$\" didn't match for benchmark Richards.", | 
| +        ["Regexp \"^Richards: (.+)$\" didn't match for test Richards.", | 
| "Not all traces have the same number of results."]) | 
| self._VerifyMock(path.join("out", "Release", "d7"), "--flag", "run.js") | 
|  | 
| @@ -331,7 +331,7 @@ class BenchmarksTest(unittest.TestCase): | 
| {"name": "DeltaBlue", "results": ["10657567"], "stddev": ""}, | 
| ]) | 
| self._VerifyErrors( | 
| -        ["Regexp \"^Richards: (.+)$\" didn't match for benchmark Richards."]) | 
| +        ["Regexp \"^Richards: (.+)$\" didn't match for test Richards."]) | 
| self._VerifyMock(path.join("out", "x64.release", "d7"), "--flag", "run.js") | 
|  | 
| def testOneRunGeneric(self): | 
|  |