Index: tools/unittests/run_perf_test.py |
diff --git a/tools/unittests/run_benchmarks_test.py b/tools/unittests/run_perf_test.py |
similarity index 91% |
copy from tools/unittests/run_benchmarks_test.py |
copy to tools/unittests/run_perf_test.py |
index d17025232c98356bbc1a69449f914ffb16c705c1..86c81993c7cfd7f8d71849feef243bc4a705787e 100644 |
--- a/tools/unittests/run_benchmarks_test.py |
+++ b/tools/unittests/run_perf_test.py |
@@ -17,7 +17,7 @@ import unittest |
# Requires python-coverage and python-mock. Native python coverage |
# version >= 3.7.1 should be installed to get the best speed. |
-TEST_WORKSPACE = path.join(tempfile.gettempdir(), "test-v8-run-benchmarks") |
+TEST_WORKSPACE = path.join(tempfile.gettempdir(), "test-v8-run-perf") |
V8_JSON = { |
"path": ["."], |
@@ -26,7 +26,7 @@ V8_JSON = { |
"main": "run.js", |
"run_count": 1, |
"results_regexp": "^%s: (.+)$", |
- "benchmarks": [ |
+ "tests": [ |
{"name": "Richards"}, |
{"name": "DeltaBlue"}, |
] |
@@ -37,7 +37,7 @@ V8_NESTED_SUITES_JSON = { |
"flags": ["--flag"], |
"run_count": 1, |
"units": "score", |
- "benchmarks": [ |
+ "tests": [ |
{"name": "Richards", |
"path": ["richards"], |
"binary": "d7", |
@@ -47,7 +47,7 @@ V8_NESTED_SUITES_JSON = { |
"results_regexp": "^Richards: (.+)$"}, |
{"name": "Sub", |
"path": ["sub"], |
- "benchmarks": [ |
+ "tests": [ |
{"name": "Leaf", |
"path": ["leaf"], |
"run_count_x64": 3, |
@@ -79,18 +79,18 @@ V8_GENERIC_JSON = { |
Output = namedtuple("Output", "stdout, stderr") |
-class BenchmarksTest(unittest.TestCase): |
+class PerfTest(unittest.TestCase): |
@classmethod |
def setUpClass(cls): |
cls.base = path.dirname(path.dirname(path.abspath(__file__))) |
sys.path.append(cls.base) |
cls._cov = coverage.coverage( |
- include=([os.path.join(cls.base, "run_benchmarks.py")])) |
+ include=([os.path.join(cls.base, "run_perf.py")])) |
cls._cov.start() |
- import run_benchmarks |
+ import run_perf |
from testrunner.local import commands |
global commands |
- global run_benchmarks |
+ global run_perf |
@classmethod |
def tearDownClass(cls): |
@@ -114,13 +114,13 @@ class BenchmarksTest(unittest.TestCase): |
f.write(json.dumps(json_content)) |
def _MockCommand(self, *args): |
- # Fake output for each benchmark run. |
- benchmark_outputs = [Output(stdout=arg, stderr=None) for arg in args[1]] |
+ # Fake output for each test run. |
+ test_outputs = [Output(stdout=arg, stderr=None) for arg in args[1]] |
def execute(*args, **kwargs): |
- return benchmark_outputs.pop() |
+ return test_outputs.pop() |
commands.Execute = MagicMock(side_effect=execute) |
- # Check that d8 is called from the correct cwd for each benchmark run. |
+ # Check that d8 is called from the correct cwd for each test run. |
dirs = [path.join(TEST_WORKSPACE, arg) for arg in args[0]] |
def chdir(*args, **kwargs): |
self.assertEquals(dirs.pop(), args[0]) |
@@ -134,7 +134,7 @@ class BenchmarksTest(unittest.TestCase): |
self._test_input, |
] |
all_args += args |
- return run_benchmarks.Main(all_args) |
+ return run_perf.Main(all_args) |
def _LoadResults(self): |
with open(self._test_output) as f: |
@@ -196,8 +196,8 @@ class BenchmarksTest(unittest.TestCase): |
test_input = dict(V8_JSON) |
test_input["run_count"] = 2 |
del test_input["results_regexp"] |
- test_input["benchmarks"][0]["results_regexp"] = "^Richards: (.+)$" |
- test_input["benchmarks"][1]["results_regexp"] = "^DeltaBlue: (.+)$" |
+ test_input["tests"][0]["results_regexp"] = "^Richards: (.+)$" |
+ test_input["tests"][1]["results_regexp"] = "^DeltaBlue: (.+)$" |
self._WriteTestInput(test_input) |
self._MockCommand([".", "."], |
["Richards: 100\nDeltaBlue: 200\n", |
@@ -274,11 +274,11 @@ class BenchmarksTest(unittest.TestCase): |
{"name": "DeltaBlue", "results": ["5", "6"], "stddev": "0.8"}, |
]) |
self._VerifyErrors( |
- ["Benchmark Richards should only run once since a stddev is provided " |
- "by the benchmark.", |
- "Benchmark DeltaBlue should only run once since a stddev is provided " |
- "by the benchmark.", |
- "Regexp \"^DeltaBlue\-stddev: (.+)$\" didn't match for benchmark " |
+ ["Test Richards should only run once since a stddev is provided " |
+ "by the test.", |
+ "Test DeltaBlue should only run once since a stddev is provided " |
+ "by the test.", |
+ "Regexp \"^DeltaBlue\-stddev: (.+)$\" didn't match for test " |
"DeltaBlue."]) |
self._VerifyMock(path.join("out", "x64.release", "d7"), "--flag", "run.js") |
@@ -318,7 +318,7 @@ class BenchmarksTest(unittest.TestCase): |
{"name": "DeltaBlue", "results": ["10657567"], "stddev": ""}, |
]) |
self._VerifyErrors( |
- ["Regexp \"^Richards: (.+)$\" didn't match for benchmark Richards.", |
+ ["Regexp \"^Richards: (.+)$\" didn't match for test Richards.", |
"Not all traces have the same number of results."]) |
self._VerifyMock(path.join("out", "Release", "d7"), "--flag", "run.js") |
@@ -331,7 +331,7 @@ class BenchmarksTest(unittest.TestCase): |
{"name": "DeltaBlue", "results": ["10657567"], "stddev": ""}, |
]) |
self._VerifyErrors( |
- ["Regexp \"^Richards: (.+)$\" didn't match for benchmark Richards."]) |
+ ["Regexp \"^Richards: (.+)$\" didn't match for test Richards."]) |
self._VerifyMock(path.join("out", "x64.release", "d7"), "--flag", "run.js") |
def testOneRunGeneric(self): |