Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(91)

Unified Diff: third_party/WebKit/Tools/Scripts/webkitpy/performance_tests/perftestsrunner_unittest.py

Issue 2136793002: Remove all unused variables. (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@master
Patch Set: Created 4 years, 5 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: third_party/WebKit/Tools/Scripts/webkitpy/performance_tests/perftestsrunner_unittest.py
diff --git a/third_party/WebKit/Tools/Scripts/webkitpy/performance_tests/perftestsrunner_unittest.py b/third_party/WebKit/Tools/Scripts/webkitpy/performance_tests/perftestsrunner_unittest.py
index 23c1346d1f80021b76847052717e65b200ca36a7..2047e2bcac2936428e450fc412e37622c4d5cdc9 100644
--- a/third_party/WebKit/Tools/Scripts/webkitpy/performance_tests/perftestsrunner_unittest.py
+++ b/third_party/WebKit/Tools/Scripts/webkitpy/performance_tests/perftestsrunner_unittest.py
@@ -47,7 +47,7 @@ from webkitpy.performance_tests.perftestsrunner import PerfTestsRunner
class MainTest(unittest.TestCase):
def create_runner(self, args=[]):
- options, parsed_args = PerfTestsRunner._parse_args(args)
+ options, _ = PerfTestsRunner._parse_args(args)
test_port = TestPort(host=MockHost(), options=options)
runner = PerfTestsRunner(args=args, port=test_port)
runner._host.filesystem.maybe_make_directory(runner._base_path, 'inspector')
@@ -61,7 +61,7 @@ class MainTest(unittest.TestCase):
runner._host.filesystem.files[runner._host.filesystem.join(dirname, filename)] = content
def test_collect_tests(self):
- runner, port = self.create_runner()
+ runner, _ = self.create_runner()
self._add_file(runner, 'inspector', 'a_file.html', 'a content')
tests = runner._collect_tests()
self.assertEqual(len(tests), 1)
@@ -118,8 +118,7 @@ class MainTest(unittest.TestCase):
'inspector/test1.html', 'inspector/test2.html', 'inspector/unsupported_test1.html', 'unsupported/unsupported_test2.html'])
def test_default_args(self):
- runner, port = self.create_runner()
- options, args = PerfTestsRunner._parse_args([])
+ options, _ = PerfTestsRunner._parse_args([])
self.assertTrue(options.build)
self.assertEqual(options.time_out_ms, 600 * 1000)
self.assertTrue(options.generate_results)
@@ -129,8 +128,7 @@ class MainTest(unittest.TestCase):
self.assertEqual(options.test_runner_count, DEFAULT_TEST_RUNNER_COUNT)
def test_parse_args(self):
- runner, port = self.create_runner()
- options, args = PerfTestsRunner._parse_args([
+ options, _ = PerfTestsRunner._parse_args([
'--build-directory=folder42',
'--platform=platform42',
'--builder-name', 'webkit-mac-1',
@@ -385,7 +383,7 @@ class IntegrationTest(unittest.TestCase):
return json.loads(re.sub(r'("stdev":\s*\d+\.\d{5})\d+', r'\1', json_content))
def create_runner(self, args=[], driver_class=TestDriver):
- options, parsed_args = PerfTestsRunner._parse_args(args)
+ options, _ = PerfTestsRunner._parse_args(args)
test_port = TestPort(host=MockHost(), options=options)
test_port.create_driver = lambda worker_number=None, no_timeout=False: driver_class()
@@ -424,7 +422,6 @@ class IntegrationTest(unittest.TestCase):
tests = []
for test in test_names:
path = filesystem.join(runner._base_path, test)
- dirname = filesystem.dirname(path)
if test.startswith('inspector/'):
tests.append(ChromiumStylePerfTest(runner._port, test, path))
else:
@@ -432,7 +429,7 @@ class IntegrationTest(unittest.TestCase):
return tests
def test_run_test_set(self):
- runner, port = self.create_runner()
+ runner, _ = self.create_runner()
tests = self._tests_for_runner(runner, ['inspector/pass.html', 'inspector/silent.html', 'inspector/failed.html',
'inspector/tonguey.html', 'inspector/timeout.html', 'inspector/crash.html'])
output = OutputCapture()
@@ -440,7 +437,7 @@ class IntegrationTest(unittest.TestCase):
try:
unexpected_result_count = runner._run_tests_set(tests)
finally:
- stdout, stderr, log = output.restore_output()
+ _, _, log = output.restore_output()
self.assertEqual(unexpected_result_count, len(tests) - 1)
self.assertTrue('\nRESULT group_name: test_name= 42 ms\n' in log)
@@ -452,23 +449,23 @@ class IntegrationTest(unittest.TestCase):
def stop(self):
TestDriverWithStopCount.stop_count += 1
- runner, port = self.create_runner(driver_class=TestDriverWithStopCount)
+ runner, _ = self.create_runner(driver_class=TestDriverWithStopCount)
tests = self._tests_for_runner(runner, ['inspector/pass.html', 'inspector/silent.html', 'inspector/failed.html',
'inspector/tonguey.html', 'inspector/timeout.html', 'inspector/crash.html'])
- unexpected_result_count = runner._run_tests_set(tests)
+ runner._run_tests_set(tests)
self.assertEqual(TestDriverWithStopCount.stop_count, 6)
def test_run_test_set_for_parser_tests(self):
- runner, port = self.create_runner()
+ runner, _ = self.create_runner()
tests = self._tests_for_runner(runner, ['Bindings/event-target-wrapper.html', 'Parser/some-parser.html'])
output = OutputCapture()
output.capture_output()
try:
unexpected_result_count = runner._run_tests_set(tests)
finally:
- stdout, stderr, log = output.restore_output()
+ _, _, log = output.restore_output()
self.assertEqual(unexpected_result_count, 0)
self.assertEqual(self._normalize_output(log), EventTargetWrapperTestData.output + SomeParserTestData.output)
@@ -482,7 +479,7 @@ class IntegrationTest(unittest.TestCase):
try:
unexpected_result_count = runner.run()
finally:
- stdout, stderr, log = output.restore_output()
+ _, _, log = output.restore_output()
self.assertEqual(unexpected_result_count, 0)
self.assertEqual(self._normalize_output(log), MemoryTestData.output + '\nMOCK: user.open_url: file://...\n')
parser_tests = self._load_output_json(runner)[0]['tests']['Parser']['tests']
@@ -513,7 +510,7 @@ class IntegrationTest(unittest.TestCase):
try:
self.assertEqual(runner.run(), expected_exit_code)
finally:
- stdout, stderr, logs = output_capture.restore_output()
+ _, _, logs = output_capture.restore_output()
if not expected_exit_code and compare_logs:
expected_logs = ''

Powered by Google App Engine
This is Rietveld 408576698