Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(2948)

Unified Diff: Tools/Scripts/webkitpy/layout_tests/controllers/single_test_runner.py

Issue 546133003: Reformat webkitpy.layout_tests w/ format-webkitpy. (Closed) Base URL: svn://svn.chromium.org/blink/trunk
Patch Set: Created 6 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: Tools/Scripts/webkitpy/layout_tests/controllers/single_test_runner.py
diff --git a/Tools/Scripts/webkitpy/layout_tests/controllers/single_test_runner.py b/Tools/Scripts/webkitpy/layout_tests/controllers/single_test_runner.py
index 382bfccdb25e0af3258f33666096047fda71981b..a4207fe66262eebfcac0326ba18cf40c7bc84302 100644
--- a/Tools/Scripts/webkitpy/layout_tests/controllers/single_test_runner.py
+++ b/Tools/Scripts/webkitpy/layout_tests/controllers/single_test_runner.py
@@ -48,7 +48,7 @@ def run_single_test(port, options, results_directory, worker_name, driver, test_
try:
return runner.run()
except DeviceFailure as e:
- _log.error("device failed: %s", str(e))
+ _log.error('device failed: %s', str(e))
return TestResult(test_input.test_name, device_failed=True)
@@ -78,13 +78,13 @@ class SingleTestRunner(object):
expected_filename = self._port.expected_filename(self._test_name, suffix)
if self._filesystem.exists(expected_filename):
_log.error('%s is a reftest, but has an unused expectation file. Please remove %s.',
- self._test_name, expected_filename)
+ self._test_name, expected_filename)
def _expected_driver_output(self):
return DriverOutput(self._port.expected_text(self._test_name),
- self._port.expected_image(self._test_name),
- self._port.expected_checksum(self._test_name),
- self._port.expected_audio(self._test_name))
+ self._port.expected_image(self._test_name),
+ self._port.expected_checksum(self._test_name),
+ self._port.expected_audio(self._test_name))
def _should_fetch_expected_checksum(self):
return self._should_run_pixel_test and not (self._options.new_baseline or self._options.reset_results)
@@ -143,13 +143,27 @@ class SingleTestRunner(object):
test_result = self._compare_output(expected_driver_output, driver_output)
if self._should_add_missing_baselines:
self._add_missing_baselines(test_result, driver_output)
- test_result_writer.write_test_result(self._filesystem, self._port, self._results_directory, self._test_name, driver_output, expected_driver_output, test_result.failures)
+ test_result_writer.write_test_result(
+ self._filesystem,
+ self._port,
+ self._results_directory,
+ self._test_name,
+ driver_output,
+ expected_driver_output,
+ test_result.failures)
return test_result
def _run_rebaseline(self):
driver_output = self._driver.run_test(self._driver_input(), self._stop_when_done)
failures = self._handle_error(driver_output)
- test_result_writer.write_test_result(self._filesystem, self._port, self._results_directory, self._test_name, driver_output, None, failures)
+ test_result_writer.write_test_result(
+ self._filesystem,
+ self._port,
+ self._results_directory,
+ self._test_name,
+ driver_output,
+ None,
+ failures)
# FIXME: It the test crashed or timed out, it might be better to avoid
# to write new baselines.
self._overwrite_baselines(driver_output)
@@ -159,7 +173,9 @@ class SingleTestRunner(object):
_render_tree_dump_pattern = re.compile(r"^layer at \(\d+,\d+\) size \d+x\d+\n")
def _add_missing_baselines(self, test_result, driver_output):
- missingImage = test_result.has_failure_matching_types(test_failures.FailureMissingImage, test_failures.FailureMissingImageHash)
+ missingImage = test_result.has_failure_matching_types(
+ test_failures.FailureMissingImage,
+ test_failures.FailureMissingImageHash)
if test_result.has_failure_matching_types(test_failures.FailureMissingResult):
self._save_baseline_data(driver_output.text, '.txt', self._location_for_new_baseline(driver_output.text, '.txt'))
if test_result.has_failure_matching_types(test_failures.FailureMissingAudio):
@@ -202,7 +218,7 @@ class SingleTestRunner(object):
raise AssertionError('unrecognized baseline location: %s' % location)
fs.maybe_make_directory(output_dir)
- output_basename = fs.basename(fs.splitext(self._test_name)[0] + "-expected" + extension)
+ output_basename = fs.basename(fs.splitext(self._test_name)[0] + '-expected' + extension)
output_path = fs.join(output_dir, output_basename)
_log.info('Writing new expected result "%s"' % port.relative_test_filename(output_path))
port.update_baseline(output_path, data)
@@ -231,17 +247,17 @@ class SingleTestRunner(object):
driver_output.crashed_process_name,
driver_output.crashed_pid))
if driver_output.error:
- _log.debug("%s %s crashed, (stderr lines):" % (self._worker_name, testname))
+ _log.debug('%s %s crashed, (stderr lines):' % (self._worker_name, testname))
else:
- _log.debug("%s %s crashed, (no stderr)" % (self._worker_name, testname))
+ _log.debug('%s %s crashed, (no stderr)' % (self._worker_name, testname))
elif driver_output.leak:
failures.append(test_failures.FailureLeak(bool(reference_filename),
driver_output.leak_log))
- _log.debug("%s %s leaked" % (self._worker_name, testname))
+ _log.debug('%s %s leaked' % (self._worker_name, testname))
elif driver_output.error:
- _log.debug("%s %s output stderr lines:" % (self._worker_name, testname))
+ _log.debug('%s %s output stderr lines:' % (self._worker_name, testname))
for line in driver_output.error.splitlines():
- _log.debug(" %s" % line)
+ _log.debug(' %s' % line)
return failures
def _compare_output(self, expected_driver_output, driver_output):
@@ -283,13 +299,13 @@ class SingleTestRunner(object):
return True, []
def _is_render_tree(self, text):
- return text and "layer at (0,0) size 800x600" in text
+ return text and 'layer at (0,0) size 800x600' in text
def _compare_text(self, expected_text, actual_text):
failures = []
if (expected_text and actual_text and
- # Assuming expected_text is already normalized.
- self._port.do_text_results_differ(expected_text, self._get_normalized_output_text(actual_text))):
+ # Assuming expected_text is already normalized.
+ self._port.do_text_results_differ(expected_text, self._get_normalized_output_text(actual_text))):
failures.append(test_failures.FailureTextMismatch())
elif actual_text and not expected_text:
failures.append(test_failures.FailureMissingResult())
@@ -298,7 +314,7 @@ class SingleTestRunner(object):
def _compare_audio(self, expected_audio, actual_audio):
failures = []
if (expected_audio and actual_audio and
- self._port.do_audio_results_differ(expected_audio, actual_audio)):
+ self._port.do_audio_results_differ(expected_audio, actual_audio)):
failures.append(test_failures.FailureAudioMismatch())
elif actual_audio and not expected_audio:
failures.append(test_failures.FailureMissingAudio())
@@ -311,7 +327,7 @@ class SingleTestRunner(object):
# changed to "\r\n" by our system (Python/Cygwin), resulting in
# "\r\r\n", when, in fact, we wanted to compare the text output with
# the normalized text expectation files.
- return output.replace("\r\r\n", "\r\n").replace("\r\n", "\n")
+ return output.replace('\r\r\n', '\r\n').replace('\r\n', '\n')
# FIXME: This function also creates the image diff. Maybe that work should
# be handled elsewhere?
@@ -368,14 +384,25 @@ class SingleTestRunner(object):
reference_test_names.append(reference_test_name)
driver_input = DriverInput(reference_test_name, self._timeout, image_hash=None, should_run_pixel_test=True, args=args)
reference_output = self._driver.run_test(driver_input, self._stop_when_done)
- test_result = self._compare_output_with_reference(reference_output, test_output, reference_filename, expectation == '!=')
+ test_result = self._compare_output_with_reference(
+ reference_output,
+ test_output,
+ reference_filename,
+ expectation == '!=')
if (expectation == '!=' and test_result.failures) or (expectation == '==' and not test_result.failures):
break
total_test_time += test_result.test_run_time
assert(reference_output)
- test_result_writer.write_test_result(self._filesystem, self._port, self._results_directory, self._test_name, test_output, reference_output, test_result.failures)
+ test_result_writer.write_test_result(
+ self._filesystem,
+ self._port,
+ self._results_directory,
+ self._test_name,
+ test_output,
+ reference_output,
+ test_result.failures)
# FIXME: We don't really deal with a mix of reftest types properly. We pass in a set() to reftest_type
# and only really handle the first of the references in the result.
@@ -406,7 +433,7 @@ class SingleTestRunner(object):
elif err_str:
_log.error(err_str)
else:
- _log.warning(" %s -> ref test hashes matched but diff failed" % self._test_name)
+ _log.warning(' %s -> ref test hashes matched but diff failed' % self._test_name)
elif reference_driver_output.image_hash != actual_driver_output.image_hash:
diff, err_str = self._port.diff_image(reference_driver_output.image, actual_driver_output.image)

Powered by Google App Engine
This is Rietveld 408576698