Index: Tools/Scripts/webkitpy/layout_tests/views/buildbot_results.py |
diff --git a/Tools/Scripts/webkitpy/layout_tests/views/buildbot_results.py b/Tools/Scripts/webkitpy/layout_tests/views/buildbot_results.py |
index 82d334af27ad1288da61633ae4cc0586106f1974..78bbf41c5d0697a7b94b2e513c1a47708de75cb3 100644 |
--- a/Tools/Scripts/webkitpy/layout_tests/views/buildbot_results.py |
+++ b/Tools/Scripts/webkitpy/layout_tests/views/buildbot_results.py |
@@ -62,22 +62,22 @@ class BuildBotPrinter(object): |
if total > 0: |
percent_passed = float(passed) * 100 / total |
- self._print("=> Results: %d/%d tests passed (%.1f%%)" % (passed, total, percent_passed)) |
- self._print("") |
- self._print_run_results_entry(run_results, test_expectations.NOW, "Tests to be fixed") |
+ self._print('=> Results: %d/%d tests passed (%.1f%%)' % (passed, total, percent_passed)) |
+ self._print('') |
+ self._print_run_results_entry(run_results, test_expectations.NOW, 'Tests to be fixed') |
- self._print("") |
+ self._print('') |
# FIXME: We should be skipping anything marked WONTFIX, so we shouldn't bother logging these stats. |
self._print_run_results_entry(run_results, test_expectations.WONTFIX, |
- "Tests that will only be fixed if they crash (WONTFIX)") |
- self._print("") |
+ 'Tests that will only be fixed if they crash (WONTFIX)') |
+ self._print('') |
def _print_run_results_entry(self, run_results, timeline, heading): |
total = len(run_results.tests_by_timeline[timeline]) |
not_passing = (total - |
- len(run_results.tests_by_expectation[test_expectations.PASS] & |
- run_results.tests_by_timeline[timeline])) |
- self._print("=> %s (%d):" % (heading, not_passing)) |
+ len(run_results.tests_by_expectation[test_expectations.PASS] & |
+ run_results.tests_by_timeline[timeline])) |
+ self._print('=> %s (%d):' % (heading, not_passing)) |
for result in TestExpectations.EXPECTATION_DESCRIPTIONS.keys(): |
if result in (test_expectations.PASS, test_expectations.SKIP): |
@@ -86,7 +86,7 @@ class BuildBotPrinter(object): |
desc = TestExpectations.EXPECTATION_DESCRIPTIONS[result] |
if not_passing and len(results): |
pct = len(results) * 100.0 / not_passing |
- self._print(" %5d %-24s (%4.1f%%)" % (len(results), desc, pct)) |
+ self._print(' %5d %-24s (%4.1f%%)' % (len(results), desc, pct)) |
def print_unexpected_results(self, summarized_results, enabled_pixel_tests_in_retry=False): |
passes = {} |
@@ -97,8 +97,8 @@ class BuildBotPrinter(object): |
dict.setdefault(key, []).append(value) |
def add_result(test, results, passes=passes, flaky=flaky, regressions=regressions): |
- actual = results['actual'].split(" ") |
- expected = results['expected'].split(" ") |
+ actual = results['actual'].split(' ') |
+ expected = results['expected'].split(' ') |
if 'is_unexpected' not in results or not results['is_unexpected']: |
# Don't print anything for tests that ran as expected. |
@@ -122,43 +122,44 @@ class BuildBotPrinter(object): |
layouttestresults.for_each_test(summarized_results['tests'], add_result) |
if len(passes) or len(flaky) or len(regressions): |
- self._print("") |
+ self._print('') |
if len(passes): |
for key, tests in passes.iteritems(): |
- self._print("%s: (%d)" % (key, len(tests))) |
+ self._print('%s: (%d)' % (key, len(tests))) |
tests.sort() |
for test in tests: |
- self._print(" %s" % test) |
- self._print("") |
- self._print("") |
+ self._print(' %s' % test) |
+ self._print('') |
+ self._print('') |
if len(flaky): |
descriptions = TestExpectations.EXPECTATION_DESCRIPTIONS |
for key, tests in flaky.iteritems(): |
result = TestExpectations.EXPECTATIONS[key.lower()] |
- self._print("Unexpected flakiness: %s (%d)" % (descriptions[result], len(tests))) |
+ self._print('Unexpected flakiness: %s (%d)' % (descriptions[result], len(tests))) |
tests.sort() |
for test in tests: |
result = layouttestresults.result_for_test(summarized_results['tests'], test) |
- actual = result['actual'].split(" ") |
- expected = result['expected'].split(" ") |
+ actual = result['actual'].split(' ') |
+ expected = result['expected'].split(' ') |
result = TestExpectations.EXPECTATIONS[key.lower()] |
# FIXME: clean this up once the old syntax is gone |
- new_expectations_list = [TestExpectationParser._inverted_expectation_tokens[exp] for exp in list(set(actual) | set(expected))] |
- self._print(" %s [ %s ]" % (test, " ".join(new_expectations_list))) |
- self._print("") |
- self._print("") |
+ new_expectations_list = [TestExpectationParser._inverted_expectation_tokens[exp] |
+ for exp in list(set(actual) | set(expected))] |
+ self._print(' %s [ %s ]' % (test, ' '.join(new_expectations_list))) |
+ self._print('') |
+ self._print('') |
if len(regressions): |
descriptions = TestExpectations.EXPECTATION_DESCRIPTIONS |
for key, tests in regressions.iteritems(): |
result = TestExpectations.EXPECTATIONS[key.lower()] |
- self._print("Regressions: Unexpected %s (%d)" % (descriptions[result], len(tests))) |
+ self._print('Regressions: Unexpected %s (%d)' % (descriptions[result], len(tests))) |
tests.sort() |
for test in tests: |
- self._print(" %s [ %s ]" % (test, TestExpectationParser._inverted_expectation_tokens[key])) |
- self._print("") |
+ self._print(' %s [ %s ]' % (test, TestExpectationParser._inverted_expectation_tokens[key])) |
+ self._print('') |
if len(summarized_results['tests']) and self.debug_logging: |
- self._print("%s" % ("-" * 78)) |
+ self._print('%s' % ('-' * 78)) |