Index: tools/telemetry/telemetry/unittest/json_results.py |
diff --git a/tools/telemetry/telemetry/unittest/json_results.py b/tools/telemetry/telemetry/unittest/json_results.py |
deleted file mode 100644 |
index 256e5c93efecd3d9773c6ce337a606abe3bb0357..0000000000000000000000000000000000000000 |
--- a/tools/telemetry/telemetry/unittest/json_results.py |
+++ /dev/null |
@@ -1,247 +0,0 @@ |
-# Copyright 2014 The Chromium Authors. All rights reserved. |
-# Use of this source code is governed by a BSD-style license that can be |
-# found in the LICENSE file. |
- |
-import functools |
-import json |
-import re |
-import time |
-import unittest |
-import urllib2 |
- |
- |
-# TODO(dpranke): This code is largely cloned from, and redundant with, |
-# src/mojo/tools/run_mojo_python_tests.py, and also duplicates logic |
-# in test-webkitpy and run-webkit-tests. We should consolidate the |
-# python TestResult parsing/converting/uploading code as much as possible. |
- |
- |
-def AddOptions(parser): |
- parser.add_option('--metadata', action='append', default=[], |
- help=('optional key=value metadata that will be stored ' |
- 'in the results files (can be used for revision ' |
- 'numbers, etc.)')) |
- parser.add_option('--write-full-results-to', metavar='FILENAME', |
- action='store', |
- help='The path to write the list of full results to.') |
- parser.add_option('--builder-name', |
- help='The name of the builder as shown on the waterfall.') |
- parser.add_option('--master-name', |
- help='The name of the buildbot master.') |
- parser.add_option("--test-results-server", default="", |
- help=('If specified, upload full_results.json file to ' |
- 'this server.')) |
- parser.add_option('--test-type', |
- help=('Name of test type / step on the waterfall ' |
- '(e.g., "telemetry_unittests").')) |
- |
- |
-def ValidateArgs(parser, args): |
- for val in args.metadata: |
- if '=' not in val: |
- parser.error('Error: malformed metadata "%s"' % val) |
- |
- if (args.test_results_server and |
- (not args.builder_name or not args.master_name or not args.test_type)): |
- parser.error('Error: --builder-name, --master-name, and --test-type ' |
- 'must be specified along with --test-result-server.') |
- |
- |
-def WriteFullResultsIfNecessary(args, full_results): |
- if not args.write_full_results_to: |
- return |
- |
- with open(args.write_full_results_to, 'w') as fp: |
- json.dump(full_results, fp, indent=2) |
- fp.write("\n") |
- |
- |
-def UploadFullResultsIfNecessary(args, full_results): |
- if not args.test_results_server: |
- return False, '' |
- |
- url = 'http://%s/testfile/upload' % args.test_results_server |
- attrs = [('builder', args.builder_name), |
- ('master', args.master_name), |
- ('testtype', args.test_type)] |
- content_type, data = _EncodeMultiPartFormData(attrs, full_results) |
- return _UploadData(url, data, content_type) |
- |
- |
-TEST_SEPARATOR = '.' |
- |
- |
-def FullResults(args, suite, results): |
- """Convert the unittest results to the Chromium JSON test result format. |
- |
- This matches run-webkit-tests (the layout tests) and the flakiness dashboard. |
- """ |
- |
- full_results = {} |
- full_results['interrupted'] = False |
- full_results['path_delimiter'] = TEST_SEPARATOR |
- full_results['version'] = 3 |
- full_results['seconds_since_epoch'] = time.time() |
- full_results['builder_name'] = args.builder_name or '' |
- for md in args.metadata: |
- key, val = md.split('=', 1) |
- full_results[key] = val |
- |
- all_test_names = AllTestNames(suite) |
- sets_of_passing_test_names = map(PassingTestNames, results) |
- sets_of_failing_test_names = map(functools.partial(FailedTestNames, suite), |
- results) |
- |
- # TODO(crbug.com/405379): This handles tests that are skipped via the |
- # unittest skip decorators (like skipUnless). The tests that are skipped via |
- # telemetry's decorators package are not included in the test suite at all so |
- # we need those to be passed in in order to include them. |
- skipped_tests = (set(all_test_names) - sets_of_passing_test_names[0] |
- - sets_of_failing_test_names[0]) |
- |
- num_tests = len(all_test_names) |
- num_failures = NumFailuresAfterRetries(suite, results) |
- num_skips = len(skipped_tests) |
- num_passes = num_tests - num_failures - num_skips |
- full_results['num_failures_by_type'] = { |
- 'FAIL': num_failures, |
- 'PASS': num_passes, |
- 'SKIP': num_skips, |
- } |
- |
- full_results['tests'] = {} |
- |
- for test_name in all_test_names: |
- if test_name in skipped_tests: |
- value = { |
- 'expected': 'SKIP', |
- 'actual': 'SKIP', |
- } |
- else: |
- value = { |
- 'expected': 'PASS', |
- 'actual': ActualResultsForTest(test_name, |
- sets_of_failing_test_names, |
- sets_of_passing_test_names), |
- } |
- if value['actual'].endswith('FAIL'): |
- value['is_unexpected'] = True |
- _AddPathToTrie(full_results['tests'], test_name, value) |
- |
- return full_results |
- |
- |
-def ActualResultsForTest(test_name, sets_of_failing_test_names, |
- sets_of_passing_test_names): |
- actuals = [] |
- for retry_num in range(len(sets_of_failing_test_names)): |
- if test_name in sets_of_failing_test_names[retry_num]: |
- actuals.append('FAIL') |
- elif test_name in sets_of_passing_test_names[retry_num]: |
- assert ((retry_num == 0) or |
- (test_name in sets_of_failing_test_names[retry_num - 1])), ( |
- 'We should not have run a test that did not fail ' |
- 'on the previous run.') |
- actuals.append('PASS') |
- |
- assert actuals, 'We did not find any result data for %s.' % test_name |
- return ' '.join(actuals) |
- |
- |
-def ExitCodeFromFullResults(full_results): |
- return 1 if full_results['num_failures_by_type']['FAIL'] else 0 |
- |
- |
-def AllTestNames(suite): |
- test_names = [] |
- # _tests is protected pylint: disable=W0212 |
- for test in suite._tests: |
- if isinstance(test, unittest.suite.TestSuite): |
- test_names.extend(AllTestNames(test)) |
- else: |
- test_names.append(test.id()) |
- return test_names |
- |
- |
-def NumFailuresAfterRetries(suite, results): |
- return len(FailedTestNames(suite, results[-1])) |
- |
- |
-def FailedTestNames(suite, result): |
- failed_test_names = set() |
- for test, error in result.failures + result.errors: |
- if isinstance(test, unittest.TestCase): |
- failed_test_names.add(test.id()) |
- elif isinstance(test, unittest.suite._ErrorHolder): # pylint: disable=W0212 |
- # If there's an error in setUpClass or setUpModule, unittest gives us an |
- # _ErrorHolder object. We can parse the object's id for the class or |
- # module that failed, then find all tests in that class or module. |
- match = re.match('setUp[a-zA-Z]+ \\((.+)\\)', test.id()) |
- assert match, "Don't know how to retry after this error:\n%s" % error |
- module_or_class = match.groups()[0] |
- failed_test_names |= _FindChildren(module_or_class, AllTestNames(suite)) |
- else: |
- assert False, 'Unknown test type: %s' % test.__class__ |
- return failed_test_names |
- |
- |
-def _FindChildren(parent, potential_children): |
- children = set() |
- parent_name_parts = parent.split('.') |
- for potential_child in potential_children: |
- child_name_parts = potential_child.split('.') |
- if parent_name_parts == child_name_parts[:len(parent_name_parts)]: |
- children.add(potential_child) |
- return children |
- |
- |
-def PassingTestNames(result): |
- return set(test.id() for test in result.successes) |
- |
- |
-def _AddPathToTrie(trie, path, value): |
- if TEST_SEPARATOR not in path: |
- trie[path] = value |
- return |
- directory, rest = path.split(TEST_SEPARATOR, 1) |
- if directory not in trie: |
- trie[directory] = {} |
- _AddPathToTrie(trie[directory], rest, value) |
- |
- |
-def _EncodeMultiPartFormData(attrs, full_results): |
- # Cloned from webkitpy/common/net/file_uploader.py |
- BOUNDARY = '-M-A-G-I-C---B-O-U-N-D-A-R-Y-' |
- CRLF = '\r\n' |
- lines = [] |
- |
- for key, value in attrs: |
- lines.append('--' + BOUNDARY) |
- lines.append('Content-Disposition: form-data; name="%s"' % key) |
- lines.append('') |
- lines.append(value) |
- |
- lines.append('--' + BOUNDARY) |
- lines.append('Content-Disposition: form-data; name="file"; ' |
- 'filename="full_results.json"') |
- lines.append('Content-Type: application/json') |
- lines.append('') |
- lines.append(json.dumps(full_results)) |
- |
- lines.append('--' + BOUNDARY + '--') |
- lines.append('') |
- body = CRLF.join(lines) |
- content_type = 'multipart/form-data; boundary=%s' % BOUNDARY |
- return content_type, body |
- |
- |
-def _UploadData(url, data, content_type): |
- request = urllib2.Request(url, data, {'Content-Type': content_type}) |
- try: |
- response = urllib2.urlopen(request) |
- if response.code == 200: |
- return False, '' |
- return True, ('Uploading the JSON results failed with %d: "%s"' % |
- (response.code, response.read())) |
- except Exception as e: |
- return True, 'Uploading the JSON results raised "%s"\n' % str(e) |