| Index: tools/telemetry/telemetry/unittest_util/json_results.py
|
| diff --git a/tools/telemetry/telemetry/unittest_util/json_results.py b/tools/telemetry/telemetry/unittest_util/json_results.py
|
| deleted file mode 100644
|
| index 256e5c93efecd3d9773c6ce337a606abe3bb0357..0000000000000000000000000000000000000000
|
| --- a/tools/telemetry/telemetry/unittest_util/json_results.py
|
| +++ /dev/null
|
| @@ -1,247 +0,0 @@
|
| -# Copyright 2014 The Chromium Authors. All rights reserved.
|
| -# Use of this source code is governed by a BSD-style license that can be
|
| -# found in the LICENSE file.
|
| -
|
| -import functools
|
| -import json
|
| -import re
|
| -import time
|
| -import unittest
|
| -import urllib2
|
| -
|
| -
|
| -# TODO(dpranke): This code is largely cloned from, and redundant with,
|
| -# src/mojo/tools/run_mojo_python_tests.py, and also duplicates logic
|
| -# in test-webkitpy and run-webkit-tests. We should consolidate the
|
| -# python TestResult parsing/converting/uploading code as much as possible.
|
| -
|
| -
|
| -def AddOptions(parser):
|
| - parser.add_option('--metadata', action='append', default=[],
|
| - help=('optional key=value metadata that will be stored '
|
| - 'in the results files (can be used for revision '
|
| - 'numbers, etc.)'))
|
| - parser.add_option('--write-full-results-to', metavar='FILENAME',
|
| - action='store',
|
| - help='The path to write the list of full results to.')
|
| - parser.add_option('--builder-name',
|
| - help='The name of the builder as shown on the waterfall.')
|
| - parser.add_option('--master-name',
|
| - help='The name of the buildbot master.')
|
| - parser.add_option("--test-results-server", default="",
|
| - help=('If specified, upload full_results.json file to '
|
| - 'this server.'))
|
| - parser.add_option('--test-type',
|
| - help=('Name of test type / step on the waterfall '
|
| - '(e.g., "telemetry_unittests").'))
|
| -
|
| -
|
| -def ValidateArgs(parser, args):
|
| - for val in args.metadata:
|
| - if '=' not in val:
|
| - parser.error('Error: malformed metadata "%s"' % val)
|
| -
|
| - if (args.test_results_server and
|
| - (not args.builder_name or not args.master_name or not args.test_type)):
|
| - parser.error('Error: --builder-name, --master-name, and --test-type '
|
| - 'must be specified along with --test-result-server.')
|
| -
|
| -
|
| -def WriteFullResultsIfNecessary(args, full_results):
|
| - if not args.write_full_results_to:
|
| - return
|
| -
|
| - with open(args.write_full_results_to, 'w') as fp:
|
| - json.dump(full_results, fp, indent=2)
|
| - fp.write("\n")
|
| -
|
| -
|
| -def UploadFullResultsIfNecessary(args, full_results):
|
| - if not args.test_results_server:
|
| - return False, ''
|
| -
|
| - url = 'http://%s/testfile/upload' % args.test_results_server
|
| - attrs = [('builder', args.builder_name),
|
| - ('master', args.master_name),
|
| - ('testtype', args.test_type)]
|
| - content_type, data = _EncodeMultiPartFormData(attrs, full_results)
|
| - return _UploadData(url, data, content_type)
|
| -
|
| -
|
| -TEST_SEPARATOR = '.'
|
| -
|
| -
|
| -def FullResults(args, suite, results):
|
| - """Convert the unittest results to the Chromium JSON test result format.
|
| -
|
| - This matches run-webkit-tests (the layout tests) and the flakiness dashboard.
|
| - """
|
| -
|
| - full_results = {}
|
| - full_results['interrupted'] = False
|
| - full_results['path_delimiter'] = TEST_SEPARATOR
|
| - full_results['version'] = 3
|
| - full_results['seconds_since_epoch'] = time.time()
|
| - full_results['builder_name'] = args.builder_name or ''
|
| - for md in args.metadata:
|
| - key, val = md.split('=', 1)
|
| - full_results[key] = val
|
| -
|
| - all_test_names = AllTestNames(suite)
|
| - sets_of_passing_test_names = map(PassingTestNames, results)
|
| - sets_of_failing_test_names = map(functools.partial(FailedTestNames, suite),
|
| - results)
|
| -
|
| - # TODO(crbug.com/405379): This handles tests that are skipped via the
|
| - # unittest skip decorators (like skipUnless). The tests that are skipped via
|
| - # telemetry's decorators package are not included in the test suite at all so
|
| - # we need those to be passed in in order to include them.
|
| - skipped_tests = (set(all_test_names) - sets_of_passing_test_names[0]
|
| - - sets_of_failing_test_names[0])
|
| -
|
| - num_tests = len(all_test_names)
|
| - num_failures = NumFailuresAfterRetries(suite, results)
|
| - num_skips = len(skipped_tests)
|
| - num_passes = num_tests - num_failures - num_skips
|
| - full_results['num_failures_by_type'] = {
|
| - 'FAIL': num_failures,
|
| - 'PASS': num_passes,
|
| - 'SKIP': num_skips,
|
| - }
|
| -
|
| - full_results['tests'] = {}
|
| -
|
| - for test_name in all_test_names:
|
| - if test_name in skipped_tests:
|
| - value = {
|
| - 'expected': 'SKIP',
|
| - 'actual': 'SKIP',
|
| - }
|
| - else:
|
| - value = {
|
| - 'expected': 'PASS',
|
| - 'actual': ActualResultsForTest(test_name,
|
| - sets_of_failing_test_names,
|
| - sets_of_passing_test_names),
|
| - }
|
| - if value['actual'].endswith('FAIL'):
|
| - value['is_unexpected'] = True
|
| - _AddPathToTrie(full_results['tests'], test_name, value)
|
| -
|
| - return full_results
|
| -
|
| -
|
| -def ActualResultsForTest(test_name, sets_of_failing_test_names,
|
| - sets_of_passing_test_names):
|
| - actuals = []
|
| - for retry_num in range(len(sets_of_failing_test_names)):
|
| - if test_name in sets_of_failing_test_names[retry_num]:
|
| - actuals.append('FAIL')
|
| - elif test_name in sets_of_passing_test_names[retry_num]:
|
| - assert ((retry_num == 0) or
|
| - (test_name in sets_of_failing_test_names[retry_num - 1])), (
|
| - 'We should not have run a test that did not fail '
|
| - 'on the previous run.')
|
| - actuals.append('PASS')
|
| -
|
| - assert actuals, 'We did not find any result data for %s.' % test_name
|
| - return ' '.join(actuals)
|
| -
|
| -
|
| -def ExitCodeFromFullResults(full_results):
|
| - return 1 if full_results['num_failures_by_type']['FAIL'] else 0
|
| -
|
| -
|
| -def AllTestNames(suite):
|
| - test_names = []
|
| - # _tests is protected pylint: disable=W0212
|
| - for test in suite._tests:
|
| - if isinstance(test, unittest.suite.TestSuite):
|
| - test_names.extend(AllTestNames(test))
|
| - else:
|
| - test_names.append(test.id())
|
| - return test_names
|
| -
|
| -
|
| -def NumFailuresAfterRetries(suite, results):
|
| - return len(FailedTestNames(suite, results[-1]))
|
| -
|
| -
|
| -def FailedTestNames(suite, result):
|
| - failed_test_names = set()
|
| - for test, error in result.failures + result.errors:
|
| - if isinstance(test, unittest.TestCase):
|
| - failed_test_names.add(test.id())
|
| - elif isinstance(test, unittest.suite._ErrorHolder): # pylint: disable=W0212
|
| - # If there's an error in setUpClass or setUpModule, unittest gives us an
|
| - # _ErrorHolder object. We can parse the object's id for the class or
|
| - # module that failed, then find all tests in that class or module.
|
| - match = re.match('setUp[a-zA-Z]+ \\((.+)\\)', test.id())
|
| - assert match, "Don't know how to retry after this error:\n%s" % error
|
| - module_or_class = match.groups()[0]
|
| - failed_test_names |= _FindChildren(module_or_class, AllTestNames(suite))
|
| - else:
|
| - assert False, 'Unknown test type: %s' % test.__class__
|
| - return failed_test_names
|
| -
|
| -
|
| -def _FindChildren(parent, potential_children):
|
| - children = set()
|
| - parent_name_parts = parent.split('.')
|
| - for potential_child in potential_children:
|
| - child_name_parts = potential_child.split('.')
|
| - if parent_name_parts == child_name_parts[:len(parent_name_parts)]:
|
| - children.add(potential_child)
|
| - return children
|
| -
|
| -
|
| -def PassingTestNames(result):
|
| - return set(test.id() for test in result.successes)
|
| -
|
| -
|
| -def _AddPathToTrie(trie, path, value):
|
| - if TEST_SEPARATOR not in path:
|
| - trie[path] = value
|
| - return
|
| - directory, rest = path.split(TEST_SEPARATOR, 1)
|
| - if directory not in trie:
|
| - trie[directory] = {}
|
| - _AddPathToTrie(trie[directory], rest, value)
|
| -
|
| -
|
| -def _EncodeMultiPartFormData(attrs, full_results):
|
| - # Cloned from webkitpy/common/net/file_uploader.py
|
| - BOUNDARY = '-M-A-G-I-C---B-O-U-N-D-A-R-Y-'
|
| - CRLF = '\r\n'
|
| - lines = []
|
| -
|
| - for key, value in attrs:
|
| - lines.append('--' + BOUNDARY)
|
| - lines.append('Content-Disposition: form-data; name="%s"' % key)
|
| - lines.append('')
|
| - lines.append(value)
|
| -
|
| - lines.append('--' + BOUNDARY)
|
| - lines.append('Content-Disposition: form-data; name="file"; '
|
| - 'filename="full_results.json"')
|
| - lines.append('Content-Type: application/json')
|
| - lines.append('')
|
| - lines.append(json.dumps(full_results))
|
| -
|
| - lines.append('--' + BOUNDARY + '--')
|
| - lines.append('')
|
| - body = CRLF.join(lines)
|
| - content_type = 'multipart/form-data; boundary=%s' % BOUNDARY
|
| - return content_type, body
|
| -
|
| -
|
| -def _UploadData(url, data, content_type):
|
| - request = urllib2.Request(url, data, {'Content-Type': content_type})
|
| - try:
|
| - response = urllib2.urlopen(request)
|
| - if response.code == 200:
|
| - return False, ''
|
| - return True, ('Uploading the JSON results failed with %d: "%s"' %
|
| - (response.code, response.read()))
|
| - except Exception as e:
|
| - return True, 'Uploading the JSON results raised "%s"\n' % str(e)
|
|
|