Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(200)

Unified Diff: tools/telemetry/telemetry/unittest/json_results.py

Issue 659293003: Switch telemetry over to use typ to run the unit tests. (Closed) Base URL: https://chromium.googlesource.com/chromium/src.git@rev_typ_v086
Patch Set: fix merge error in browser_test_case Created 6 years, 1 month ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: tools/telemetry/telemetry/unittest/json_results.py
diff --git a/tools/telemetry/telemetry/unittest/json_results.py b/tools/telemetry/telemetry/unittest/json_results.py
deleted file mode 100644
index 256e5c93efecd3d9773c6ce337a606abe3bb0357..0000000000000000000000000000000000000000
--- a/tools/telemetry/telemetry/unittest/json_results.py
+++ /dev/null
@@ -1,247 +0,0 @@
-# Copyright 2014 The Chromium Authors. All rights reserved.
-# Use of this source code is governed by a BSD-style license that can be
-# found in the LICENSE file.
-
-import functools
-import json
-import re
-import time
-import unittest
-import urllib2
-
-
-# TODO(dpranke): This code is largely cloned from, and redundant with,
-# src/mojo/tools/run_mojo_python_tests.py, and also duplicates logic
-# in test-webkitpy and run-webkit-tests. We should consolidate the
-# python TestResult parsing/converting/uploading code as much as possible.
-
-
-def AddOptions(parser):
- parser.add_option('--metadata', action='append', default=[],
- help=('optional key=value metadata that will be stored '
- 'in the results files (can be used for revision '
- 'numbers, etc.)'))
- parser.add_option('--write-full-results-to', metavar='FILENAME',
- action='store',
- help='The path to write the list of full results to.')
- parser.add_option('--builder-name',
- help='The name of the builder as shown on the waterfall.')
- parser.add_option('--master-name',
- help='The name of the buildbot master.')
- parser.add_option("--test-results-server", default="",
- help=('If specified, upload full_results.json file to '
- 'this server.'))
- parser.add_option('--test-type',
- help=('Name of test type / step on the waterfall '
- '(e.g., "telemetry_unittests").'))
-
-
-def ValidateArgs(parser, args):
- for val in args.metadata:
- if '=' not in val:
- parser.error('Error: malformed metadata "%s"' % val)
-
- if (args.test_results_server and
- (not args.builder_name or not args.master_name or not args.test_type)):
- parser.error('Error: --builder-name, --master-name, and --test-type '
- 'must be specified along with --test-result-server.')
-
-
-def WriteFullResultsIfNecessary(args, full_results):
- if not args.write_full_results_to:
- return
-
- with open(args.write_full_results_to, 'w') as fp:
- json.dump(full_results, fp, indent=2)
- fp.write("\n")
-
-
-def UploadFullResultsIfNecessary(args, full_results):
- if not args.test_results_server:
- return False, ''
-
- url = 'http://%s/testfile/upload' % args.test_results_server
- attrs = [('builder', args.builder_name),
- ('master', args.master_name),
- ('testtype', args.test_type)]
- content_type, data = _EncodeMultiPartFormData(attrs, full_results)
- return _UploadData(url, data, content_type)
-
-
-TEST_SEPARATOR = '.'
-
-
-def FullResults(args, suite, results):
- """Convert the unittest results to the Chromium JSON test result format.
-
- This matches run-webkit-tests (the layout tests) and the flakiness dashboard.
- """
-
- full_results = {}
- full_results['interrupted'] = False
- full_results['path_delimiter'] = TEST_SEPARATOR
- full_results['version'] = 3
- full_results['seconds_since_epoch'] = time.time()
- full_results['builder_name'] = args.builder_name or ''
- for md in args.metadata:
- key, val = md.split('=', 1)
- full_results[key] = val
-
- all_test_names = AllTestNames(suite)
- sets_of_passing_test_names = map(PassingTestNames, results)
- sets_of_failing_test_names = map(functools.partial(FailedTestNames, suite),
- results)
-
- # TODO(crbug.com/405379): This handles tests that are skipped via the
- # unittest skip decorators (like skipUnless). The tests that are skipped via
- # telemetry's decorators package are not included in the test suite at all so
- # we need those to be passed in in order to include them.
- skipped_tests = (set(all_test_names) - sets_of_passing_test_names[0]
- - sets_of_failing_test_names[0])
-
- num_tests = len(all_test_names)
- num_failures = NumFailuresAfterRetries(suite, results)
- num_skips = len(skipped_tests)
- num_passes = num_tests - num_failures - num_skips
- full_results['num_failures_by_type'] = {
- 'FAIL': num_failures,
- 'PASS': num_passes,
- 'SKIP': num_skips,
- }
-
- full_results['tests'] = {}
-
- for test_name in all_test_names:
- if test_name in skipped_tests:
- value = {
- 'expected': 'SKIP',
- 'actual': 'SKIP',
- }
- else:
- value = {
- 'expected': 'PASS',
- 'actual': ActualResultsForTest(test_name,
- sets_of_failing_test_names,
- sets_of_passing_test_names),
- }
- if value['actual'].endswith('FAIL'):
- value['is_unexpected'] = True
- _AddPathToTrie(full_results['tests'], test_name, value)
-
- return full_results
-
-
-def ActualResultsForTest(test_name, sets_of_failing_test_names,
- sets_of_passing_test_names):
- actuals = []
- for retry_num in range(len(sets_of_failing_test_names)):
- if test_name in sets_of_failing_test_names[retry_num]:
- actuals.append('FAIL')
- elif test_name in sets_of_passing_test_names[retry_num]:
- assert ((retry_num == 0) or
- (test_name in sets_of_failing_test_names[retry_num - 1])), (
- 'We should not have run a test that did not fail '
- 'on the previous run.')
- actuals.append('PASS')
-
- assert actuals, 'We did not find any result data for %s.' % test_name
- return ' '.join(actuals)
-
-
-def ExitCodeFromFullResults(full_results):
- return 1 if full_results['num_failures_by_type']['FAIL'] else 0
-
-
-def AllTestNames(suite):
- test_names = []
- # _tests is protected pylint: disable=W0212
- for test in suite._tests:
- if isinstance(test, unittest.suite.TestSuite):
- test_names.extend(AllTestNames(test))
- else:
- test_names.append(test.id())
- return test_names
-
-
-def NumFailuresAfterRetries(suite, results):
- return len(FailedTestNames(suite, results[-1]))
-
-
-def FailedTestNames(suite, result):
- failed_test_names = set()
- for test, error in result.failures + result.errors:
- if isinstance(test, unittest.TestCase):
- failed_test_names.add(test.id())
- elif isinstance(test, unittest.suite._ErrorHolder): # pylint: disable=W0212
- # If there's an error in setUpClass or setUpModule, unittest gives us an
- # _ErrorHolder object. We can parse the object's id for the class or
- # module that failed, then find all tests in that class or module.
- match = re.match('setUp[a-zA-Z]+ \\((.+)\\)', test.id())
- assert match, "Don't know how to retry after this error:\n%s" % error
- module_or_class = match.groups()[0]
- failed_test_names |= _FindChildren(module_or_class, AllTestNames(suite))
- else:
- assert False, 'Unknown test type: %s' % test.__class__
- return failed_test_names
-
-
-def _FindChildren(parent, potential_children):
- children = set()
- parent_name_parts = parent.split('.')
- for potential_child in potential_children:
- child_name_parts = potential_child.split('.')
- if parent_name_parts == child_name_parts[:len(parent_name_parts)]:
- children.add(potential_child)
- return children
-
-
-def PassingTestNames(result):
- return set(test.id() for test in result.successes)
-
-
-def _AddPathToTrie(trie, path, value):
- if TEST_SEPARATOR not in path:
- trie[path] = value
- return
- directory, rest = path.split(TEST_SEPARATOR, 1)
- if directory not in trie:
- trie[directory] = {}
- _AddPathToTrie(trie[directory], rest, value)
-
-
-def _EncodeMultiPartFormData(attrs, full_results):
- # Cloned from webkitpy/common/net/file_uploader.py
- BOUNDARY = '-M-A-G-I-C---B-O-U-N-D-A-R-Y-'
- CRLF = '\r\n'
- lines = []
-
- for key, value in attrs:
- lines.append('--' + BOUNDARY)
- lines.append('Content-Disposition: form-data; name="%s"' % key)
- lines.append('')
- lines.append(value)
-
- lines.append('--' + BOUNDARY)
- lines.append('Content-Disposition: form-data; name="file"; '
- 'filename="full_results.json"')
- lines.append('Content-Type: application/json')
- lines.append('')
- lines.append(json.dumps(full_results))
-
- lines.append('--' + BOUNDARY + '--')
- lines.append('')
- body = CRLF.join(lines)
- content_type = 'multipart/form-data; boundary=%s' % BOUNDARY
- return content_type, body
-
-
-def _UploadData(url, data, content_type):
- request = urllib2.Request(url, data, {'Content-Type': content_type})
- try:
- response = urllib2.urlopen(request)
- if response.code == 200:
- return False, ''
- return True, ('Uploading the JSON results failed with %d: "%s"' %
- (response.code, response.read()))
- except Exception as e:
- return True, 'Uploading the JSON results raised "%s"\n' % str(e)
« no previous file with comments | « tools/telemetry/telemetry/unittest/browser_test_case.py ('k') | tools/telemetry/telemetry/unittest/run_chromeos_tests.py » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698