Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(49)

Unified Diff: gm/rebaseline_server/results.py

Issue 24274003: Create HTTP-based GM results viewer. (Closed) Base URL: http://skia.googlecode.com/svn/trunk/
Patch Set: linewrap Created 7 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « no previous file | gm/rebaseline_server/server.py » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: gm/rebaseline_server/results.py
===================================================================
--- gm/rebaseline_server/results.py (revision 0)
+++ gm/rebaseline_server/results.py (revision 0)
@@ -0,0 +1,178 @@
+#!/usr/bin/python
+
+'''
+Copyright 2013 Google Inc.
+
+Use of this source code is governed by a BSD-style license that can be
+found in the LICENSE file.
+'''
+
+'''
+Repackage expected/actual GM results as needed by our HTML rebaseline viewer.
+'''
+
+# System-level imports
+import fnmatch
+import json
+import os
+import re
+import sys
+
+# Imports from within Skia
+#
+# We need to add the 'gm' directory, so that we can import gm_json.py within
+# that directory. That script allows us to parse the actual-results.json file
+# written out by the GM tool.
+# Make sure that the 'gm' dir is in the PYTHONPATH, but add it at the *end*
+# so any dirs that are already in the PYTHONPATH will be preferred.
+GM_DIRECTORY = os.path.dirname(os.path.dirname(os.path.realpath(__file__)))
+if GM_DIRECTORY not in sys.path:
+ sys.path.append(GM_DIRECTORY)
+import gm_json
+
+IMAGE_FILENAME_RE = re.compile(gm_json.IMAGE_FILENAME_PATTERN)
+
+class Results(object):
+ """ Loads actual and expected results from all builders, supplying combined
+ reports as requested. """
+
+ def __init__(self, actuals_root, expected_root):
+ """
+ params:
+ actuals_root: root directory containing all actual-results.json files
+ expected_root: root directory containing all expected-results.json files
+ """
+ self._actual_builder_dicts = Results._GetDictsFromRoot(actuals_root)
+ self._expected_builder_dicts = Results._GetDictsFromRoot(expected_root)
+ self._all_results = self._Combine()
+
+ def GetAll(self):
+ """Return results of all tests, as a list in this form:
+
+ [
+ {
+ "builder": "Test-Mac10.6-MacMini4.1-GeForce320M-x86-Debug",
+ "test": "bigmatrix",
+ "config": "8888",
+ "resultType": "failed",
+ "expectedHashType": "bitmap-64bitMD5",
+ "expectedHashDigest": "10894408024079689926",
+ "actualHashType": "bitmap-64bitMD5",
+ "actualHashDigest": "2409857384569",
+ },
+ ...
+ ]
+ """
+ return self._all_results
+
+ @staticmethod
+ def _GetDictsFromRoot(root, pattern='*.json'):
+ """Read all JSON dictionaries within a directory tree, returning them within
+ a meta-dictionary (keyed by the builder name for each dictionary).
+
+ params:
+ root: path to root of directory tree
+ pattern: which files to read within root (fnmatch-style pattern)
+ """
+ meta_dict = {}
+ for dirpath, dirnames, filenames in os.walk(root):
+ for matching_filename in fnmatch.filter(filenames, pattern):
+ builder = os.path.basename(dirpath)
+ if builder.endswith('-Trybot'):
+ continue
+ fullpath = os.path.join(dirpath, matching_filename)
+ meta_dict[builder] = gm_json.LoadFromFile(fullpath)
+ return meta_dict
+
+ def _Combine(self):
+ """Returns a list of all tests, across all builders, based on the
+ contents of self._actual_builder_dicts and self._expected_builder_dicts .
+ Returns the list in the same form needed for GetAllResults().
+ """
+ all_tests = []
+ for builder in sorted(self._actual_builder_dicts.keys()):
+ actual_results_for_this_builder = (
+ self._actual_builder_dicts[builder][gm_json.JSONKEY_ACTUALRESULTS])
+ for result_type in sorted(actual_results_for_this_builder.keys()):
+ results_of_this_type = actual_results_for_this_builder[result_type]
+ if not results_of_this_type:
+ continue
+ for image_name in sorted(results_of_this_type.keys()):
+ actual_image = results_of_this_type[image_name]
+ try:
+ # TODO(epoger): assumes a single allowed digest per test
+ expected_image = (
+ self._expected_builder_dicts
+ [builder][gm_json.JSONKEY_EXPECTEDRESULTS]
+ [image_name][gm_json.JSONKEY_EXPECTEDRESULTS_ALLOWEDDIGESTS]
+ [0])
+ except (KeyError, TypeError):
+ # There are several cases in which we would expect to find
+ # no expectations for a given test:
+ #
+ # 1. result_type == NOCOMPARISON
+ # There are no expectations for this test yet!
+ #
+ # 2. ignore-tests.txt
+ # If a test has been listed in ignore-tests.txt, then its status
+ # may show as FAILUREIGNORED even if it doesn't have any
+ # expectations yet.
+ #
+ # 3. alternate rendering mode failures (e.g. serialized)
+ # In cases like
+ # https://code.google.com/p/skia/issues/detail?id=1684
+ # ('tileimagefilter GM test failing in serialized render mode'),
+ # the gm-actuals will list a failure for the alternate
+ # rendering mode even though we don't have explicit expectations
+ # for the test (the implicit expectation is that it must
+ # render the same in all rendering modes).
+ #
+ # Don't log types 1 or 2, because they are common.
+ # Log other types, because they are rare and we should know about
+ # them, but don't throw an exception, because we need to keep our
+ # tools working in the meanwhile!
+ if result_type not in [
+ gm_json.JSONKEY_ACTUALRESULTS_NOCOMPARISON,
+ gm_json.JSONKEY_ACTUALRESULTS_FAILUREIGNORED] :
+ print 'WARNING: No expectations found for test: %s' % {
+ 'builder': builder,
+ 'image_name': image_name,
+ 'result_type': result_type,
+ }
+ expected_image = [None, None]
+
+ # If this test was recently rebaselined, it will remain in
+ # the "failed" set of actuals until all the bots have
+ # cycled (although the expectations have indeed been set
+ # from the most recent actuals). Treat these as successes
+ # instead of failures.
+ #
+ # TODO(epoger): Do we need to do something similar in
+ # other cases, such as when we have recently marked a test
+ # as ignoreFailure but it still shows up in the "failed"
+ # category? Maybe we should not rely on the result_type
+ # categories recorded within the gm_actuals AT ALL, and
+ # instead evaluate the result_type ourselves based on what
+ # we see in expectations vs actual checksum?
+ if expected_image == actual_image:
+ updated_result_type = gm_json.JSONKEY_ACTUALRESULTS_SUCCEEDED
+ else:
+ updated_result_type = result_type
+
+ # TODO(epoger): For now, don't include succeeded results.
+ # There are so many of them that they make the client too slow.
+ if updated_result_type == gm_json.JSONKEY_ACTUALRESULTS_SUCCEEDED:
+ continue
+
+ (test, config) = IMAGE_FILENAME_RE.match(image_name).groups()
+ all_tests.append({
+ "builder": builder,
+ "test": test,
+ "config": config,
+ "resultType": updated_result_type,
+ "actualHashType": actual_image[0],
+ "actualHashDigest": str(actual_image[1]),
+ "expectedHashType": expected_image[0],
+ "expectedHashDigest": str(expected_image[1]),
+ })
+ return all_tests
Property changes on: gm/rebaseline_server/results.py
___________________________________________________________________
Added: svn:executable
+ *
« no previous file with comments | « no previous file | gm/rebaseline_server/server.py » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698