| Index: gm/rebaseline_server/server.py
|
| diff --git a/gm/rebaseline_server/server.py b/gm/rebaseline_server/server.py
|
| new file mode 100755
|
| index 0000000000000000000000000000000000000000..85874eb6667bb1353741e9fbe7b4e07dfc53d125
|
| --- /dev/null
|
| +++ b/gm/rebaseline_server/server.py
|
| @@ -0,0 +1,967 @@
|
| +#!/usr/bin/python
|
| +
|
| +"""
|
| +Copyright 2013 Google Inc.
|
| +
|
| +Use of this source code is governed by a BSD-style license that can be
|
| +found in the LICENSE file.
|
| +
|
| +HTTP server for our HTML rebaseline viewer.
|
| +"""
|
| +
|
| +# System-level imports
|
| +import argparse
|
| +import BaseHTTPServer
|
| +import json
|
| +import logging
|
| +import os
|
| +import posixpath
|
| +import re
|
| +import shutil
|
| +import socket
|
| +import subprocess
|
| +import thread
|
| +import threading
|
| +import time
|
| +import urllib
|
| +import urlparse
|
| +
|
| +# Must fix up PYTHONPATH before importing from within Skia
|
| +import rs_fixpypath # pylint: disable=W0611
|
| +
|
| +# Imports from within Skia
|
| +from py.utils import gs_utils
|
| +import buildbot_globals
|
| +import gm_json
|
| +
|
| +# Imports from local dir
|
| +#
|
| +# pylint: disable=C0301
|
| +# Note: we import results under a different name, to avoid confusion with the
|
| +# Server.results() property. See discussion at
|
| +# https://codereview.chromium.org/195943004/diff/1/gm/rebaseline_server/server.py#newcode44
|
| +# pylint: enable=C0301
|
| +import compare_configs
|
| +import compare_rendered_pictures
|
| +import compare_to_expectations
|
| +import download_actuals
|
| +import imagediffdb
|
| +import imagepairset
|
| +import results as results_mod
|
| +import writable_expectations as writable_expectations_mod
|
| +
|
| +
|
| +PATHSPLIT_RE = re.compile('/([^/]+)/(.+)')
|
| +
|
| +# A simple dictionary of file name extensions to MIME types. The empty string
|
| +# entry is used as the default when no extension was given or if the extension
|
| +# has no entry in this dictionary.
|
| +MIME_TYPE_MAP = {'': 'application/octet-stream',
|
| + 'html': 'text/html',
|
| + 'css': 'text/css',
|
| + 'png': 'image/png',
|
| + 'js': 'application/javascript',
|
| + 'json': 'application/json'
|
| + }
|
| +
|
| +# Keys that server.py uses to create the toplevel content header.
|
| +# NOTE: Keep these in sync with static/constants.js
|
| +KEY__EDITS__MODIFICATIONS = 'modifications'
|
| +KEY__EDITS__OLD_RESULTS_HASH = 'oldResultsHash'
|
| +KEY__EDITS__OLD_RESULTS_TYPE = 'oldResultsType'
|
| +KEY__LIVE_EDITS__MODIFICATIONS = 'modifications'
|
| +KEY__LIVE_EDITS__SET_A_DESCRIPTIONS = 'setA'
|
| +KEY__LIVE_EDITS__SET_B_DESCRIPTIONS = 'setB'
|
| +
|
| +DEFAULT_ACTUALS_DIR = results_mod.DEFAULT_ACTUALS_DIR
|
| +DEFAULT_GM_SUMMARIES_BUCKET = download_actuals.GM_SUMMARIES_BUCKET
|
| +DEFAULT_JSON_FILENAME = download_actuals.DEFAULT_JSON_FILENAME
|
| +DEFAULT_PORT = 8888
|
| +
|
| +PARENT_DIRECTORY = os.path.dirname(os.path.realpath(__file__))
|
| +TRUNK_DIRECTORY = os.path.dirname(os.path.dirname(PARENT_DIRECTORY))
|
| +
|
| +# Directory, relative to PARENT_DIRECTORY, within which the server will serve
|
| +# out static files.
|
| +STATIC_CONTENTS_SUBDIR = 'static'
|
| +# All of the GENERATED_*_SUBDIRS are relative to STATIC_CONTENTS_SUBDIR
|
| +GENERATED_HTML_SUBDIR = 'generated-html'
|
| +GENERATED_IMAGES_SUBDIR = 'generated-images'
|
| +GENERATED_JSON_SUBDIR = 'generated-json'
|
| +
|
| +# Directives associated with various HTTP GET requests.
|
| +GET__LIVE_RESULTS = 'live-results'
|
| +GET__PRECOMPUTED_RESULTS = 'results'
|
| +GET__PREFETCH_RESULTS = 'prefetch'
|
| +GET__STATIC_CONTENTS = 'static'
|
| +
|
| +# Parameters we use within do_GET_live_results() and do_GET_prefetch_results()
|
| +LIVE_PARAM__DOWNLOAD_ONLY_DIFFERING = 'downloadOnlyDifferingImages'
|
| +LIVE_PARAM__SET_A_DIR = 'setADir'
|
| +LIVE_PARAM__SET_A_SECTION = 'setASection'
|
| +LIVE_PARAM__SET_B_DIR = 'setBDir'
|
| +LIVE_PARAM__SET_B_SECTION = 'setBSection'
|
| +
|
| +# How often (in seconds) clients should reload while waiting for initial
|
| +# results to load.
|
| +RELOAD_INTERVAL_UNTIL_READY = 10
|
| +
|
| +_GM_SUMMARY_TYPES = [
|
| + results_mod.KEY__HEADER__RESULTS_FAILURES,
|
| + results_mod.KEY__HEADER__RESULTS_ALL,
|
| +]
|
| +# If --compare-configs is specified, compare these configs.
|
| +CONFIG_PAIRS_TO_COMPARE = [('8888', 'gpu')]
|
| +
|
| +# SKP results that are available to compare.
|
| +#
|
| +# TODO(stephana): We don't actually want to maintain this list of platforms.
|
| +# We are just putting them in here for now, as "convenience" links for testing
|
| +# SKP diffs.
|
| +# Ultimately, we will depend on buildbot steps linking to their own diffs on
|
| +# the shared rebaseline_server instance.
|
| +_SKP_BASE_GS_URL = 'gs://' + buildbot_globals.Get('skp_summaries_bucket')
|
| +_SKP_BASE_REPO_URL = (
|
| + compare_rendered_pictures.REPO_URL_PREFIX + posixpath.join(
|
| + 'expectations', 'skp'))
|
| +_SKP_PLATFORMS = [
|
| + 'Test-Mac10.8-MacMini4.1-GeForce320M-x86_64-Debug',
|
| + 'Test-Ubuntu12-ShuttleA-GTX660-x86-Release',
|
| +]
|
| +
|
| +_HTTP_HEADER_CONTENT_LENGTH = 'Content-Length'
|
| +_HTTP_HEADER_CONTENT_TYPE = 'Content-Type'
|
| +
|
| +_SERVER = None # This gets filled in by main()
|
| +
|
| +
|
| +def _run_command(args, directory):
|
| + """Runs a command and returns stdout as a single string.
|
| +
|
| + Args:
|
| + args: the command to run, as a list of arguments
|
| + directory: directory within which to run the command
|
| +
|
| + Returns: stdout, as a string
|
| +
|
| + Raises an Exception if the command failed (exited with nonzero return code).
|
| + """
|
| + logging.debug('_run_command: %s in directory %s' % (args, directory))
|
| + proc = subprocess.Popen(args, cwd=directory,
|
| + stdout=subprocess.PIPE,
|
| + stderr=subprocess.PIPE)
|
| + (stdout, stderr) = proc.communicate()
|
| + if proc.returncode is not 0:
|
| + raise Exception('command "%s" failed in dir "%s": %s' %
|
| + (args, directory, stderr))
|
| + return stdout
|
| +
|
| +
|
| +def _get_routable_ip_address():
|
| + """Returns routable IP address of this host (the IP address of its network
|
| + interface that would be used for most traffic, not its localhost
|
| + interface). See http://stackoverflow.com/a/166589 """
|
| + sock = socket.socket(socket.AF_INET, socket.SOCK_DGRAM)
|
| + sock.connect(('8.8.8.8', 80))
|
| + host = sock.getsockname()[0]
|
| + sock.close()
|
| + return host
|
| +
|
| +
|
| +def _create_index(file_path, config_pairs):
|
| + """Creates an index file linking to all results available from this server.
|
| +
|
| + Prior to https://codereview.chromium.org/215503002 , we had a static
|
| + index.html within our repo. But now that the results may or may not include
|
| + config comparisons, index.html needs to be generated differently depending
|
| + on which results are included.
|
| +
|
| + TODO(epoger): Instead of including raw HTML within the Python code,
|
| + consider restoring the index.html file as a template and using django (or
|
| + similar) to fill in dynamic content.
|
| +
|
| + Args:
|
| + file_path: path on local disk to write index to; any directory components
|
| + of this path that do not already exist will be created
|
| + config_pairs: what pairs of configs (if any) we compare actual results of
|
| + """
|
| + dir_path = os.path.dirname(file_path)
|
| + if not os.path.isdir(dir_path):
|
| + os.makedirs(dir_path)
|
| + with open(file_path, 'w') as file_handle:
|
| + file_handle.write(
|
| + '<!DOCTYPE html><html>'
|
| + '<head><title>rebaseline_server</title></head>'
|
| + '<body><ul>')
|
| +
|
| + if _GM_SUMMARY_TYPES:
|
| + file_handle.write('<li>GM Expectations vs Actuals</li><ul>')
|
| + for summary_type in _GM_SUMMARY_TYPES:
|
| + file_handle.write(
|
| + '\n<li><a href="/{static_directive}/view.html#/view.html?'
|
| + 'resultsToLoad=/{results_directive}/{summary_type}">'
|
| + '{summary_type}</a></li>'.format(
|
| + results_directive=GET__PRECOMPUTED_RESULTS,
|
| + static_directive=GET__STATIC_CONTENTS,
|
| + summary_type=summary_type))
|
| + file_handle.write('</ul>')
|
| +
|
| + if config_pairs:
|
| + file_handle.write(
|
| + '\n<li>Comparing configs within actual GM results</li><ul>')
|
| + for config_pair in config_pairs:
|
| + file_handle.write('<li>%s vs %s:' % config_pair)
|
| + for summary_type in _GM_SUMMARY_TYPES:
|
| + file_handle.write(
|
| + ' <a href="/%s/view.html#/view.html?'
|
| + 'resultsToLoad=/%s/%s/%s-vs-%s_%s.json">%s</a>' % (
|
| + GET__STATIC_CONTENTS, GET__STATIC_CONTENTS,
|
| + GENERATED_JSON_SUBDIR, config_pair[0], config_pair[1],
|
| + summary_type, summary_type))
|
| + file_handle.write('</li>')
|
| + file_handle.write('</ul>')
|
| +
|
| + if _SKP_PLATFORMS:
|
| + file_handle.write('\n<li>Rendered SKPs:<ul>')
|
| + for builder in _SKP_PLATFORMS:
|
| + file_handle.write(
|
| + '\n<li><a href="../live-view.html#live-view.html?%s">' %
|
| + urllib.urlencode({
|
| + LIVE_PARAM__SET_A_SECTION:
|
| + gm_json.JSONKEY_EXPECTEDRESULTS,
|
| + LIVE_PARAM__SET_A_DIR:
|
| + posixpath.join(_SKP_BASE_REPO_URL, builder),
|
| + LIVE_PARAM__SET_B_SECTION:
|
| + gm_json.JSONKEY_ACTUALRESULTS,
|
| + LIVE_PARAM__SET_B_DIR:
|
| + posixpath.join(_SKP_BASE_GS_URL, builder),
|
| + }))
|
| + file_handle.write('expected vs actuals on %s</a></li>' % builder)
|
| + file_handle.write(
|
| + '\n<li><a href="../live-view.html#live-view.html?%s">' %
|
| + urllib.urlencode({
|
| + LIVE_PARAM__SET_A_SECTION:
|
| + gm_json.JSONKEY_ACTUALRESULTS,
|
| + LIVE_PARAM__SET_A_DIR:
|
| + posixpath.join(_SKP_BASE_GS_URL, _SKP_PLATFORMS[0]),
|
| + LIVE_PARAM__SET_B_SECTION:
|
| + gm_json.JSONKEY_ACTUALRESULTS,
|
| + LIVE_PARAM__SET_B_DIR:
|
| + posixpath.join(_SKP_BASE_GS_URL, _SKP_PLATFORMS[1]),
|
| + }))
|
| + file_handle.write('actuals on %s vs %s</a></li>' % (
|
| + _SKP_PLATFORMS[0], _SKP_PLATFORMS[1]))
|
| + file_handle.write('</li>')
|
| +
|
| + file_handle.write('\n</ul></body></html>')
|
| +
|
| +
|
| +class Server(object):
|
| + """ HTTP server for our HTML rebaseline viewer. """
|
| +
|
| + def __init__(self,
|
| + actuals_source,
|
| + actuals_dir=DEFAULT_ACTUALS_DIR,
|
| + json_filename=DEFAULT_JSON_FILENAME,
|
| + port=DEFAULT_PORT, export=False, editable=True,
|
| + reload_seconds=0, config_pairs=None, builder_regex_list=None,
|
| + boto_file_path=None,
|
| + imagediffdb_threads=imagediffdb.DEFAULT_NUM_WORKER_THREADS):
|
| + """
|
| + Args:
|
| + actuals_source: actuals_source.get_builders() ->
|
| + {builder:string -> [ bucket:string, path:string, generation:string ]}
|
| + If None, don't fetch new actual-results files
|
| + at all, just compare to whatever files are already in actuals_dir
|
| + actuals_dir: directory under which we will check out the latest actual
|
| + GM results
|
| + json_filename: basename of the JSON summary file to load for each builder
|
| + port: which TCP port to listen on for HTTP requests
|
| + export: whether to allow HTTP clients on other hosts to access this server
|
| + editable: whether HTTP clients are allowed to submit new GM baselines
|
| + (SKP baseline modifications are performed using an entirely different
|
| + mechanism, not affected by this parameter)
|
| + reload_seconds: polling interval with which to check for new results;
|
| + if 0, don't check for new results at all
|
| + config_pairs: List of (string, string) tuples; for each tuple, compare
|
| + actual results of these two configs. If None or empty,
|
| + don't compare configs at all.
|
| + builder_regex_list: List of regular expressions specifying which builders
|
| + we will process. If None, process all builders.
|
| + boto_file_path: Path to .boto file giving us credentials to access
|
| + Google Storage buckets; if None, we will only be able to access
|
| + public GS buckets.
|
| + imagediffdb_threads: How many threads to spin up within imagediffdb.
|
| + """
|
| + self._actuals_source = actuals_source
|
| + self._actuals_dir = actuals_dir
|
| + self._json_filename = json_filename
|
| + self._port = port
|
| + self._export = export
|
| + self._editable = editable
|
| + self._reload_seconds = reload_seconds
|
| + self._config_pairs = config_pairs or []
|
| + self._builder_regex_list = builder_regex_list
|
| + self.truncate_results = False
|
| +
|
| + if boto_file_path:
|
| + self._gs = gs_utils.GSUtils(boto_file_path=boto_file_path)
|
| + else:
|
| + self._gs = gs_utils.GSUtils()
|
| +
|
| + _create_index(
|
| + file_path=os.path.join(
|
| + PARENT_DIRECTORY, STATIC_CONTENTS_SUBDIR, GENERATED_HTML_SUBDIR,
|
| + "index.html"),
|
| + config_pairs=config_pairs)
|
| +
|
| + # Reentrant lock that must be held whenever updating EITHER of:
|
| + # 1. self._results
|
| + # 2. the expected or actual results on local disk
|
| + self.results_rlock = threading.RLock()
|
| +
|
| + # Create a single ImageDiffDB instance that is used by all our differs.
|
| + self._image_diff_db = imagediffdb.ImageDiffDB(
|
| + gs=self._gs,
|
| + storage_root=os.path.join(
|
| + PARENT_DIRECTORY, STATIC_CONTENTS_SUBDIR,
|
| + GENERATED_IMAGES_SUBDIR),
|
| + num_worker_threads=imagediffdb_threads)
|
| +
|
| + # This will be filled in by calls to update_results()
|
| + self._results = None
|
| +
|
| + @property
|
| + def results(self):
|
| + """ Returns the most recently generated results, or None if we don't have
|
| + any valid results (update_results() has not completed yet). """
|
| + return self._results
|
| +
|
| + @property
|
| + def image_diff_db(self):
|
| + """ Returns reference to our ImageDiffDB object."""
|
| + return self._image_diff_db
|
| +
|
| + @property
|
| + def gs(self):
|
| + """ Returns reference to our GSUtils object."""
|
| + return self._gs
|
| +
|
| + @property
|
| + def is_exported(self):
|
| + """ Returns true iff HTTP clients on other hosts are allowed to access
|
| + this server. """
|
| + return self._export
|
| +
|
| + @property
|
| + def is_editable(self):
|
| + """ True iff HTTP clients are allowed to submit new GM baselines.
|
| +
|
| + TODO(epoger): This only pertains to GM baselines; SKP baselines are
|
| + editable whenever expectations vs actuals are shown.
|
| + Once we move the GM baselines to use the same code as the SKP baselines,
|
| + we can delete this property.
|
| + """
|
| + return self._editable
|
| +
|
| + @property
|
| + def reload_seconds(self):
|
| + """ Returns the result reload period in seconds, or 0 if we don't reload
|
| + results. """
|
| + return self._reload_seconds
|
| +
|
| + def update_results(self, invalidate=False):
|
| + """ Create or update self._results, based on the latest expectations and
|
| + actuals.
|
| +
|
| + We hold self.results_rlock while we do this, to guarantee that no other
|
| + thread attempts to update either self._results or the underlying files at
|
| + the same time.
|
| +
|
| + Args:
|
| + invalidate: if True, invalidate self._results immediately upon entry;
|
| + otherwise, we will let readers see those results until we
|
| + replace them
|
| + """
|
| + with self.results_rlock:
|
| + if invalidate:
|
| + self._results = None
|
| +
|
| + if self._actuals_source:
|
| + logging.info(
|
| + 'Updating GM result summaries in %s from %s ...'
|
| + % (self._actuals_dir, self._actuals_source.description()))
|
| +
|
| + # Clean out actuals_dir first, in case some builders have gone away
|
| + # since we last ran.
|
| + if os.path.isdir(self._actuals_dir):
|
| + shutil.rmtree(self._actuals_dir)
|
| +
|
| + # Get the list of actuals we care about.
|
| + all_actuals = self._actuals_source.get_builders()
|
| +
|
| + if self._builder_regex_list:
|
| + matching_builders = []
|
| + for builder in all_actuals:
|
| + for regex in self._builder_regex_list:
|
| + if re.match(regex, builder):
|
| + matching_builders.append(builder)
|
| + break # go on to the next builder, no need to try more regexes
|
| + else:
|
| + matching_builders = all_actuals.keys()
|
| +
|
| + # Download the JSON file for each builder we care about.
|
| + #
|
| + # TODO(epoger): When this is a large number of builders, we would be
|
| + # better off downloading them in parallel!
|
| + for builder in matching_builders:
|
| + self._gs.download_file(
|
| + source_bucket=all_actuals[builder].bucket,
|
| + source_path=all_actuals[builder].path,
|
| + source_generation=all_actuals[builder].generation,
|
| + dest_path=os.path.join(self._actuals_dir, builder,
|
| + self._json_filename),
|
| + create_subdirs_if_needed=True)
|
| +
|
| + # We only update the expectations dir if the server was run with a
|
| + # nonzero --reload argument; otherwise, we expect the user to maintain
|
| + # her own expectations as she sees fit.
|
| + #
|
| + # Because the Skia repo is hosted using git, and git does not
|
| + # support updating a single directory tree, we have to update the entire
|
| + # repo checkout.
|
| + #
|
| + # Because Skia uses depot_tools, we have to update using "gclient sync"
|
| + # instead of raw git commands.
|
| + #
|
| + # TODO(epoger): Fetch latest expectations in some other way.
|
| + # Eric points out that our official documentation recommends an
|
| + # unmanaged Skia checkout, so "gclient sync" will not bring down updated
|
| + # expectations from origin/master-- you'd have to do a "git pull" of
|
| + # some sort instead.
|
| + # However, the live rebaseline_server at
|
| + # http://skia-tree-status.appspot.com/redirect/rebaseline-server (which
|
| + # is probably the only user of the --reload flag!) uses a managed
|
| + # checkout, so "gclient sync" works in that case.
|
| + # Probably the best idea is to avoid all of this nonsense by fetching
|
| + # updated expectations into a temp directory, and leaving the rest of
|
| + # the checkout alone. This could be done using "git show", or by
|
| + # downloading individual expectation JSON files from
|
| + # skia.googlesource.com .
|
| + if self._reload_seconds:
|
| + logging.info(
|
| + 'Updating expected GM results in %s by syncing Skia repo ...' %
|
| + compare_to_expectations.DEFAULT_EXPECTATIONS_DIR)
|
| + _run_command(['gclient', 'sync'], TRUNK_DIRECTORY)
|
| +
|
| + self._results = compare_to_expectations.ExpectationComparisons(
|
| + image_diff_db=self._image_diff_db,
|
| + actuals_root=self._actuals_dir,
|
| + diff_base_url=posixpath.join(
|
| + os.pardir, STATIC_CONTENTS_SUBDIR, GENERATED_IMAGES_SUBDIR),
|
| + builder_regex_list=self._builder_regex_list)
|
| +
|
| + json_dir = os.path.join(
|
| + PARENT_DIRECTORY, STATIC_CONTENTS_SUBDIR, GENERATED_JSON_SUBDIR)
|
| + if not os.path.isdir(json_dir):
|
| + os.makedirs(json_dir)
|
| +
|
| + for config_pair in self._config_pairs:
|
| + config_comparisons = compare_configs.ConfigComparisons(
|
| + configs=config_pair,
|
| + actuals_root=self._actuals_dir,
|
| + generated_images_root=os.path.join(
|
| + PARENT_DIRECTORY, STATIC_CONTENTS_SUBDIR,
|
| + GENERATED_IMAGES_SUBDIR),
|
| + diff_base_url=posixpath.join(
|
| + os.pardir, GENERATED_IMAGES_SUBDIR),
|
| + builder_regex_list=self._builder_regex_list)
|
| + for summary_type in _GM_SUMMARY_TYPES:
|
| + gm_json.WriteToFile(
|
| + config_comparisons.get_packaged_results_of_type(
|
| + results_type=summary_type),
|
| + os.path.join(
|
| + json_dir, '%s-vs-%s_%s.json' % (
|
| + config_pair[0], config_pair[1], summary_type)))
|
| +
|
| + def _result_loader(self, reload_seconds=0):
|
| + """ Call self.update_results(), either once or periodically.
|
| +
|
| + Params:
|
| + reload_seconds: integer; if nonzero, reload results at this interval
|
| + (in which case, this method will never return!)
|
| + """
|
| + self.update_results()
|
| + logging.info('Initial results loaded. Ready for requests on %s' % self._url)
|
| + if reload_seconds:
|
| + while True:
|
| + time.sleep(reload_seconds)
|
| + self.update_results()
|
| +
|
| + def run(self):
|
| + arg_tuple = (self._reload_seconds,) # start_new_thread needs a tuple,
|
| + # even though it holds just one param
|
| + thread.start_new_thread(self._result_loader, arg_tuple)
|
| +
|
| + if self._export:
|
| + server_address = ('', self._port)
|
| + host = _get_routable_ip_address()
|
| + if self._editable:
|
| + logging.warning('Running with combination of "export" and "editable" '
|
| + 'flags. Users on other machines will '
|
| + 'be able to modify your GM expectations!')
|
| + else:
|
| + host = '127.0.0.1'
|
| + server_address = (host, self._port)
|
| + # pylint: disable=W0201
|
| + http_server = BaseHTTPServer.HTTPServer(server_address, HTTPRequestHandler)
|
| + self._url = 'http://%s:%d' % (host, http_server.server_port)
|
| + logging.info('Listening for requests on %s' % self._url)
|
| + http_server.serve_forever()
|
| +
|
| +
|
| +class HTTPRequestHandler(BaseHTTPServer.BaseHTTPRequestHandler):
|
| + """ HTTP request handlers for various types of queries this server knows
|
| + how to handle (static HTML and Javascript, expected/actual results, etc.)
|
| + """
|
| + def do_GET(self):
|
| + """
|
| + Handles all GET requests, forwarding them to the appropriate
|
| + do_GET_* dispatcher.
|
| +
|
| + If we see any Exceptions, return a 404. This fixes http://skbug.com/2147
|
| + """
|
| + try:
|
| + logging.debug('do_GET: path="%s"' % self.path)
|
| + if self.path == '' or self.path == '/' or self.path == '/index.html' :
|
| + self.redirect_to('/%s/%s/index.html' % (
|
| + GET__STATIC_CONTENTS, GENERATED_HTML_SUBDIR))
|
| + return
|
| + if self.path == '/favicon.ico' :
|
| + self.redirect_to('/%s/favicon.ico' % GET__STATIC_CONTENTS)
|
| + return
|
| +
|
| + # All requests must be of this form:
|
| + # /dispatcher/remainder
|
| + # where 'dispatcher' indicates which do_GET_* dispatcher to run
|
| + # and 'remainder' is the remaining path sent to that dispatcher.
|
| + (dispatcher_name, remainder) = PATHSPLIT_RE.match(self.path).groups()
|
| + dispatchers = {
|
| + GET__LIVE_RESULTS: self.do_GET_live_results,
|
| + GET__PRECOMPUTED_RESULTS: self.do_GET_precomputed_results,
|
| + GET__PREFETCH_RESULTS: self.do_GET_prefetch_results,
|
| + GET__STATIC_CONTENTS: self.do_GET_static,
|
| + }
|
| + dispatcher = dispatchers[dispatcher_name]
|
| + dispatcher(remainder)
|
| + except:
|
| + self.send_error(404)
|
| + raise
|
| +
|
| + def do_GET_precomputed_results(self, results_type):
|
| + """ Handle a GET request for part of the precomputed _SERVER.results object.
|
| +
|
| + Args:
|
| + results_type: string indicating which set of results to return;
|
| + must be one of the results_mod.RESULTS_* constants
|
| + """
|
| + logging.debug('do_GET_precomputed_results: sending results of type "%s"' %
|
| + results_type)
|
| + # Since we must make multiple calls to the ExpectationComparisons object,
|
| + # grab a reference to it in case it is updated to point at a new
|
| + # ExpectationComparisons object within another thread.
|
| + #
|
| + # TODO(epoger): Rather than using a global variable for the handler
|
| + # to refer to the Server object, make Server a subclass of
|
| + # HTTPServer, and then it could be available to the handler via
|
| + # the handler's .server instance variable.
|
| + results_obj = _SERVER.results
|
| + if results_obj:
|
| + response_dict = results_obj.get_packaged_results_of_type(
|
| + results_type=results_type, reload_seconds=_SERVER.reload_seconds,
|
| + is_editable=_SERVER.is_editable, is_exported=_SERVER.is_exported)
|
| + else:
|
| + now = int(time.time())
|
| + response_dict = {
|
| + imagepairset.KEY__ROOT__HEADER: {
|
| + results_mod.KEY__HEADER__SCHEMA_VERSION: (
|
| + results_mod.VALUE__HEADER__SCHEMA_VERSION),
|
| + results_mod.KEY__HEADER__IS_STILL_LOADING: True,
|
| + results_mod.KEY__HEADER__TIME_UPDATED: now,
|
| + results_mod.KEY__HEADER__TIME_NEXT_UPDATE_AVAILABLE: (
|
| + now + RELOAD_INTERVAL_UNTIL_READY),
|
| + },
|
| + }
|
| + self.send_json_dict(response_dict)
|
| +
|
| + def _get_live_results_or_prefetch(self, url_remainder, prefetch_only=False):
|
| + """ Handle a GET request for live-generated image diff data.
|
| +
|
| + Args:
|
| + url_remainder: string indicating which image diffs to generate
|
| + prefetch_only: if True, the user isn't waiting around for results
|
| + """
|
| + param_dict = urlparse.parse_qs(url_remainder)
|
| + download_all_images = (
|
| + param_dict.get(LIVE_PARAM__DOWNLOAD_ONLY_DIFFERING, [''])[0].lower()
|
| + not in ['1', 'true'])
|
| + setA_dir = param_dict[LIVE_PARAM__SET_A_DIR][0]
|
| + setB_dir = param_dict[LIVE_PARAM__SET_B_DIR][0]
|
| + setA_section = self._validate_summary_section(
|
| + param_dict.get(LIVE_PARAM__SET_A_SECTION, [None])[0])
|
| + setB_section = self._validate_summary_section(
|
| + param_dict.get(LIVE_PARAM__SET_B_SECTION, [None])[0])
|
| +
|
| + # If the sets show expectations vs actuals, always show expectations on
|
| + # the left (setA).
|
| + if ((setA_section == gm_json.JSONKEY_ACTUALRESULTS) and
|
| + (setB_section == gm_json.JSONKEY_EXPECTEDRESULTS)):
|
| + setA_dir, setB_dir = setB_dir, setA_dir
|
| + setA_section, setB_section = setB_section, setA_section
|
| +
|
| + # Are we comparing some actuals against expectations stored in the repo?
|
| + # If so, we can allow the user to submit new baselines.
|
| + is_editable = (
|
| + (setA_section == gm_json.JSONKEY_EXPECTEDRESULTS) and
|
| + (setA_dir.startswith(compare_rendered_pictures.REPO_URL_PREFIX)) and
|
| + (setB_section == gm_json.JSONKEY_ACTUALRESULTS))
|
| +
|
| + results_obj = compare_rendered_pictures.RenderedPicturesComparisons(
|
| + setA_dir=setA_dir, setB_dir=setB_dir,
|
| + setA_section=setA_section, setB_section=setB_section,
|
| + image_diff_db=_SERVER.image_diff_db,
|
| + diff_base_url='/static/generated-images',
|
| + gs=_SERVER.gs, truncate_results=_SERVER.truncate_results,
|
| + prefetch_only=prefetch_only, download_all_images=download_all_images)
|
| + if prefetch_only:
|
| + self.send_response(200)
|
| + else:
|
| + self.send_json_dict(results_obj.get_packaged_results_of_type(
|
| + results_type=results_mod.KEY__HEADER__RESULTS_ALL,
|
| + is_editable=is_editable))
|
| +
|
| + def do_GET_live_results(self, url_remainder):
|
| + """ Handle a GET request for live-generated image diff data.
|
| +
|
| + Args:
|
| + url_remainder: string indicating which image diffs to generate
|
| + """
|
| + logging.debug('do_GET_live_results: url_remainder="%s"' % url_remainder)
|
| + self._get_live_results_or_prefetch(
|
| + url_remainder=url_remainder, prefetch_only=False)
|
| +
|
| + def do_GET_prefetch_results(self, url_remainder):
|
| + """ Prefetch image diff data for a future do_GET_live_results() call.
|
| +
|
| + Args:
|
| + url_remainder: string indicating which image diffs to generate
|
| + """
|
| + logging.debug('do_GET_prefetch_results: url_remainder="%s"' % url_remainder)
|
| + self._get_live_results_or_prefetch(
|
| + url_remainder=url_remainder, prefetch_only=True)
|
| +
|
| + def do_GET_static(self, path):
|
| + """ Handle a GET request for a file under STATIC_CONTENTS_SUBDIR .
|
| + Only allow serving of files within STATIC_CONTENTS_SUBDIR that is a
|
| + filesystem sibling of this script.
|
| +
|
| + Args:
|
| + path: path to file (within STATIC_CONTENTS_SUBDIR) to retrieve
|
| + """
|
| + # Strip arguments ('?resultsToLoad=all') from the path
|
| + path = urlparse.urlparse(path).path
|
| +
|
| + logging.debug('do_GET_static: sending file "%s"' % path)
|
| + static_dir = os.path.realpath(os.path.join(
|
| + PARENT_DIRECTORY, STATIC_CONTENTS_SUBDIR))
|
| + full_path = os.path.realpath(os.path.join(static_dir, path))
|
| + if full_path.startswith(static_dir):
|
| + self.send_file(full_path)
|
| + else:
|
| + logging.error(
|
| + 'Attempted do_GET_static() of path [%s] outside of static dir [%s]'
|
| + % (full_path, static_dir))
|
| + self.send_error(404)
|
| +
|
| + def do_POST(self):
|
| + """ Handles all POST requests, forwarding them to the appropriate
|
| + do_POST_* dispatcher. """
|
| + # All requests must be of this form:
|
| + # /dispatcher
|
| + # where 'dispatcher' indicates which do_POST_* dispatcher to run.
|
| + logging.debug('do_POST: path="%s"' % self.path)
|
| + normpath = posixpath.normpath(self.path)
|
| + dispatchers = {
|
| + '/edits': self.do_POST_edits,
|
| + '/live-edits': self.do_POST_live_edits,
|
| + }
|
| + try:
|
| + dispatcher = dispatchers[normpath]
|
| + dispatcher()
|
| + except:
|
| + self.send_error(404)
|
| + raise
|
| +
|
| + def do_POST_edits(self):
|
| + """ Handle a POST request with modifications to GM expectations, in this
|
| + format:
|
| +
|
| + {
|
| + KEY__EDITS__OLD_RESULTS_TYPE: 'all', # type of results that the client
|
| + # loaded and then made
|
| + # modifications to
|
| + KEY__EDITS__OLD_RESULTS_HASH: 39850913, # hash of results when the client
|
| + # loaded them (ensures that the
|
| + # client and server apply
|
| + # modifications to the same base)
|
| + KEY__EDITS__MODIFICATIONS: [
|
| + # as needed by compare_to_expectations.edit_expectations()
|
| + ...
|
| + ],
|
| + }
|
| +
|
| + Raises an Exception if there were any problems.
|
| + """
|
| + if not _SERVER.is_editable:
|
| + raise Exception('this server is not running in --editable mode')
|
| +
|
| + content_type = self.headers[_HTTP_HEADER_CONTENT_TYPE]
|
| + if content_type != 'application/json;charset=UTF-8':
|
| + raise Exception('unsupported %s [%s]' % (
|
| + _HTTP_HEADER_CONTENT_TYPE, content_type))
|
| +
|
| + content_length = int(self.headers[_HTTP_HEADER_CONTENT_LENGTH])
|
| + json_data = self.rfile.read(content_length)
|
| + data = json.loads(json_data)
|
| + logging.debug('do_POST_edits: received new GM expectations data [%s]' %
|
| + data)
|
| +
|
| + # Update the results on disk with the information we received from the
|
| + # client.
|
| + # We must hold _SERVER.results_rlock while we do this, to guarantee that
|
| + # no other thread updates expectations (from the Skia repo) while we are
|
| + # updating them (using the info we received from the client).
|
| + with _SERVER.results_rlock:
|
| + oldResultsType = data[KEY__EDITS__OLD_RESULTS_TYPE]
|
| + oldResults = _SERVER.results.get_results_of_type(oldResultsType)
|
| + oldResultsHash = str(hash(repr(
|
| + oldResults[imagepairset.KEY__ROOT__IMAGEPAIRS])))
|
| + if oldResultsHash != data[KEY__EDITS__OLD_RESULTS_HASH]:
|
| + raise Exception('results of type "%s" changed while the client was '
|
| + 'making modifications. The client should reload the '
|
| + 'results and submit the modifications again.' %
|
| + oldResultsType)
|
| + _SERVER.results.edit_expectations(data[KEY__EDITS__MODIFICATIONS])
|
| +
|
| + # Read the updated results back from disk.
|
| + # We can do this in a separate thread; we should return our success message
|
| + # to the UI as soon as possible.
|
| + thread.start_new_thread(_SERVER.update_results, (True,))
|
| + self.send_response(200)
|
| +
|
| + def do_POST_live_edits(self):
|
| + """ Handle a POST request with modifications to SKP expectations, in this
|
| + format:
|
| +
|
| + {
|
| + KEY__LIVE_EDITS__SET_A_DESCRIPTIONS: {
|
| + # setA descriptions from the original data
|
| + },
|
| + KEY__LIVE_EDITS__SET_B_DESCRIPTIONS: {
|
| + # setB descriptions from the original data
|
| + },
|
| + KEY__LIVE_EDITS__MODIFICATIONS: [
|
| + # as needed by writable_expectations.modify()
|
| + ],
|
| + }
|
| +
|
| + Raises an Exception if there were any problems.
|
| + """
|
| + content_type = self.headers[_HTTP_HEADER_CONTENT_TYPE]
|
| + if content_type != 'application/json;charset=UTF-8':
|
| + raise Exception('unsupported %s [%s]' % (
|
| + _HTTP_HEADER_CONTENT_TYPE, content_type))
|
| +
|
| + content_length = int(self.headers[_HTTP_HEADER_CONTENT_LENGTH])
|
| + json_data = self.rfile.read(content_length)
|
| + data = json.loads(json_data)
|
| + logging.debug('do_POST_live_edits: received new GM expectations data [%s]' %
|
| + data)
|
| + with writable_expectations_mod.WritableExpectations(
|
| + data[KEY__LIVE_EDITS__SET_A_DESCRIPTIONS]) as writable_expectations:
|
| + writable_expectations.modify(data[KEY__LIVE_EDITS__MODIFICATIONS])
|
| + diffs = writable_expectations.get_diffs()
|
| + # TODO(stephana): Move to a simpler web framework so we don't have to
|
| + # call these functions. See http://skbug.com/2856 ('rebaseline_server:
|
| + # Refactor server to use a simple web framework')
|
| + self.send_response(200)
|
| + self.send_header('Content-type', 'text/plain')
|
| + self.end_headers()
|
| + self.wfile.write(diffs)
|
| +
|
| + def redirect_to(self, url):
|
| + """ Redirect the HTTP client to a different url.
|
| +
|
| + Args:
|
| + url: URL to redirect the HTTP client to
|
| + """
|
| + self.send_response(301)
|
| + self.send_header('Location', url)
|
| + self.end_headers()
|
| +
|
| + def send_file(self, path):
|
| + """ Send the contents of the file at this path, with a mimetype based
|
| + on the filename extension.
|
| +
|
| + Args:
|
| + path: path of file whose contents to send to the HTTP client
|
| + """
|
| + # Grab the extension if there is one
|
| + extension = os.path.splitext(path)[1]
|
| + if len(extension) >= 1:
|
| + extension = extension[1:]
|
| +
|
| + # Determine the MIME type of the file from its extension
|
| + mime_type = MIME_TYPE_MAP.get(extension, MIME_TYPE_MAP[''])
|
| +
|
| + # Open the file and send it over HTTP
|
| + if os.path.isfile(path):
|
| + with open(path, 'rb') as sending_file:
|
| + self.send_response(200)
|
| + self.send_header('Content-type', mime_type)
|
| + self.end_headers()
|
| + self.wfile.write(sending_file.read())
|
| + else:
|
| + self.send_error(404)
|
| +
|
| + def send_json_dict(self, json_dict):
|
| + """ Send the contents of this dictionary in JSON format, with a JSON
|
| + mimetype.
|
| +
|
| + Args:
|
| + json_dict: dictionary to send
|
| + """
|
| + self.send_response(200)
|
| + self.send_header('Content-type', 'application/json')
|
| + self.end_headers()
|
| + json.dump(json_dict, self.wfile)
|
| +
|
| + def _validate_summary_section(self, section_name):
|
| + """Validates the section we have been requested to read within JSON summary.
|
| +
|
| + Args:
|
| + section_name: which section of the JSON summary file has been requested
|
| +
|
| + Returns: the validated section name
|
| +
|
| + Raises: Exception if an invalid section_name was requested.
|
| + """
|
| + if section_name not in compare_rendered_pictures.ALLOWED_SECTION_NAMES:
|
| + raise Exception('requested section name "%s" not in allowed list %s' % (
|
| + section_name, compare_rendered_pictures.ALLOWED_SECTION_NAMES))
|
| + return section_name
|
| +
|
| +
|
| +def main():
|
| + logging.basicConfig(format='%(asctime)s %(levelname)s %(message)s',
|
| + datefmt='%m/%d/%Y %H:%M:%S',
|
| + level=logging.INFO)
|
| + parser = argparse.ArgumentParser()
|
| + parser.add_argument('--actuals-dir',
|
| + help=('Directory into which we will check out the latest '
|
| + 'actual GM results. If this directory does not '
|
| + 'exist, it will be created. Defaults to %(default)s'),
|
| + default=DEFAULT_ACTUALS_DIR)
|
| + parser.add_argument('--boto',
|
| + help=('Path to .boto file giving us credentials to access '
|
| + 'Google Storage buckets. If not specified, we will '
|
| + 'only be able to access public GS buckets (and thus '
|
| + 'won\'t be able to download SKP images).'),
|
| + default='')
|
| + # TODO(epoger): Before https://codereview.chromium.org/310093003 ,
|
| + # when this tool downloaded the JSON summaries from skia-autogen,
|
| + # it had an --actuals-revision the caller could specify to download
|
| + # actual results as of a specific point in time. We should add similar
|
| + # functionality when retrieving the summaries from Google Storage.
|
| + parser.add_argument('--builders', metavar='BUILDER_REGEX', nargs='+',
|
| + help=('Only process builders matching these regular '
|
| + 'expressions. If unspecified, process all '
|
| + 'builders.'))
|
| + parser.add_argument('--compare-configs', action='store_true',
|
| + help=('In addition to generating differences between '
|
| + 'expectations and actuals, also generate '
|
| + 'differences between these config pairs: '
|
| + + str(CONFIG_PAIRS_TO_COMPARE)))
|
| + parser.add_argument('--editable', action='store_true',
|
| + help=('Allow HTTP clients to submit new GM baselines; '
|
| + 'SKP baselines can be edited regardless of this '
|
| + 'setting.'))
|
| + parser.add_argument('--export', action='store_true',
|
| + help=('Instead of only allowing access from HTTP clients '
|
| + 'on localhost, allow HTTP clients on other hosts '
|
| + 'to access this server. WARNING: doing so will '
|
| + 'allow users on other hosts to modify your '
|
| + 'GM expectations, if combined with --editable.'))
|
| + parser.add_argument('--rietveld-issue',
|
| + help=('Download json_filename files from latest trybot'
|
| + 'runs on this codereview.chromium.org issue.'
|
| + 'Overrides --gm-summaries-bucket.'))
|
| + parser.add_argument('--gm-summaries-bucket',
|
| + help=('Google Cloud Storage bucket to download '
|
| + 'JSON_FILENAME files from. '
|
| + 'Defaults to %(default)s ; if set to '
|
| + 'empty string, just compare to actual-results '
|
| + 'already found in ACTUALS_DIR.'),
|
| + default=DEFAULT_GM_SUMMARIES_BUCKET)
|
| + parser.add_argument('--json-filename',
|
| + help=('JSON summary filename to read for each builder; '
|
| + 'defaults to %(default)s.'),
|
| + default=DEFAULT_JSON_FILENAME)
|
| + parser.add_argument('--port', type=int,
|
| + help=('Which TCP port to listen on for HTTP requests; '
|
| + 'defaults to %(default)s'),
|
| + default=DEFAULT_PORT)
|
| + parser.add_argument('--reload', type=int,
|
| + help=('How often (a period in seconds) to update the '
|
| + 'results. If specified, both expected and actual '
|
| + 'results will be updated by running "gclient sync" '
|
| + 'on your Skia checkout as a whole. '
|
| + 'By default, we do not reload at all, and you '
|
| + 'must restart the server to pick up new data.'),
|
| + default=0)
|
| + parser.add_argument('--threads', type=int,
|
| + help=('How many parallel threads we use to download '
|
| + 'images and generate diffs; defaults to '
|
| + '%(default)s'),
|
| + default=imagediffdb.DEFAULT_NUM_WORKER_THREADS)
|
| + parser.add_argument('--truncate', action='store_true',
|
| + help=('FOR TESTING ONLY: truncate the set of images we '
|
| + 'process, to speed up testing.'))
|
| + args = parser.parse_args()
|
| + if args.compare_configs:
|
| + config_pairs = CONFIG_PAIRS_TO_COMPARE
|
| + else:
|
| + config_pairs = None
|
| +
|
| + if args.rietveld_issue:
|
| + actuals_source = download_actuals.RietveldIssueActuals(args.rietveld_issue,
|
| + args.json_filename)
|
| + else:
|
| + actuals_source = download_actuals.TipOfTreeActuals(args.gm_summaries_bucket,
|
| + args.json_filename)
|
| +
|
| + global _SERVER
|
| + _SERVER = Server(actuals_source,
|
| + actuals_dir=args.actuals_dir,
|
| + json_filename=args.json_filename,
|
| + port=args.port, export=args.export, editable=args.editable,
|
| + reload_seconds=args.reload, config_pairs=config_pairs,
|
| + builder_regex_list=args.builders, boto_file_path=args.boto,
|
| + imagediffdb_threads=args.threads)
|
| + if args.truncate:
|
| + _SERVER.truncate_results = True
|
| + _SERVER.run()
|
| +
|
| +
|
| +if __name__ == '__main__':
|
| + main()
|
|
|