Index: gm/rebaseline_server/download_actuals.py |
diff --git a/gm/rebaseline_server/download_actuals.py b/gm/rebaseline_server/download_actuals.py |
new file mode 100755 |
index 0000000000000000000000000000000000000000..3f3f640f1216a1a5b041d809a8f316e6e11bec3a |
--- /dev/null |
+++ b/gm/rebaseline_server/download_actuals.py |
@@ -0,0 +1,201 @@ |
+#!/usr/bin/python |
+ |
+""" |
+Copyright 2014 Google Inc. |
+ |
+Use of this source code is governed by a BSD-style license that can be |
+found in the LICENSE file. |
+ |
+Download actual GM results for a particular builder. |
+""" |
+ |
+# System-level imports |
+import contextlib |
+import optparse |
+import os |
+import posixpath |
+import re |
+import shutil |
+import sys |
+import urllib |
+import urllib2 |
+import urlparse |
+ |
+# Imports from within Skia |
+# |
+# We need to add the 'gm' and 'tools' directories, so that we can import |
+# gm_json.py and buildbot_globals.py. |
+# |
+# Make sure that these dirs are in the PYTHONPATH, but add them at the *end* |
+# so any dirs that are already in the PYTHONPATH will be preferred. |
+# |
+# TODO(epoger): Is it OK for this to depend on the 'tools' dir, given that |
+# the tools dir is dependent on the 'gm' dir (to import gm_json.py)? |
+TRUNK_DIRECTORY = os.path.dirname(os.path.dirname(os.path.dirname(__file__))) |
+GM_DIRECTORY = os.path.join(TRUNK_DIRECTORY, 'gm') |
+TOOLS_DIRECTORY = os.path.join(TRUNK_DIRECTORY, 'tools') |
+if GM_DIRECTORY not in sys.path: |
+ sys.path.append(GM_DIRECTORY) |
+if TOOLS_DIRECTORY not in sys.path: |
+ sys.path.append(TOOLS_DIRECTORY) |
+import buildbot_globals |
+import gm_json |
+ |
+DEFAULT_ACTUALS_BASE_URL = posixpath.join( |
+ buildbot_globals.Get('autogen_svn_url'), 'gm-actual') |
+DEFAULT_JSON_FILENAME = 'actual-results.json' |
+ |
+ |
+class Download(object): |
+ |
+ def __init__(self, actuals_base_url=DEFAULT_ACTUALS_BASE_URL, |
+ json_filename=DEFAULT_JSON_FILENAME, |
+ gm_actuals_root_url=gm_json.GM_ACTUALS_ROOT_HTTP_URL): |
+ """ |
+ Args: |
+ actuals_base_url: URL pointing at the root directory |
+ containing all actual-results.json files, e.g., |
+ http://domain.name/path/to/dir OR |
+ file:///absolute/path/to/localdir |
+ json_filename: The JSON filename to read from within each directory. |
+ gm_actuals_root_url: Base URL under which the actually-generated-by-bots |
+ GM images are stored. |
+ """ |
+ self._actuals_base_url = actuals_base_url |
+ self._json_filename = json_filename |
+ self._gm_actuals_root_url = gm_actuals_root_url |
+ self._image_filename_re = re.compile(gm_json.IMAGE_FILENAME_PATTERN) |
+ |
+ def fetch(self, builder_name, dest_dir): |
+ """ Downloads actual GM results for a particular builder. |
+ |
+ Args: |
+ builder_name: which builder to download results of |
+ dest_dir: path to directory where the image files will be written; |
+ if the directory does not exist yet, it will be created |
+ |
+ TODO(epoger): Display progress info. Right now, it can take a long time |
+ to download all of the results, and there is no indication of progress. |
+ |
+ TODO(epoger): Download multiple images in parallel to speed things up. |
+ """ |
+ json_url = posixpath.join(self._actuals_base_url, builder_name, |
+ self._json_filename) |
+ json_contents = urllib2.urlopen(json_url).read() |
+ results_dict = gm_json.LoadFromString(json_contents) |
+ |
+ actual_results_dict = results_dict[gm_json.JSONKEY_ACTUALRESULTS] |
+ for result_type in sorted(actual_results_dict.keys()): |
+ results_of_this_type = actual_results_dict[result_type] |
+ if not results_of_this_type: |
+ continue |
+ for image_name in sorted(results_of_this_type.keys()): |
+ (test, config) = self._image_filename_re.match(image_name).groups() |
+ (hash_type, hash_digest) = results_of_this_type[image_name] |
+ source_url = gm_json.CreateGmActualUrl( |
+ test_name=test, hash_type=hash_type, hash_digest=hash_digest, |
+ gm_actuals_root_url=self._gm_actuals_root_url) |
+ dest_path = os.path.join(dest_dir, config, test + '.png') |
+ copy_contents(source_url=source_url, dest_path=dest_path, |
+ create_subdirs_if_needed=True) |
+ |
+ |
+def create_filepath_url(filepath): |
+ """ Returns a file:/// URL pointing at the given filepath on local disk. |
+ |
+ For now, this is only used by unittests, but I anticipate it being useful |
+ in production, as a way for developers to run rebaseline_server over locally |
+ generated images. |
+ |
+ TODO(epoger): Move this function, and copy_contents(), into a shared |
+ utility module. They are generally useful. |
+ |
+ Args: |
+ filepath: string; path to a file on local disk (may be absolute or relative, |
+ and the file does not need to exist) |
+ |
+ Returns: |
+ A file:/// URL pointing at the file. Regardless of whether filepath was |
+ specified as a relative or absolute path, the URL will contain an |
+ absolute path to the file. |
+ |
+ Raises: |
+ An Exception, if filepath is already a URL. |
+ """ |
+ if urlparse.urlparse(filepath).scheme: |
+ raise Exception('"%s" is already a URL' % filepath) |
+ return urlparse.urljoin( |
+ 'file:', urllib.pathname2url(os.path.abspath(filepath))) |
+ |
+ |
+def copy_contents(source_url, dest_path, create_subdirs_if_needed=False): |
+ """ Copies the full contents of the URL 'source_url' into |
+ filepath 'dest_path'. |
+ |
+ Args: |
+ source_url: string; complete URL to read from |
+ dest_path: string; complete filepath to write to (may be absolute or |
+ relative) |
+ create_subdirs_if_needed: boolean; whether to create subdirectories as |
+ needed to create dest_path |
+ |
+ Raises: |
+ Some subclass of Exception if unable to read source_url or write dest_path. |
+ """ |
+ if create_subdirs_if_needed: |
+ dest_dir = os.path.dirname(dest_path) |
+ if not os.path.exists(dest_dir): |
+ os.makedirs(dest_dir) |
+ with contextlib.closing(urllib.urlopen(source_url)) as source_handle: |
+ with open(dest_path, 'wb') as dest_handle: |
+ shutil.copyfileobj(fsrc=source_handle, fdst=dest_handle) |
+ |
+ |
+def main(): |
+ parser = optparse.OptionParser() |
+ required_params = [] |
+ parser.add_option('--actuals-base-url', |
+ action='store', type='string', |
+ default=DEFAULT_ACTUALS_BASE_URL, |
+ help=('Base URL from which to read files containing JSON ' |
+ 'summaries of actual GM results; defaults to ' |
+ '"%default". To get a specific revision (useful for ' |
+ 'trybots) replace "svn" with "svn-history/r123".')) |
+ # TODO(epoger): Rather than telling the user to run "svn ls" to get the list |
+ # of builders, add a --list-builders option that will print the list. |
+ required_params.append('builder') |
+ parser.add_option('--builder', |
+ action='store', type='string', |
+ help=('REQUIRED: Which builder to download results for. ' |
+ 'To see a list of builders, run "svn ls %s".' % |
+ DEFAULT_ACTUALS_BASE_URL)) |
+ required_params.append('dest_dir') |
+ parser.add_option('--dest-dir', |
+ action='store', type='string', |
+ help=('REQUIRED: Directory where all images should be ' |
+ 'written. If this directory does not exist yet, it ' |
+ 'will be created.')) |
+ parser.add_option('--json-filename', |
+ action='store', type='string', |
+ default=DEFAULT_JSON_FILENAME, |
+ help=('JSON summary filename to read for each builder; ' |
+ 'defaults to "%default".')) |
+ (params, remaining_args) = parser.parse_args() |
+ |
+ # Make sure all required options were set, |
+ # and that there were no items left over in the command line. |
+ for required_param in required_params: |
+ if not getattr(params, required_param): |
+ raise Exception('required option \'%s\' was not set' % required_param) |
+ if len(remaining_args) is not 0: |
+ raise Exception('extra items specified in the command line: %s' % |
+ remaining_args) |
+ |
+ downloader = Download(actuals_base_url=params.actuals_base_url) |
+ downloader.fetch(builder_name=params.builder, |
+ dest_dir=params.dest_dir) |
+ |
+ |
+ |
+if __name__ == '__main__': |
+ main() |