Index: gm/rebaseline_server/download_actuals.py |
diff --git a/gm/rebaseline_server/download_actuals.py b/gm/rebaseline_server/download_actuals.py |
index 636958be6b545c2251628a3b0f65f2d2db0c25d1..2f92898fd4b0451feb3dd1895157f21b51283919 100755 |
--- a/gm/rebaseline_server/download_actuals.py |
+++ b/gm/rebaseline_server/download_actuals.py |
@@ -10,44 +10,19 @@ Download actual GM results for a particular builder. |
""" |
# System-level imports |
-import contextlib |
import optparse |
import os |
import posixpath |
import re |
-import shutil |
-import sys |
-import urllib |
import urllib2 |
-import urlparse |
# Imports from within Skia |
-# |
-# We need to add the 'gm' and 'tools' directories, so that we can import |
-# gm_json.py and buildbot_globals.py. |
-# |
-# Make sure that these dirs are in the PYTHONPATH, but add them at the *end* |
-# so any dirs that are already in the PYTHONPATH will be preferred. |
-# |
-# TODO(epoger): Is it OK for this to depend on the 'tools' dir, given that |
-# the tools dir is dependent on the 'gm' dir (to import gm_json.py)? |
-TRUNK_DIRECTORY = os.path.dirname(os.path.dirname(os.path.dirname(__file__))) |
-GM_DIRECTORY = os.path.join(TRUNK_DIRECTORY, 'gm') |
-TOOLS_DIRECTORY = os.path.join(TRUNK_DIRECTORY, 'tools') |
-if GM_DIRECTORY not in sys.path: |
- sys.path.append(GM_DIRECTORY) |
-if TOOLS_DIRECTORY not in sys.path: |
- sys.path.append(TOOLS_DIRECTORY) |
+import fix_pythonpath # must do this first |
+from pyutils import gs_utils |
+from pyutils import url_utils |
import buildbot_globals |
import gm_json |
-# Imports from third-party code |
-APICLIENT_DIRECTORY = os.path.join( |
- TRUNK_DIRECTORY, 'third_party', 'externals', 'google-api-python-client') |
-if APICLIENT_DIRECTORY not in sys.path: |
- sys.path.append(APICLIENT_DIRECTORY) |
-from googleapiclient.discovery import build as build_service |
- |
GM_SUMMARIES_BUCKET = buildbot_globals.Get('gm_summaries_bucket') |
DEFAULT_ACTUALS_BASE_URL = ( |
@@ -105,98 +80,19 @@ class Download(object): |
test_name=test, hash_type=hash_type, hash_digest=hash_digest, |
gm_actuals_root_url=self._gm_actuals_root_url) |
dest_path = os.path.join(dest_dir, config, test + '.png') |
- # TODO(epoger): To speed this up, we should only download files that |
- # we don't already have on local disk. |
- copy_contents(source_url=source_url, dest_path=dest_path, |
- create_subdirs_if_needed=True) |
- |
- |
-def create_filepath_url(filepath): |
- """ Returns a file:/// URL pointing at the given filepath on local disk. |
+ url_utils.copy_contents(source_url=source_url, dest_path=dest_path, |
+ create_subdirs_if_needed=True) |
- For now, this is only used by unittests, but I anticipate it being useful |
- in production, as a way for developers to run rebaseline_server over locally |
- generated images. |
- TODO(epoger): Move this function, and copy_contents(), into a shared |
- utility module. They are generally useful. |
+def get_builders_list(summaries_bucket=GM_SUMMARIES_BUCKET): |
+ """ Returns the list of builders we have actual results for. |
Args: |
- filepath: string; path to a file on local disk (may be absolute or relative, |
- and the file does not need to exist) |
- |
- Returns: |
- A file:/// URL pointing at the file. Regardless of whether filepath was |
- specified as a relative or absolute path, the URL will contain an |
- absolute path to the file. |
- |
- Raises: |
- An Exception, if filepath is already a URL. |
+ summaries_bucket: Google Cloud Storage bucket containing the summary |
+ JSON files |
""" |
- if urlparse.urlparse(filepath).scheme: |
- raise Exception('"%s" is already a URL' % filepath) |
- return urlparse.urljoin( |
- 'file:', urllib.pathname2url(os.path.abspath(filepath))) |
- |
- |
-def copy_contents(source_url, dest_path, create_subdirs_if_needed=False): |
- """ Copies the full contents of the URL 'source_url' into |
- filepath 'dest_path'. |
- |
- Args: |
- source_url: string; complete URL to read from |
- dest_path: string; complete filepath to write to (may be absolute or |
- relative) |
- create_subdirs_if_needed: boolean; whether to create subdirectories as |
- needed to create dest_path |
- |
- Raises: |
- Some subclass of Exception if unable to read source_url or write dest_path. |
- """ |
- if create_subdirs_if_needed: |
- dest_dir = os.path.dirname(dest_path) |
- if not os.path.exists(dest_dir): |
- os.makedirs(dest_dir) |
- with contextlib.closing(urllib.urlopen(source_url)) as source_handle: |
- with open(dest_path, 'wb') as dest_handle: |
- shutil.copyfileobj(fsrc=source_handle, fdst=dest_handle) |
- |
- |
-def gcs_list_bucket_contents(bucket, subdir=None): |
- """ Returns files in the Google Cloud Storage bucket as a (dirs, files) tuple. |
- |
- Uses the API documented at |
- https://developers.google.com/storage/docs/json_api/v1/objects/list |
- |
- Args: |
- bucket: name of the Google Storage bucket |
- subdir: directory within the bucket to list, or None for root directory |
- """ |
- # The GCS command relies on the subdir name (if any) ending with a slash. |
- if subdir and not subdir.endswith('/'): |
- subdir += '/' |
- subdir_length = len(subdir) if subdir else 0 |
- |
- storage = build_service('storage', 'v1') |
- command = storage.objects().list( |
- bucket=bucket, delimiter='/', fields='items(name),prefixes', |
- prefix=subdir) |
- results = command.execute() |
- |
- # The GCS command returned two subdicts: |
- # prefixes: the full path of every directory within subdir, with trailing '/' |
- # items: property dict for each file object within subdir |
- # (including 'name', which is full path of the object) |
- dirs = [] |
- for dir_fullpath in results.get('prefixes', []): |
- dir_basename = dir_fullpath[subdir_length:] |
- dirs.append(dir_basename[:-1]) # strip trailing slash |
- files = [] |
- for file_properties in results.get('items', []): |
- file_fullpath = file_properties['name'] |
- file_basename = file_fullpath[subdir_length:] |
- files.append(file_basename) |
- return (dirs, files) |
+ dirs, _ = gs_utils.list_bucket_contents(bucket=GM_SUMMARIES_BUCKET) |
+ return dirs |
def main(): |
@@ -234,8 +130,7 @@ def main(): |
(params, remaining_args) = parser.parse_args() |
if params.list_builders: |
- dirs, _ = gcs_list_bucket_contents(bucket=GM_SUMMARIES_BUCKET) |
- print '\n'.join(dirs) |
+ print '\n'.join(get_builders_list()) |
return |
# Make sure all required options were set, |