| Index: gm/rebaseline_server/compare_rendered_pictures.py
|
| diff --git a/gm/rebaseline_server/compare_rendered_pictures.py b/gm/rebaseline_server/compare_rendered_pictures.py
|
| deleted file mode 100755
|
| index 73cb36b430fa6e0816e3feec892f93616c28e3ac..0000000000000000000000000000000000000000
|
| --- a/gm/rebaseline_server/compare_rendered_pictures.py
|
| +++ /dev/null
|
| @@ -1,504 +0,0 @@
|
| -#!/usr/bin/python
|
| -
|
| -"""
|
| -Copyright 2014 Google Inc.
|
| -
|
| -Use of this source code is governed by a BSD-style license that can be
|
| -found in the LICENSE file.
|
| -
|
| -Compare results of two render_pictures runs.
|
| -
|
| -TODO(epoger): Start using this module to compare ALL images (whether they
|
| -were generated from GMs or SKPs), and rename it accordingly.
|
| -"""
|
| -
|
| -# System-level imports
|
| -import logging
|
| -import os
|
| -import shutil
|
| -import subprocess
|
| -import tempfile
|
| -import time
|
| -
|
| -# Must fix up PYTHONPATH before importing from within Skia
|
| -import rs_fixpypath # pylint: disable=W0611
|
| -
|
| -# Imports from within Skia
|
| -from py.utils import git_utils
|
| -from py.utils import gs_utils
|
| -from py.utils import url_utils
|
| -import buildbot_globals
|
| -import column
|
| -import gm_json
|
| -import imagediffdb
|
| -import imagepair
|
| -import imagepairset
|
| -import results
|
| -
|
| -# URL under which all render_pictures images can be found in Google Storage.
|
| -#
|
| -# TODO(epoger): In order to allow live-view of GMs and other images, read this
|
| -# from the input summary files, or allow the caller to set it within the
|
| -# GET_live_results call.
|
| -DEFAULT_IMAGE_BASE_GS_URL = 'gs://' + buildbot_globals.Get('skp_images_bucket')
|
| -
|
| -# Column descriptors, and display preferences for them.
|
| -COLUMN__RESULT_TYPE = results.KEY__EXTRACOLUMNS__RESULT_TYPE
|
| -COLUMN__SOURCE_SKP = 'sourceSkpFile'
|
| -COLUMN__TILED_OR_WHOLE = 'tiledOrWhole'
|
| -COLUMN__TILENUM = 'tilenum'
|
| -COLUMN__BUILDER_A = 'builderA'
|
| -COLUMN__RENDER_MODE_A = 'renderModeA'
|
| -COLUMN__BUILDER_B = 'builderB'
|
| -COLUMN__RENDER_MODE_B = 'renderModeB'
|
| -# Known values for some of those columns.
|
| -COLUMN__TILED_OR_WHOLE__TILED = 'tiled'
|
| -COLUMN__TILED_OR_WHOLE__WHOLE = 'whole'
|
| -
|
| -FREEFORM_COLUMN_IDS = [
|
| - COLUMN__SOURCE_SKP,
|
| - COLUMN__TILENUM,
|
| -]
|
| -ORDERED_COLUMN_IDS = [
|
| - COLUMN__RESULT_TYPE,
|
| - COLUMN__SOURCE_SKP,
|
| - COLUMN__TILED_OR_WHOLE,
|
| - COLUMN__TILENUM,
|
| - COLUMN__BUILDER_A,
|
| - COLUMN__RENDER_MODE_A,
|
| - COLUMN__BUILDER_B,
|
| - COLUMN__RENDER_MODE_B,
|
| -]
|
| -
|
| -# A special "repo:" URL type that we use to refer to Skia repo contents.
|
| -# (Useful for comparing against expectations files we store in our repo.)
|
| -REPO_URL_PREFIX = 'repo:'
|
| -REPO_BASEPATH = os.path.abspath(os.path.join(
|
| - os.path.dirname(os.path.abspath(__file__)), os.pardir, os.pardir))
|
| -
|
| -# Which sections within a JSON summary file can contain results.
|
| -ALLOWED_SECTION_NAMES = [
|
| - gm_json.JSONKEY_ACTUALRESULTS,
|
| - gm_json.JSONKEY_EXPECTEDRESULTS,
|
| -]
|
| -
|
| -
|
| -class RenderedPicturesComparisons(results.BaseComparisons):
|
| - """Loads results from multiple render_pictures runs into an ImagePairSet.
|
| - """
|
| -
|
| - def __init__(self,
|
| - setA_dir, setB_dir,
|
| - setA_section, setB_section,
|
| - image_diff_db,
|
| - image_base_gs_url=DEFAULT_IMAGE_BASE_GS_URL, diff_base_url=None,
|
| - setA_label=None, setB_label=None,
|
| - gs=None, truncate_results=False, prefetch_only=False,
|
| - download_all_images=False):
|
| - """Constructor: downloads images and generates diffs.
|
| -
|
| - Once the object has been created (which may take a while), you can call its
|
| - get_packaged_results_of_type() method to quickly retrieve the results...
|
| - unless you have set prefetch_only to True, in which case we will
|
| - asynchronously warm up the ImageDiffDB cache but not fill in self._results.
|
| -
|
| - Args:
|
| - setA_dir: root directory to copy all JSON summaries from, and to use as
|
| - setA within the comparisons. This directory may be specified as a
|
| - gs:// URL, special "repo:" URL, or local filepath.
|
| - setB_dir: root directory to copy all JSON summaries from, and to use as
|
| - setB within the comparisons. This directory may be specified as a
|
| - gs:// URL, special "repo:" URL, or local filepath.
|
| - setA_section: which section within setA to examine; must be one of
|
| - ALLOWED_SECTION_NAMES
|
| - setB_section: which section within setB to examine; must be one of
|
| - ALLOWED_SECTION_NAMES
|
| - image_diff_db: ImageDiffDB instance
|
| - image_base_gs_url: "gs://" URL pointing at the Google Storage bucket/dir
|
| - under which all render_pictures result images can
|
| - be found; this will be used to read images for comparison within
|
| - this code, and included in the ImagePairSet (as an HTTP URL) so its
|
| - consumers know where to download the images from
|
| - diff_base_url: base URL within which the client should look for diff
|
| - images; if not specified, defaults to a "file:///" URL representation
|
| - of image_diff_db's storage_root
|
| - setA_label: description to use for results in setA; if None, will be
|
| - set to a reasonable default
|
| - setB_label: description to use for results in setB; if None, will be
|
| - set to a reasonable default
|
| - gs: instance of GSUtils object we can use to download summary files
|
| - truncate_results: FOR MANUAL TESTING: if True, truncate the set of images
|
| - we process, to speed up testing.
|
| - prefetch_only: if True, return the new object as quickly as possible
|
| - with empty self._results (just queue up all the files to process,
|
| - don't wait around for them to be processed and recorded); otherwise,
|
| - block until the results have been assembled and recorded in
|
| - self._results.
|
| - download_all_images: if True, download all images, even if we don't
|
| - need them to generate diffs. This will take much longer to complete,
|
| - but is useful for warming up the bitmap cache on local disk.
|
| - """
|
| - super(RenderedPicturesComparisons, self).__init__()
|
| - self._image_diff_db = image_diff_db
|
| - self._image_base_gs_url = image_base_gs_url
|
| - self._diff_base_url = (
|
| - diff_base_url or
|
| - url_utils.create_filepath_url(image_diff_db.storage_root))
|
| - self._gs = gs
|
| - self.truncate_results = truncate_results
|
| - self._prefetch_only = prefetch_only
|
| - self._download_all_images = download_all_images
|
| -
|
| - # If we are comparing two different section types, we can use those
|
| - # as the default labels for setA and setB.
|
| - if setA_section != setB_section:
|
| - self._setA_label = setA_label or setA_section
|
| - self._setB_label = setB_label or setB_section
|
| - else:
|
| - self._setA_label = setA_label or 'setA'
|
| - self._setB_label = setB_label or 'setB'
|
| -
|
| - tempdir = tempfile.mkdtemp()
|
| - try:
|
| - setA_root = os.path.join(tempdir, 'setA')
|
| - setB_root = os.path.join(tempdir, 'setB')
|
| - # TODO(stephana): There is a potential race condition here... we copy
|
| - # the contents out of the source_dir, and THEN we get the commithash
|
| - # of source_dir. If source_dir points at a git checkout, and that
|
| - # checkout is updated (by a different thread/process) during this
|
| - # operation, then the contents and commithash will be out of sync.
|
| - self._copy_dir_contents(source_dir=setA_dir, dest_dir=setA_root)
|
| - setA_repo_revision = self._get_repo_revision(source_dir=setA_dir)
|
| - self._copy_dir_contents(source_dir=setB_dir, dest_dir=setB_root)
|
| - setB_repo_revision = self._get_repo_revision(source_dir=setB_dir)
|
| -
|
| - self._setA_descriptions = {
|
| - results.KEY__SET_DESCRIPTIONS__DIR: setA_dir,
|
| - results.KEY__SET_DESCRIPTIONS__REPO_REVISION: setA_repo_revision,
|
| - results.KEY__SET_DESCRIPTIONS__SECTION: setA_section,
|
| - }
|
| - self._setB_descriptions = {
|
| - results.KEY__SET_DESCRIPTIONS__DIR: setB_dir,
|
| - results.KEY__SET_DESCRIPTIONS__REPO_REVISION: setB_repo_revision,
|
| - results.KEY__SET_DESCRIPTIONS__SECTION: setB_section,
|
| - }
|
| -
|
| - time_start = int(time.time())
|
| - self._results = self._load_result_pairs(
|
| - setA_root=setA_root, setB_root=setB_root,
|
| - setA_section=setA_section, setB_section=setB_section)
|
| - if self._results:
|
| - self._timestamp = int(time.time())
|
| - logging.info('Number of download file collisions: %s' %
|
| - imagediffdb.global_file_collisions)
|
| - logging.info('Results complete; took %d seconds.' %
|
| - (self._timestamp - time_start))
|
| - finally:
|
| - shutil.rmtree(tempdir)
|
| -
|
| - def _load_result_pairs(self, setA_root, setB_root,
|
| - setA_section, setB_section):
|
| - """Loads all JSON image summaries from 2 directory trees and compares them.
|
| -
|
| - TODO(stephana): This method is only called from within __init__(); it might
|
| - make more sense to just roll the content of this method into __init__().
|
| -
|
| - Args:
|
| - setA_root: root directory containing JSON summaries of rendering results
|
| - setB_root: root directory containing JSON summaries of rendering results
|
| - setA_section: which section (gm_json.JSONKEY_ACTUALRESULTS or
|
| - gm_json.JSONKEY_EXPECTEDRESULTS) to load from the summaries in setA
|
| - setB_section: which section (gm_json.JSONKEY_ACTUALRESULTS or
|
| - gm_json.JSONKEY_EXPECTEDRESULTS) to load from the summaries in setB
|
| -
|
| - Returns the summary of all image diff results (or None, depending on
|
| - self._prefetch_only).
|
| - """
|
| - logging.info('Reading JSON image summaries from dirs %s and %s...' % (
|
| - setA_root, setB_root))
|
| - setA_dicts = self.read_dicts_from_root(setA_root)
|
| - setB_dicts = self.read_dicts_from_root(setB_root)
|
| - logging.info('Comparing summary dicts...')
|
| -
|
| - all_image_pairs = imagepairset.ImagePairSet(
|
| - descriptions=(self._setA_label, self._setB_label),
|
| - diff_base_url=self._diff_base_url)
|
| - failing_image_pairs = imagepairset.ImagePairSet(
|
| - descriptions=(self._setA_label, self._setB_label),
|
| - diff_base_url=self._diff_base_url)
|
| -
|
| - # Override settings for columns that should be filtered using freeform text.
|
| - for column_id in FREEFORM_COLUMN_IDS:
|
| - factory = column.ColumnHeaderFactory(
|
| - header_text=column_id, use_freeform_filter=True)
|
| - all_image_pairs.set_column_header_factory(
|
| - column_id=column_id, column_header_factory=factory)
|
| - failing_image_pairs.set_column_header_factory(
|
| - column_id=column_id, column_header_factory=factory)
|
| -
|
| - all_image_pairs.ensure_extra_column_values_in_summary(
|
| - column_id=COLUMN__RESULT_TYPE, values=[
|
| - results.KEY__RESULT_TYPE__FAILED,
|
| - results.KEY__RESULT_TYPE__NOCOMPARISON,
|
| - results.KEY__RESULT_TYPE__SUCCEEDED,
|
| - ])
|
| - failing_image_pairs.ensure_extra_column_values_in_summary(
|
| - column_id=COLUMN__RESULT_TYPE, values=[
|
| - results.KEY__RESULT_TYPE__FAILED,
|
| - results.KEY__RESULT_TYPE__NOCOMPARISON,
|
| - ])
|
| -
|
| - logging.info('Starting to add imagepairs to queue.')
|
| - self._image_diff_db.log_queue_size_if_changed(limit_verbosity=False)
|
| -
|
| - union_dict_paths = sorted(set(setA_dicts.keys() + setB_dicts.keys()))
|
| - num_union_dict_paths = len(union_dict_paths)
|
| - dict_num = 0
|
| - for dict_path in union_dict_paths:
|
| - dict_num += 1
|
| - logging.info(
|
| - 'Asynchronously requesting pixel diffs for dict #%d of %d, "%s"...' %
|
| - (dict_num, num_union_dict_paths, dict_path))
|
| -
|
| - dictA = self.get_default(setA_dicts, None, dict_path)
|
| - self._validate_dict_version(dictA)
|
| - dictA_results = self.get_default(dictA, {}, setA_section)
|
| -
|
| - dictB = self.get_default(setB_dicts, None, dict_path)
|
| - self._validate_dict_version(dictB)
|
| - dictB_results = self.get_default(dictB, {}, setB_section)
|
| -
|
| - image_A_base_url = self.get_default(
|
| - setA_dicts, self._image_base_gs_url, dict_path,
|
| - gm_json.JSONKEY_IMAGE_BASE_GS_URL)
|
| - image_B_base_url = self.get_default(
|
| - setB_dicts, self._image_base_gs_url, dict_path,
|
| - gm_json.JSONKEY_IMAGE_BASE_GS_URL)
|
| -
|
| - # get the builders and render modes for each set
|
| - builder_A = self.get_default(dictA, None,
|
| - gm_json.JSONKEY_DESCRIPTIONS,
|
| - gm_json.JSONKEY_DESCRIPTIONS_BUILDER)
|
| - render_mode_A = self.get_default(dictA, None,
|
| - gm_json.JSONKEY_DESCRIPTIONS,
|
| - gm_json.JSONKEY_DESCRIPTIONS_RENDER_MODE)
|
| - builder_B = self.get_default(dictB, None,
|
| - gm_json.JSONKEY_DESCRIPTIONS,
|
| - gm_json.JSONKEY_DESCRIPTIONS_BUILDER)
|
| - render_mode_B = self.get_default(dictB, None,
|
| - gm_json.JSONKEY_DESCRIPTIONS,
|
| - gm_json.JSONKEY_DESCRIPTIONS_RENDER_MODE)
|
| -
|
| - skp_names = sorted(set(dictA_results.keys() + dictB_results.keys()))
|
| - # Just for manual testing... truncate to an arbitrary subset.
|
| - if self.truncate_results:
|
| - skp_names = skp_names[1:3]
|
| - for skp_name in skp_names:
|
| - imagepairs_for_this_skp = []
|
| -
|
| - whole_image_A = self.get_default(
|
| - dictA_results, None,
|
| - skp_name, gm_json.JSONKEY_SOURCE_WHOLEIMAGE)
|
| - whole_image_B = self.get_default(
|
| - dictB_results, None,
|
| - skp_name, gm_json.JSONKEY_SOURCE_WHOLEIMAGE)
|
| -
|
| - imagepairs_for_this_skp.append(self._create_image_pair(
|
| - image_dict_A=whole_image_A, image_dict_B=whole_image_B,
|
| - image_A_base_url=image_A_base_url,
|
| - image_B_base_url=image_B_base_url,
|
| - builder_A=builder_A, render_mode_A=render_mode_A,
|
| - builder_B=builder_B, render_mode_B=render_mode_B,
|
| - source_json_file=dict_path,
|
| - source_skp_name=skp_name, tilenum=None))
|
| -
|
| - tiled_images_A = self.get_default(
|
| - dictA_results, [],
|
| - skp_name, gm_json.JSONKEY_SOURCE_TILEDIMAGES)
|
| - tiled_images_B = self.get_default(
|
| - dictB_results, [],
|
| - skp_name, gm_json.JSONKEY_SOURCE_TILEDIMAGES)
|
| - if tiled_images_A or tiled_images_B:
|
| - num_tiles_A = len(tiled_images_A)
|
| - num_tiles_B = len(tiled_images_B)
|
| - num_tiles = max(num_tiles_A, num_tiles_B)
|
| - for tile_num in range(num_tiles):
|
| - imagepairs_for_this_skp.append(self._create_image_pair(
|
| - image_dict_A=(tiled_images_A[tile_num]
|
| - if tile_num < num_tiles_A else None),
|
| - image_dict_B=(tiled_images_B[tile_num]
|
| - if tile_num < num_tiles_B else None),
|
| - image_A_base_url=image_A_base_url,
|
| - image_B_base_url=image_B_base_url,
|
| - builder_A=builder_A, render_mode_A=render_mode_A,
|
| - builder_B=builder_B, render_mode_B=render_mode_B,
|
| - source_json_file=dict_path,
|
| - source_skp_name=skp_name, tilenum=tile_num))
|
| -
|
| - for one_imagepair in imagepairs_for_this_skp:
|
| - if one_imagepair:
|
| - all_image_pairs.add_image_pair(one_imagepair)
|
| - result_type = one_imagepair.extra_columns_dict\
|
| - [COLUMN__RESULT_TYPE]
|
| - if result_type != results.KEY__RESULT_TYPE__SUCCEEDED:
|
| - failing_image_pairs.add_image_pair(one_imagepair)
|
| -
|
| - logging.info('Finished adding imagepairs to queue.')
|
| - self._image_diff_db.log_queue_size_if_changed(limit_verbosity=False)
|
| -
|
| - if self._prefetch_only:
|
| - return None
|
| - else:
|
| - return {
|
| - results.KEY__HEADER__RESULTS_ALL: all_image_pairs.as_dict(
|
| - column_ids_in_order=ORDERED_COLUMN_IDS),
|
| - results.KEY__HEADER__RESULTS_FAILURES: failing_image_pairs.as_dict(
|
| - column_ids_in_order=ORDERED_COLUMN_IDS),
|
| - }
|
| -
|
| - def _validate_dict_version(self, result_dict):
|
| - """Raises Exception if the dict is not the type/version we know how to read.
|
| -
|
| - Args:
|
| - result_dict: dictionary holding output of render_pictures; if None,
|
| - this method will return without raising an Exception
|
| - """
|
| - # TODO(stephana): These values should be defined as constants somewhere,
|
| - # to be kept in sync between this file and writable_expectations.py
|
| - expected_header_type = 'ChecksummedImages'
|
| - expected_header_revision = 1
|
| -
|
| - if result_dict == None:
|
| - return
|
| - header = result_dict[gm_json.JSONKEY_HEADER]
|
| - header_type = header[gm_json.JSONKEY_HEADER_TYPE]
|
| - if header_type != expected_header_type:
|
| - raise Exception('expected header_type "%s", but got "%s"' % (
|
| - expected_header_type, header_type))
|
| - header_revision = header[gm_json.JSONKEY_HEADER_REVISION]
|
| - if header_revision != expected_header_revision:
|
| - raise Exception('expected header_revision %d, but got %d' % (
|
| - expected_header_revision, header_revision))
|
| -
|
| - def _create_image_pair(self, image_dict_A, image_dict_B,
|
| - image_A_base_url, image_B_base_url,
|
| - builder_A, render_mode_A,
|
| - builder_B, render_mode_B,
|
| - source_json_file,
|
| - source_skp_name, tilenum):
|
| - """Creates an ImagePair object for this pair of images.
|
| -
|
| - Args:
|
| - image_dict_A: dict with JSONKEY_IMAGE_* keys, or None if no image
|
| - image_dict_B: dict with JSONKEY_IMAGE_* keys, or None if no image
|
| - image_A_base_url: base URL for image A
|
| - image_B_base_url: base URL for image B
|
| - builder_A: builder that created image set A or None if unknow
|
| - render_mode_A: render mode used to generate image set A or None if
|
| - unknown.
|
| - builder_B: builder that created image set A or None if unknow
|
| - render_mode_B: render mode used to generate image set A or None if
|
| - unknown.
|
| - source_json_file: string; relative path of the JSON file where this
|
| - result came from, within setA and setB.
|
| - source_skp_name: string; name of the source SKP file
|
| - tilenum: which tile, or None if a wholeimage
|
| -
|
| - Returns:
|
| - An ImagePair object, or None if both image_dict_A and image_dict_B are
|
| - None.
|
| - """
|
| - if (not image_dict_A) and (not image_dict_B):
|
| - return None
|
| -
|
| - def _checksum_and_relative_url(dic):
|
| - if dic:
|
| - return ((dic[gm_json.JSONKEY_IMAGE_CHECKSUMALGORITHM],
|
| - int(dic[gm_json.JSONKEY_IMAGE_CHECKSUMVALUE])),
|
| - dic[gm_json.JSONKEY_IMAGE_FILEPATH])
|
| - else:
|
| - return None, None
|
| -
|
| - imageA_checksum, imageA_relative_url = _checksum_and_relative_url(
|
| - image_dict_A)
|
| - imageB_checksum, imageB_relative_url = _checksum_and_relative_url(
|
| - image_dict_B)
|
| -
|
| - if not imageA_checksum:
|
| - result_type = results.KEY__RESULT_TYPE__NOCOMPARISON
|
| - elif not imageB_checksum:
|
| - result_type = results.KEY__RESULT_TYPE__NOCOMPARISON
|
| - elif imageA_checksum == imageB_checksum:
|
| - result_type = results.KEY__RESULT_TYPE__SUCCEEDED
|
| - else:
|
| - result_type = results.KEY__RESULT_TYPE__FAILED
|
| -
|
| - extra_columns_dict = {
|
| - COLUMN__RESULT_TYPE: result_type,
|
| - COLUMN__SOURCE_SKP: source_skp_name,
|
| - COLUMN__BUILDER_A: builder_A,
|
| - COLUMN__RENDER_MODE_A: render_mode_A,
|
| - COLUMN__BUILDER_B: builder_B,
|
| - COLUMN__RENDER_MODE_B: render_mode_B,
|
| - }
|
| - if tilenum == None:
|
| - extra_columns_dict[COLUMN__TILED_OR_WHOLE] = COLUMN__TILED_OR_WHOLE__WHOLE
|
| - extra_columns_dict[COLUMN__TILENUM] = 'N/A'
|
| - else:
|
| - extra_columns_dict[COLUMN__TILED_OR_WHOLE] = COLUMN__TILED_OR_WHOLE__TILED
|
| - extra_columns_dict[COLUMN__TILENUM] = str(tilenum)
|
| -
|
| - try:
|
| - return imagepair.ImagePair(
|
| - image_diff_db=self._image_diff_db,
|
| - imageA_base_url=image_A_base_url,
|
| - imageB_base_url=image_B_base_url,
|
| - imageA_relative_url=imageA_relative_url,
|
| - imageB_relative_url=imageB_relative_url,
|
| - extra_columns=extra_columns_dict,
|
| - source_json_file=source_json_file,
|
| - download_all_images=self._download_all_images)
|
| - except (KeyError, TypeError):
|
| - logging.exception(
|
| - 'got exception while creating ImagePair for'
|
| - ' urlPair=("%s","%s"), source_skp_name="%s", tilenum="%s"' % (
|
| - imageA_relative_url, imageB_relative_url, source_skp_name,
|
| - tilenum))
|
| - return None
|
| -
|
| - def _copy_dir_contents(self, source_dir, dest_dir):
|
| - """Copy all contents of source_dir into dest_dir, recursing into subdirs.
|
| -
|
| - Args:
|
| - source_dir: path to source dir (GS URL, local filepath, or a special
|
| - "repo:" URL type that points at a file within our Skia checkout)
|
| - dest_dir: path to destination dir (local filepath)
|
| -
|
| - The copy operates as a "merge with overwrite": any files in source_dir will
|
| - be "overlaid" on top of the existing content in dest_dir. Existing files
|
| - with the same names will be overwritten.
|
| - """
|
| - if gs_utils.GSUtils.is_gs_url(source_dir):
|
| - (bucket, path) = gs_utils.GSUtils.split_gs_url(source_dir)
|
| - self._gs.download_dir_contents(source_bucket=bucket, source_dir=path,
|
| - dest_dir=dest_dir)
|
| - elif source_dir.lower().startswith(REPO_URL_PREFIX):
|
| - repo_dir = os.path.join(REPO_BASEPATH, source_dir[len(REPO_URL_PREFIX):])
|
| - shutil.copytree(repo_dir, dest_dir)
|
| - else:
|
| - shutil.copytree(source_dir, dest_dir)
|
| -
|
| - def _get_repo_revision(self, source_dir):
|
| - """Get the commit hash of source_dir, IF it refers to a git checkout.
|
| -
|
| - Args:
|
| - source_dir: path to source dir (GS URL, local filepath, or a special
|
| - "repo:" URL type that points at a file within our Skia checkout;
|
| - only the "repo:" URL type will have a commit hash.
|
| - """
|
| - if source_dir.lower().startswith(REPO_URL_PREFIX):
|
| - repo_dir = os.path.join(REPO_BASEPATH, source_dir[len(REPO_URL_PREFIX):])
|
| - return subprocess.check_output(
|
| - args=[git_utils.GIT, 'rev-parse', 'HEAD'], cwd=repo_dir).strip()
|
| - else:
|
| - return None
|
|
|