| Index: gm/rebaseline_server/imagediffdb.py
|
| diff --git a/gm/rebaseline_server/imagediffdb.py b/gm/rebaseline_server/imagediffdb.py
|
| index 89f9fef319a4045909553c7c840bbc19fb1004f0..fbe71211401f812775bc7fe6a55bdacda971e13c 100644
|
| --- a/gm/rebaseline_server/imagediffdb.py
|
| +++ b/gm/rebaseline_server/imagediffdb.py
|
| @@ -11,12 +11,16 @@ Calulate differences between image pairs, and store them in a database.
|
|
|
| # System-level imports
|
| import contextlib
|
| +import errno
|
| import json
|
| import logging
|
| import os
|
| +import Queue
|
| import re
|
| import shutil
|
| import tempfile
|
| +import threading
|
| +import time
|
| import urllib
|
|
|
| # Must fix up PYTHONPATH before importing from within Skia
|
| @@ -24,11 +28,16 @@ import fix_pythonpath # pylint: disable=W0611
|
|
|
| # Imports from within Skia
|
| import find_run_binary
|
| +from py.utils import gs_utils
|
| +
|
|
|
| SKPDIFF_BINARY = find_run_binary.find_path_to_program('skpdiff')
|
|
|
| DEFAULT_IMAGE_SUFFIX = '.png'
|
| DEFAULT_IMAGES_SUBDIR = 'images'
|
| +# TODO(epoger): Figure out a better default number of threads; for now,
|
| +# using a conservative default value.
|
| +DEFAULT_NUM_WORKER_THREADS = 1
|
|
|
| DISALLOWED_FILEPATH_CHAR_REGEX = re.compile('[^\w\-]')
|
|
|
| @@ -42,11 +51,20 @@ KEY__DIFFERENCES__NUM_DIFF_PIXELS = 'numDifferingPixels'
|
| KEY__DIFFERENCES__PERCENT_DIFF_PIXELS = 'percentDifferingPixels'
|
| KEY__DIFFERENCES__PERCEPTUAL_DIFF = 'perceptualDifference'
|
|
|
| +# Special values within ImageDiffDB._diff_dict
|
| +_DIFFRECORD_FAILED = 'failed'
|
| +_DIFFRECORD_PENDING = 'pending'
|
| +
|
| +# Temporary variable to keep track of how many times we download
|
| +# the same file in multiple threads.
|
| +# TODO(epoger): Delete this, once we see that the number stays close to 0.
|
| +global_file_collisions = 0
|
| +
|
|
|
| class DiffRecord(object):
|
| """ Record of differences between two images. """
|
|
|
| - def __init__(self, storage_root,
|
| + def __init__(self, gs, storage_root,
|
| expected_image_url, expected_image_locator,
|
| actual_image_url, actual_image_locator,
|
| expected_images_subdir=DEFAULT_IMAGES_SUBDIR,
|
| @@ -55,18 +73,16 @@ class DiffRecord(object):
|
| """Download this pair of images (unless we already have them on local disk),
|
| and prepare a DiffRecord for them.
|
|
|
| - TODO(epoger): Make this asynchronously download images, rather than blocking
|
| - until the images have been downloaded and processed.
|
| -
|
| Args:
|
| + gs: instance of GSUtils object we can use to download images
|
| storage_root: root directory on local disk within which we store all
|
| images
|
| - expected_image_url: file or HTTP url from which we will download the
|
| + expected_image_url: file, GS, or HTTP url from which we will download the
|
| expected image
|
| expected_image_locator: a unique ID string under which we will store the
|
| expected image within storage_root (probably including a checksum to
|
| guarantee uniqueness)
|
| - actual_image_url: file or HTTP url from which we will download the
|
| + actual_image_url: file, GS, or HTTP url from which we will download the
|
| actual image
|
| actual_image_locator: a unique ID string under which we will store the
|
| actual image within storage_root (probably including a checksum to
|
| @@ -79,8 +95,6 @@ class DiffRecord(object):
|
| actual_image_locator = _sanitize_locator(actual_image_locator)
|
|
|
| # Download the expected/actual images, if we don't have them already.
|
| - # TODO(rmistry): Add a parameter that just tries to use already-present
|
| - # image files rather than downloading them.
|
| expected_image_file = os.path.join(
|
| storage_root, expected_images_subdir,
|
| str(expected_image_locator) + image_suffix)
|
| @@ -88,13 +102,13 @@ class DiffRecord(object):
|
| storage_root, actual_images_subdir,
|
| str(actual_image_locator) + image_suffix)
|
| try:
|
| - _download_file(expected_image_file, expected_image_url)
|
| + _download_file(gs, expected_image_file, expected_image_url)
|
| except Exception:
|
| logging.exception('unable to download expected_image_url %s to file %s' %
|
| (expected_image_url, expected_image_file))
|
| raise
|
| try:
|
| - _download_file(actual_image_file, actual_image_url)
|
| + _download_file(gs, actual_image_file, actual_image_url)
|
| except Exception:
|
| logging.exception('unable to download actual_image_url %s to file %s' %
|
| (actual_image_url, actual_image_file))
|
| @@ -112,8 +126,12 @@ class DiffRecord(object):
|
| actual_img = os.path.join(storage_root, actual_images_subdir,
|
| str(actual_image_locator) + image_suffix)
|
|
|
| - # TODO: Call skpdiff ONCE for all image pairs, instead of calling it
|
| - # repeatedly. This will allow us to parallelize a lot more work.
|
| + # TODO(epoger): Consider calling skpdiff ONCE for all image pairs,
|
| + # instead of calling it separately for each image pair.
|
| + # Pro: we'll incur less overhead from making repeated system calls,
|
| + # spinning up the skpdiff binary, etc.
|
| + # Con: we would have to wait until all image pairs were loaded before
|
| + # generating any of the diffs?
|
| find_run_binary.run_command(
|
| [SKPDIFF_BINARY, '-p', expected_img, actual_img,
|
| '--jsonp', 'false',
|
| @@ -211,16 +229,71 @@ class ImageDiffDB(object):
|
| """ Calculates differences between image pairs, maintaining a database of
|
| them for download."""
|
|
|
| - def __init__(self, storage_root):
|
| + def __init__(self, storage_root, gs=None,
|
| + num_worker_threads=DEFAULT_NUM_WORKER_THREADS):
|
| """
|
| Args:
|
| storage_root: string; root path within the DB will store all of its stuff
|
| + gs: instance of GSUtils object we can use to download images
|
| + num_worker_threads: how many threads that download images and
|
| + generate diffs simultaneously
|
| """
|
| self._storage_root = storage_root
|
| + self._gs = gs
|
|
|
| # Dictionary of DiffRecords, keyed by (expected_image_locator,
|
| # actual_image_locator) tuples.
|
| + # Values can also be _DIFFRECORD_PENDING, _DIFFRECORD_FAILED.
|
| + #
|
| + # Any thread that modifies _diff_dict must first acquire
|
| + # _diff_dict_writelock!
|
| + #
|
| + # TODO(epoger): Disk is limitless, but RAM is not... so, we should probably
|
| + # remove items from self._diff_dict if they haven't been accessed for a
|
| + # long time. We can always regenerate them by diffing the images we
|
| + # previously downloaded to local disk.
|
| + # I guess we should figure out how expensive it is to download vs diff the
|
| + # image pairs... if diffing them is expensive too, we can write these
|
| + # _diff_dict objects out to disk if there's too many to hold in RAM.
|
| + # Or we could use virtual memory to handle that automatically.
|
| self._diff_dict = {}
|
| + self._diff_dict_writelock = threading.RLock()
|
| +
|
| + # Set up the queue for asynchronously loading DiffRecords, and start the
|
| + # worker threads reading from it.
|
| + self._tasks_queue = Queue.Queue(maxsize=2*num_worker_threads)
|
| + self._workers = []
|
| + for i in range(num_worker_threads):
|
| + worker = threading.Thread(target=self.worker, args=(i,))
|
| + worker.daemon = True
|
| + worker.start()
|
| + self._workers.append(worker)
|
| +
|
| + def worker(self, worker_num):
|
| + """Launch a worker thread that pulls tasks off self._tasks_queue.
|
| +
|
| + Args:
|
| + worker_num: (integer) which worker this is
|
| + """
|
| + while True:
|
| + params = self._tasks_queue.get()
|
| + key, expected_image_url, actual_image_url = params
|
| + try:
|
| + diff_record = DiffRecord(
|
| + self._gs, self._storage_root,
|
| + expected_image_url=expected_image_url,
|
| + expected_image_locator=key[0],
|
| + actual_image_url=actual_image_url,
|
| + actual_image_locator=key[1])
|
| + except Exception:
|
| + logging.exception(
|
| + 'exception while creating DiffRecord for key %s' % str(key))
|
| + diff_record = _DIFFRECORD_FAILED
|
| + self._diff_dict_writelock.acquire()
|
| + try:
|
| + self._diff_dict[key] = diff_record
|
| + finally:
|
| + self._diff_dict_writelock.release()
|
|
|
| @property
|
| def storage_root(self):
|
| @@ -229,24 +302,21 @@ class ImageDiffDB(object):
|
| def add_image_pair(self,
|
| expected_image_url, expected_image_locator,
|
| actual_image_url, actual_image_locator):
|
| - """Download this pair of images (unless we already have them on local disk),
|
| - and prepare a DiffRecord for them.
|
| + """Asynchronously prepare a DiffRecord for a pair of images.
|
| +
|
| + This method will return quickly; calls to get_diff_record() will block
|
| + until the DiffRecord is available (or we have given up on creating it).
|
|
|
| - TODO(epoger): Make this asynchronously download images, rather than blocking
|
| - until the images have been downloaded and processed.
|
| - When we do that, we should probably add a new method that will block
|
| - until all of the images have been downloaded and processed. Otherwise,
|
| - we won't know when it's safe to start calling get_diff_record().
|
| - jcgregorio notes: maybe just make ImageDiffDB thread-safe and create a
|
| - thread-pool/worker queue at a higher level that just uses ImageDiffDB?
|
| + If we already have a DiffRecord for this particular image pair, no work
|
| + will be done.
|
|
|
| Args:
|
| - expected_image_url: file or HTTP url from which we will download the
|
| + expected_image_url: file, GS, or HTTP url from which we will download the
|
| expected image
|
| expected_image_locator: a unique ID string under which we will store the
|
| expected image within storage_root (probably including a checksum to
|
| guarantee uniqueness)
|
| - actual_image_url: file or HTTP url from which we will download the
|
| + actual_image_url: file, GS, or HTTP url from which we will download the
|
| actual image
|
| actual_image_locator: a unique ID string under which we will store the
|
| actual image within storage_root (probably including a checksum to
|
| @@ -255,49 +325,96 @@ class ImageDiffDB(object):
|
| expected_image_locator = _sanitize_locator(expected_image_locator)
|
| actual_image_locator = _sanitize_locator(actual_image_locator)
|
| key = (expected_image_locator, actual_image_locator)
|
| - if not key in self._diff_dict:
|
| - try:
|
| - new_diff_record = DiffRecord(
|
| - self._storage_root,
|
| - expected_image_url=expected_image_url,
|
| - expected_image_locator=expected_image_locator,
|
| - actual_image_url=actual_image_url,
|
| - actual_image_locator=actual_image_locator)
|
| - except Exception:
|
| - # If we can't create a real DiffRecord for this (expected, actual) pair,
|
| - # store None and the UI will show whatever information we DO have.
|
| - # Fixes http://skbug.com/2368 .
|
| - logging.exception(
|
| - 'got exception while creating a DiffRecord for '
|
| - 'expected_image_url=%s , actual_image_url=%s; returning None' % (
|
| - expected_image_url, actual_image_url))
|
| - new_diff_record = None
|
| - self._diff_dict[key] = new_diff_record
|
| + must_add_to_queue = False
|
| +
|
| + self._diff_dict_writelock.acquire()
|
| + try:
|
| + if not key in self._diff_dict:
|
| + # If we have already requested a diff between these two images,
|
| + # we don't need to request it again.
|
| + must_add_to_queue = True
|
| + self._diff_dict[key] = _DIFFRECORD_PENDING
|
| + finally:
|
| + self._diff_dict_writelock.release()
|
| +
|
| + if must_add_to_queue:
|
| + self._tasks_queue.put((key, expected_image_url, actual_image_url))
|
|
|
| def get_diff_record(self, expected_image_locator, actual_image_locator):
|
| """Returns the DiffRecord for this image pair.
|
|
|
| - Raises a KeyError if we don't have a DiffRecord for this image pair.
|
| + This call will block until the diff record is available, or we were unable
|
| + to generate it.
|
| +
|
| + Args:
|
| + expected_image_locator: a unique ID string under which we will store the
|
| + expected image within storage_root (probably including a checksum to
|
| + guarantee uniqueness)
|
| + actual_image_locator: a unique ID string under which we will store the
|
| + actual image within storage_root (probably including a checksum to
|
| + guarantee uniqueness)
|
| +
|
| + Returns the DiffRecord for this image pair, or None if we were unable to
|
| + generate one.
|
| """
|
| key = (_sanitize_locator(expected_image_locator),
|
| _sanitize_locator(actual_image_locator))
|
| - return self._diff_dict[key]
|
| + diff_record = self._diff_dict[key]
|
| +
|
| + # If we have no results yet, block until we do.
|
| + while diff_record == _DIFFRECORD_PENDING:
|
| + time.sleep(1)
|
| + diff_record = self._diff_dict[key]
|
| +
|
| + # Once we have the result...
|
| + if diff_record == _DIFFRECORD_FAILED:
|
| + logging.error(
|
| + 'failed to create a DiffRecord for expected_image_locator=%s , '
|
| + 'actual_image_locator=%s' % (
|
| + expected_image_locator, actual_image_locator))
|
| + return None
|
| + else:
|
| + return diff_record
|
|
|
|
|
| # Utility functions
|
|
|
| -def _download_file(local_filepath, url):
|
| +def _download_file(gs, local_filepath, url):
|
| """Download a file from url to local_filepath, unless it is already there.
|
|
|
| Args:
|
| + gs: instance of GSUtils object, in case the url points at Google Storage
|
| local_filepath: path on local disk where the image should be stored
|
| - url: URL from which we can download the image if we don't have it yet
|
| + url: HTTP or GS URL from which we can download the image if we don't have
|
| + it yet
|
| """
|
| + global global_file_collisions
|
| if not os.path.exists(local_filepath):
|
| _mkdir_unless_exists(os.path.dirname(local_filepath))
|
| - with contextlib.closing(urllib.urlopen(url)) as url_handle:
|
| - with open(local_filepath, 'wb') as file_handle:
|
| - shutil.copyfileobj(fsrc=url_handle, fdst=file_handle)
|
| +
|
| + # First download the file contents into a unique filename, and
|
| + # then rename that file. That way, if multiple threads are downloading
|
| + # the same filename at the same time, they won't interfere with each
|
| + # other (they will both download the file, and one will "win" in the end)
|
| + temp_filename = '%s-%d' % (local_filepath,
|
| + threading.current_thread().ident)
|
| + if gs_utils.GSUtils.is_gs_url(url):
|
| + (bucket, path) = gs_utils.GSUtils.split_gs_url(url)
|
| + gs.download_file(source_bucket=bucket, source_path=path,
|
| + dest_path=temp_filename)
|
| + else:
|
| + with contextlib.closing(urllib.urlopen(url)) as url_handle:
|
| + with open(temp_filename, 'wb') as file_handle:
|
| + shutil.copyfileobj(fsrc=url_handle, fdst=file_handle)
|
| +
|
| + # Rename the file to its real filename.
|
| + # Keep count of how many colliding downloads we encounter;
|
| + # if it's a large number, we may want to change our download strategy
|
| + # to minimize repeated downloads.
|
| + if os.path.exists(local_filepath):
|
| + global_file_collisions += 1
|
| + else:
|
| + os.rename(temp_filename, local_filepath)
|
|
|
|
|
| def _mkdir_unless_exists(path):
|
| @@ -306,8 +423,11 @@ def _mkdir_unless_exists(path):
|
| Args:
|
| path: path on local disk
|
| """
|
| - if not os.path.isdir(path):
|
| + try:
|
| os.makedirs(path)
|
| + except OSError as e:
|
| + if e.errno == errno.EEXIST:
|
| + pass
|
|
|
|
|
| def _sanitize_locator(locator):
|
|
|