1 #!/usr/bin/python 2 3 """ 4 Copyright 2013 Google Inc. 5 6 Use of this source code is governed by a BSD-style license that can be 7 found in the LICENSE file. 8 9 Calulate differences between image pairs, and store them in a database. 10 """ 11 12 # System-level imports 13 import contextlib 14 import errno 15 import json 16 import logging 17 import os 18 import Queue 19 import re 20 import shutil 21 import tempfile 22 import threading 23 import time 24 import urllib 25 26 # Must fix up PYTHONPATH before importing from within Skia 27 import rs_fixpypath # pylint: disable=W0611 28 29 # Imports from within Skia 30 import find_run_binary 31 from py.utils import gs_utils 32 33 34 SKPDIFF_BINARY = find_run_binary.find_path_to_program('skpdiff') 35 36 DEFAULT_IMAGE_SUFFIX = '.png' 37 DEFAULT_IMAGES_SUBDIR = 'images' 38 # TODO(epoger): Figure out a better default number of threads; for now, 39 # using a conservative default value. 40 DEFAULT_NUM_WORKER_THREADS = 1 41 42 DISALLOWED_FILEPATH_CHAR_REGEX = re.compile('[^\w\-]') 43 44 RGBDIFFS_SUBDIR = 'diffs' 45 WHITEDIFFS_SUBDIR = 'whitediffs' 46 47 # Keys used within DiffRecord dictionary representations. 48 # NOTE: Keep these in sync with static/constants.js 49 KEY__DIFFERENCES__MAX_DIFF_PER_CHANNEL = 'maxDiffPerChannel' 50 KEY__DIFFERENCES__NUM_DIFF_PIXELS = 'numDifferingPixels' 51 KEY__DIFFERENCES__PERCENT_DIFF_PIXELS = 'percentDifferingPixels' 52 KEY__DIFFERENCES__PERCEPTUAL_DIFF = 'perceptualDifference' 53 KEY__DIFFERENCES__DIFF_URL = 'diffUrl' 54 KEY__DIFFERENCES__WHITE_DIFF_URL = 'whiteDiffUrl' 55 56 # Special values within ImageDiffDB._diff_dict 57 _DIFFRECORD_FAILED = 'failed' 58 _DIFFRECORD_PENDING = 'pending' 59 60 # How often to report tasks_queue size 61 QUEUE_LOGGING_GRANULARITY = 1000 62 63 # Temporary variable to keep track of how many times we download 64 # the same file in multiple threads. 65 # TODO(epoger): Delete this, once we see that the number stays close to 0. 66 global_file_collisions = 0 67 68 69 class DiffRecord(object): 70 """ Record of differences between two images. """ 71 72 def __init__(self, gs, storage_root, 73 expected_image_url, expected_image_locator, 74 actual_image_url, actual_image_locator, 75 expected_images_subdir=DEFAULT_IMAGES_SUBDIR, 76 actual_images_subdir=DEFAULT_IMAGES_SUBDIR, 77 image_suffix=DEFAULT_IMAGE_SUFFIX): 78 """Download this pair of images (unless we already have them on local disk), 79 and prepare a DiffRecord for them. 80 81 Args: 82 gs: instance of GSUtils object we can use to download images 83 storage_root: root directory on local disk within which we store all 84 images 85 expected_image_url: file, GS, or HTTP url from which we will download the 86 expected image 87 expected_image_locator: a unique ID string under which we will store the 88 expected image within storage_root (probably including a checksum to 89 guarantee uniqueness) 90 actual_image_url: file, GS, or HTTP url from which we will download the 91 actual image 92 actual_image_locator: a unique ID string under which we will store the 93 actual image within storage_root (probably including a checksum to 94 guarantee uniqueness) 95 expected_images_subdir: the subdirectory expected images are stored in. 96 actual_images_subdir: the subdirectory actual images are stored in. 97 image_suffix: the suffix of images. 98 """ 99 expected_image_locator = _sanitize_locator(expected_image_locator) 100 actual_image_locator = _sanitize_locator(actual_image_locator) 101 102 # Download the expected/actual images, if we don't have them already. 103 expected_image_file = os.path.join( 104 storage_root, expected_images_subdir, 105 str(expected_image_locator) + image_suffix) 106 actual_image_file = os.path.join( 107 storage_root, actual_images_subdir, 108 str(actual_image_locator) + image_suffix) 109 for image_file, image_url in [ 110 (expected_image_file, expected_image_url), 111 (actual_image_file, actual_image_url)]: 112 if image_file and image_url: 113 try: 114 _download_file(gs, image_file, image_url) 115 except Exception: 116 logging.exception('unable to download image_url %s to file %s' % 117 (image_url, image_file)) 118 raise 119 120 # Return early if we do not need to generate diffs. 121 if (expected_image_url == actual_image_url or 122 not expected_image_url or not actual_image_url): 123 return 124 125 # Get all diff images and values using the skpdiff binary. 126 skpdiff_output_dir = tempfile.mkdtemp() 127 try: 128 skpdiff_summary_file = os.path.join(skpdiff_output_dir, 129 'skpdiff-output.json') 130 skpdiff_rgbdiff_dir = os.path.join(storage_root, RGBDIFFS_SUBDIR) 131 skpdiff_whitediff_dir = os.path.join(storage_root, WHITEDIFFS_SUBDIR) 132 _mkdir_unless_exists(skpdiff_rgbdiff_dir) 133 _mkdir_unless_exists(skpdiff_rgbdiff_dir) 134 135 # TODO(epoger): Consider calling skpdiff ONCE for all image pairs, 136 # instead of calling it separately for each image pair. 137 # Pro: we'll incur less overhead from making repeated system calls, 138 # spinning up the skpdiff binary, etc. 139 # Con: we would have to wait until all image pairs were loaded before 140 # generating any of the diffs? 141 # Note(stephana): '--longnames' was added to allow for this 142 # case (multiple files at once) versus specifying output diffs 143 # directly. 144 find_run_binary.run_command( 145 [SKPDIFF_BINARY, '-p', expected_image_file, actual_image_file, 146 '--jsonp', 'false', 147 '--longnames', 'true', 148 '--output', skpdiff_summary_file, 149 '--differs', 'perceptual', 'different_pixels', 150 '--rgbDiffDir', skpdiff_rgbdiff_dir, 151 '--whiteDiffDir', skpdiff_whitediff_dir, 152 ]) 153 154 # Get information out of the skpdiff_summary_file. 155 with contextlib.closing(open(skpdiff_summary_file)) as fp: 156 data = json.load(fp) 157 158 # For now, we can assume there is only one record in the output summary, 159 # since we passed skpdiff only one pair of images. 160 record = data['records'][0] 161 self._width = record['width'] 162 self._height = record['height'] 163 self._diffUrl = os.path.split(record['rgbDiffPath'])[1] 164 self._whiteDiffUrl = os.path.split(record['whiteDiffPath'])[1] 165 166 # TODO: make max_diff_per_channel a tuple instead of a list, because the 167 # structure is meaningful (first element is red, second is green, etc.) 168 # See http://stackoverflow.com/a/626871 169 self._max_diff_per_channel = [ 170 record['maxRedDiff'], record['maxGreenDiff'], record['maxBlueDiff']] 171 per_differ_stats = record['diffs'] 172 for stats in per_differ_stats: 173 differ_name = stats['differName'] 174 if differ_name == 'different_pixels': 175 self._num_pixels_differing = stats['pointsOfInterest'] 176 elif differ_name == 'perceptual': 177 perceptual_similarity = stats['result'] 178 179 # skpdiff returns the perceptual similarity; convert it to get the 180 # perceptual difference percentage. 181 # skpdiff outputs -1 if the images are different sizes. Treat any 182 # output that does not lie in [0, 1] as having 0% perceptual 183 # similarity. 184 if not 0 <= perceptual_similarity <= 1: 185 perceptual_similarity = 0 186 self._perceptual_difference = 100 - (perceptual_similarity * 100) 187 finally: 188 shutil.rmtree(skpdiff_output_dir) 189 190 # TODO(epoger): Use properties instead of getters throughout. 191 # See http://stackoverflow.com/a/6618176 192 def get_num_pixels_differing(self): 193 """Returns the absolute number of pixels that differ.""" 194 return self._num_pixels_differing 195 196 def get_percent_pixels_differing(self): 197 """Returns the percentage of pixels that differ, as a float between 198 0 and 100 (inclusive).""" 199 return ((float(self._num_pixels_differing) * 100) / 200 (self._width * self._height)) 201 202 def get_perceptual_difference(self): 203 """Returns the perceptual difference percentage.""" 204 return self._perceptual_difference 205 206 def get_max_diff_per_channel(self): 207 """Returns the maximum difference between the expected and actual images 208 for each R/G/B channel, as a list.""" 209 return self._max_diff_per_channel 210 211 def as_dict(self): 212 """Returns a dictionary representation of this DiffRecord, as needed when 213 constructing the JSON representation.""" 214 return { 215 KEY__DIFFERENCES__NUM_DIFF_PIXELS: self._num_pixels_differing, 216 KEY__DIFFERENCES__PERCENT_DIFF_PIXELS: 217 self.get_percent_pixels_differing(), 218 KEY__DIFFERENCES__MAX_DIFF_PER_CHANNEL: self._max_diff_per_channel, 219 KEY__DIFFERENCES__PERCEPTUAL_DIFF: self._perceptual_difference, 220 KEY__DIFFERENCES__DIFF_URL: self._diffUrl, 221 KEY__DIFFERENCES__WHITE_DIFF_URL: self._whiteDiffUrl, 222 } 223 224 225 226 class ImageDiffDB(object): 227 """ Calculates differences between image pairs, maintaining a database of 228 them for download.""" 229 230 def __init__(self, storage_root, gs=None, 231 num_worker_threads=DEFAULT_NUM_WORKER_THREADS): 232 """ 233 Args: 234 storage_root: string; root path within the DB will store all of its stuff 235 gs: instance of GSUtils object we can use to download images 236 num_worker_threads: how many threads that download images and 237 generate diffs simultaneously 238 """ 239 self._storage_root = storage_root 240 self._gs = gs 241 242 # Mechanism for reporting queue size periodically. 243 self._last_queue_size_reported = None 244 self._queue_size_report_lock = threading.RLock() 245 246 # Dictionary of DiffRecords, keyed by (expected_image_locator, 247 # actual_image_locator) tuples. 248 # Values can also be _DIFFRECORD_PENDING, _DIFFRECORD_FAILED. 249 # 250 # Any thread that modifies _diff_dict must first acquire 251 # _diff_dict_writelock! 252 # 253 # TODO(epoger): Disk is limitless, but RAM is not... so, we should probably 254 # remove items from self._diff_dict if they haven't been accessed for a 255 # long time. We can always regenerate them by diffing the images we 256 # previously downloaded to local disk. 257 # I guess we should figure out how expensive it is to download vs diff the 258 # image pairs... if diffing them is expensive too, we can write these 259 # _diff_dict objects out to disk if there's too many to hold in RAM. 260 # Or we could use virtual memory to handle that automatically. 261 self._diff_dict = {} 262 self._diff_dict_writelock = threading.RLock() 263 264 # Set up the queue for asynchronously loading DiffRecords, and start the 265 # worker threads reading from it. 266 # The queue maxsize must be 0 (infinite size queue), so that asynchronous 267 # calls can return as soon as possible. 268 self._tasks_queue = Queue.Queue(maxsize=0) 269 self._workers = [] 270 for i in range(num_worker_threads): 271 worker = threading.Thread(target=self.worker, args=(i,)) 272 worker.daemon = True 273 worker.start() 274 self._workers.append(worker) 275 276 def log_queue_size_if_changed(self, limit_verbosity=True): 277 """Log the size of self._tasks_queue, if it has changed since the last call. 278 279 Reports the current queue size, using log.info(), unless the queue is the 280 same size as the last time we reported it. 281 282 Args: 283 limit_verbosity: if True, only log if the queue size is a multiple of 284 QUEUE_LOGGING_GRANULARITY 285 """ 286 # Acquire the lock, to synchronize access to self._last_queue_size_reported 287 self._queue_size_report_lock.acquire() 288 try: 289 size = self._tasks_queue.qsize() 290 if size == self._last_queue_size_reported: 291 return 292 if limit_verbosity and (size % QUEUE_LOGGING_GRANULARITY != 0): 293 return 294 logging.info('tasks_queue size is %d' % size) 295 self._last_queue_size_reported = size 296 finally: 297 self._queue_size_report_lock.release() 298 299 def worker(self, worker_num): 300 """Launch a worker thread that pulls tasks off self._tasks_queue. 301 302 Args: 303 worker_num: (integer) which worker this is 304 """ 305 while True: 306 self.log_queue_size_if_changed() 307 params = self._tasks_queue.get() 308 key, expected_image_url, actual_image_url = params 309 try: 310 diff_record = DiffRecord( 311 self._gs, self._storage_root, 312 expected_image_url=expected_image_url, 313 expected_image_locator=key[0], 314 actual_image_url=actual_image_url, 315 actual_image_locator=key[1]) 316 except Exception: 317 logging.exception( 318 'exception while creating DiffRecord for key %s' % str(key)) 319 diff_record = _DIFFRECORD_FAILED 320 self._diff_dict_writelock.acquire() 321 try: 322 self._diff_dict[key] = diff_record 323 finally: 324 self._diff_dict_writelock.release() 325 326 @property 327 def storage_root(self): 328 return self._storage_root 329 330 def add_image_pair(self, 331 expected_image_url, expected_image_locator, 332 actual_image_url, actual_image_locator): 333 """Asynchronously prepare a DiffRecord for a pair of images. 334 335 This method will return quickly; calls to get_diff_record() will block 336 until the DiffRecord is available (or we have given up on creating it). 337 338 If we already have a DiffRecord for this particular image pair, no work 339 will be done. 340 341 If expected_image_url (or its locator) is None, just download actual_image. 342 If actual_image_url (or its locator) is None, just download expected_image. 343 344 Args: 345 expected_image_url: file, GS, or HTTP url from which we will download the 346 expected image 347 expected_image_locator: a unique ID string under which we will store the 348 expected image within storage_root (probably including a checksum to 349 guarantee uniqueness) 350 actual_image_url: file, GS, or HTTP url from which we will download the 351 actual image 352 actual_image_locator: a unique ID string under which we will store the 353 actual image within storage_root (probably including a checksum to 354 guarantee uniqueness) 355 """ 356 expected_image_locator = _sanitize_locator(expected_image_locator) 357 actual_image_locator = _sanitize_locator(actual_image_locator) 358 key = (expected_image_locator, actual_image_locator) 359 must_add_to_queue = False 360 361 self._diff_dict_writelock.acquire() 362 try: 363 if not key in self._diff_dict: 364 # If we have already requested a diff between these two images, 365 # we don't need to request it again. 366 must_add_to_queue = True 367 self._diff_dict[key] = _DIFFRECORD_PENDING 368 finally: 369 self._diff_dict_writelock.release() 370 371 if must_add_to_queue: 372 self._tasks_queue.put((key, expected_image_url, actual_image_url)) 373 self.log_queue_size_if_changed() 374 375 def get_diff_record(self, expected_image_locator, actual_image_locator): 376 """Returns the DiffRecord for this image pair. 377 378 This call will block until the diff record is available, or we were unable 379 to generate it. 380 381 Args: 382 expected_image_locator: a unique ID string under which we will store the 383 expected image within storage_root (probably including a checksum to 384 guarantee uniqueness) 385 actual_image_locator: a unique ID string under which we will store the 386 actual image within storage_root (probably including a checksum to 387 guarantee uniqueness) 388 389 Returns the DiffRecord for this image pair, or None if we were unable to 390 generate one. 391 """ 392 key = (_sanitize_locator(expected_image_locator), 393 _sanitize_locator(actual_image_locator)) 394 diff_record = self._diff_dict[key] 395 396 # If we have no results yet, block until we do. 397 while diff_record == _DIFFRECORD_PENDING: 398 time.sleep(1) 399 diff_record = self._diff_dict[key] 400 401 # Once we have the result... 402 if diff_record == _DIFFRECORD_FAILED: 403 logging.error( 404 'failed to create a DiffRecord for expected_image_locator=%s , ' 405 'actual_image_locator=%s' % ( 406 expected_image_locator, actual_image_locator)) 407 return None 408 else: 409 return diff_record 410 411 412 # Utility functions 413 414 def _download_file(gs, local_filepath, url): 415 """Download a file from url to local_filepath, unless it is already there. 416 417 Args: 418 gs: instance of GSUtils object, in case the url points at Google Storage 419 local_filepath: path on local disk where the image should be stored 420 url: HTTP or GS URL from which we can download the image if we don't have 421 it yet 422 """ 423 global global_file_collisions 424 if not os.path.exists(local_filepath): 425 _mkdir_unless_exists(os.path.dirname(local_filepath)) 426 427 # First download the file contents into a unique filename, and 428 # then rename that file. That way, if multiple threads are downloading 429 # the same filename at the same time, they won't interfere with each 430 # other (they will both download the file, and one will "win" in the end) 431 temp_filename = '%s-%d' % (local_filepath, 432 threading.current_thread().ident) 433 if gs_utils.GSUtils.is_gs_url(url): 434 (bucket, path) = gs_utils.GSUtils.split_gs_url(url) 435 gs.download_file(source_bucket=bucket, source_path=path, 436 dest_path=temp_filename) 437 else: 438 with contextlib.closing(urllib.urlopen(url)) as url_handle: 439 with open(temp_filename, 'wb') as file_handle: 440 shutil.copyfileobj(fsrc=url_handle, fdst=file_handle) 441 442 # Rename the file to its real filename. 443 # Keep count of how many colliding downloads we encounter; 444 # if it's a large number, we may want to change our download strategy 445 # to minimize repeated downloads. 446 if os.path.exists(local_filepath): 447 global_file_collisions += 1 448 else: 449 os.rename(temp_filename, local_filepath) 450 451 452 def _mkdir_unless_exists(path): 453 """Unless path refers to an already-existing directory, create it. 454 455 Args: 456 path: path on local disk 457 """ 458 try: 459 os.makedirs(path) 460 except OSError as e: 461 if e.errno == errno.EEXIST: 462 pass 463 464 465 def _sanitize_locator(locator): 466 """Returns a sanitized version of a locator (one in which we know none of the 467 characters will have special meaning in filenames). 468 469 Args: 470 locator: string, or something that can be represented as a string. 471 If None or '', it is returned without modification, because empty 472 locators have a particular meaning ("there is no image for this") 473 """ 474 if locator: 475 return DISALLOWED_FILEPATH_CHAR_REGEX.sub('_', str(locator)) 476 else: 477 return locator 478