3f04517721
When set, it will only write out images that don't match expectations. BUG=skia:1942 R=rmistry@google.com Author: epoger@google.com Review URL: https://codereview.chromium.org/283123002 git-svn-id: http://skia.googlecode.com/svn/trunk@14748 2bbb7eff-a529-9590-31e7-b0007b416f81
651 lines
25 KiB
Python
Executable File
651 lines
25 KiB
Python
Executable File
#!/usr/bin/python
|
|
|
|
"""
|
|
Copyright 2014 Google Inc.
|
|
|
|
Use of this source code is governed by a BSD-style license that can be
|
|
found in the LICENSE file.
|
|
|
|
Test the render_pictures binary.
|
|
"""
|
|
|
|
# System-level imports
|
|
import copy
|
|
import json
|
|
import os
|
|
import shutil
|
|
import tempfile
|
|
|
|
# Imports from within Skia
|
|
import base_unittest
|
|
|
|
# Maximum length of text diffs to show when tests fail
|
|
MAX_DIFF_LENGTH = 30000
|
|
|
|
EXPECTED_HEADER_CONTENTS = {
|
|
"type" : "ChecksummedImages",
|
|
"revision" : 1,
|
|
}
|
|
|
|
# Manually verified: 640x400 red rectangle with black border
|
|
# Standard expectations will be set up in such a way that this image fails
|
|
# the comparison.
|
|
RED_WHOLEIMAGE = {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 11092453015575919668,
|
|
"comparisonResult" : "failed",
|
|
"filepath" : "red_skp.png",
|
|
}
|
|
|
|
# Manually verified: 640x400 green rectangle with black border
|
|
# Standard expectations will be set up in such a way that this image passes
|
|
# the comparison.
|
|
GREEN_WHOLEIMAGE = {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 8891695120562235492,
|
|
"comparisonResult" : "succeeded",
|
|
"filepath" : "green_skp.png",
|
|
}
|
|
|
|
# Manually verified these 6 images, all 256x256 tiles,
|
|
# consistent with a tiled version of the 640x400 red rect
|
|
# with black borders.
|
|
# Standard expectations will be set up in such a way that these images fail
|
|
# the comparison.
|
|
RED_TILES = [{
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 5815827069051002745,
|
|
"comparisonResult" : "failed",
|
|
"filepath" : "red_skp-tile0.png",
|
|
},{
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 9323613075234140270,
|
|
"comparisonResult" : "failed",
|
|
"filepath" : "red_skp-tile1.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 16670399404877552232,
|
|
"comparisonResult" : "failed",
|
|
"filepath" : "red_skp-tile2.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 2507897274083364964,
|
|
"comparisonResult" : "failed",
|
|
"filepath" : "red_skp-tile3.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 7325267995523877959,
|
|
"comparisonResult" : "failed",
|
|
"filepath" : "red_skp-tile4.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 2181381724594493116,
|
|
"comparisonResult" : "failed",
|
|
"filepath" : "red_skp-tile5.png",
|
|
}]
|
|
|
|
# Manually verified these 6 images, all 256x256 tiles,
|
|
# consistent with a tiled version of the 640x400 green rect
|
|
# with black borders.
|
|
# Standard expectations will be set up in such a way that these images pass
|
|
# the comparison.
|
|
GREEN_TILES = [{
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 12587324416545178013,
|
|
"comparisonResult" : "succeeded",
|
|
"filepath" : "green_skp-tile0.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 7624374914829746293,
|
|
"comparisonResult" : "succeeded",
|
|
"filepath" : "green_skp-tile1.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 5686489729535631913,
|
|
"comparisonResult" : "succeeded",
|
|
"filepath" : "green_skp-tile2.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 7980646035555096146,
|
|
"comparisonResult" : "succeeded",
|
|
"filepath" : "green_skp-tile3.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 17817086664365875131,
|
|
"comparisonResult" : "succeeded",
|
|
"filepath" : "green_skp-tile4.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 10673669813016809363,
|
|
"comparisonResult" : "succeeded",
|
|
"filepath" : "green_skp-tile5.png",
|
|
}]
|
|
|
|
|
|
def modified_dict(input_dict, modification_dict):
|
|
"""Returns a dict, with some modifications applied to it.
|
|
|
|
Args:
|
|
input_dict: a dictionary (which will be copied, not modified in place)
|
|
modification_dict: a set of key/value pairs to overwrite in the dict
|
|
"""
|
|
output_dict = input_dict.copy()
|
|
output_dict.update(modification_dict)
|
|
return output_dict
|
|
|
|
|
|
def modified_list_of_dicts(input_list, modification_dict):
|
|
"""Returns a list of dicts, with some modifications applied to each dict.
|
|
|
|
Args:
|
|
input_list: a list of dictionaries; these dicts will be copied, not
|
|
modified in place
|
|
modification_dict: a set of key/value pairs to overwrite in each dict
|
|
within input_list
|
|
"""
|
|
output_list = []
|
|
for input_dict in input_list:
|
|
output_dict = modified_dict(input_dict, modification_dict)
|
|
output_list.append(output_dict)
|
|
return output_list
|
|
|
|
|
|
class RenderPicturesTest(base_unittest.TestCase):
|
|
|
|
def setUp(self):
|
|
self.maxDiff = MAX_DIFF_LENGTH
|
|
self._expectations_dir = tempfile.mkdtemp()
|
|
self._input_skp_dir = tempfile.mkdtemp()
|
|
# All output of render_pictures binary will go into this directory.
|
|
self._output_dir = tempfile.mkdtemp()
|
|
|
|
def tearDown(self):
|
|
shutil.rmtree(self._expectations_dir)
|
|
shutil.rmtree(self._input_skp_dir)
|
|
shutil.rmtree(self._output_dir)
|
|
|
|
def test_tiled_whole_image(self):
|
|
"""Run render_pictures with tiles and --writeWholeImage flag.
|
|
|
|
TODO(epoger): This test generates undesired results! The JSON summary
|
|
includes both whole-image and tiled-images (as it should), but only
|
|
whole-images are written out to disk. See http://skbug.com/2463
|
|
Once I fix that, I should add a similar test that exercises mismatchPath.
|
|
|
|
TODO(epoger): I noticed that when this is run without --writePath being
|
|
specified, this test writes red_skp.png and green_skp.png to the current
|
|
directory. We should fix that... if --writePath is not specified, this
|
|
probably shouldn't write out red_skp.png and green_skp.png at all!
|
|
See http://skbug.com/2464
|
|
"""
|
|
output_json_path = os.path.join(self._output_dir, 'actuals.json')
|
|
write_path_dir = self.create_empty_dir(
|
|
path=os.path.join(self._output_dir, 'writePath'))
|
|
self._generate_skps()
|
|
expectations_path = self._create_expectations()
|
|
self._run_render_pictures([
|
|
'-r', self._input_skp_dir,
|
|
'--bbh', 'grid', '256', '256',
|
|
'--mode', 'tile', '256', '256',
|
|
'--readJsonSummaryPath', expectations_path,
|
|
'--writeJsonSummaryPath', output_json_path,
|
|
'--writePath', write_path_dir,
|
|
'--writeWholeImage'])
|
|
expected_summary_dict = {
|
|
"header" : EXPECTED_HEADER_CONTENTS,
|
|
"actual-results" : {
|
|
"red.skp": {
|
|
"tiled-images": RED_TILES,
|
|
"whole-image": RED_WHOLEIMAGE,
|
|
},
|
|
"green.skp": {
|
|
"tiled-images": GREEN_TILES,
|
|
"whole-image": GREEN_WHOLEIMAGE,
|
|
}
|
|
}
|
|
}
|
|
self._assert_json_contents(output_json_path, expected_summary_dict)
|
|
self._assert_directory_contents(
|
|
write_path_dir, ['red_skp.png', 'green_skp.png'])
|
|
|
|
def test_missing_tile_and_whole_image(self):
|
|
"""test_tiled_whole_image, but missing expectations for some images.
|
|
"""
|
|
output_json_path = os.path.join(self._output_dir, 'actuals.json')
|
|
write_path_dir = self.create_empty_dir(
|
|
path=os.path.join(self._output_dir, 'writePath'))
|
|
self._generate_skps()
|
|
expectations_path = self._create_expectations(missing_some_images=True)
|
|
self._run_render_pictures([
|
|
'-r', self._input_skp_dir,
|
|
'--bbh', 'grid', '256', '256',
|
|
'--mode', 'tile', '256', '256',
|
|
'--readJsonSummaryPath', expectations_path,
|
|
'--writeJsonSummaryPath', output_json_path,
|
|
'--writePath', write_path_dir,
|
|
'--writeWholeImage'])
|
|
modified_red_tiles = copy.deepcopy(RED_TILES)
|
|
modified_red_tiles[5]['comparisonResult'] = 'no-comparison'
|
|
expected_summary_dict = {
|
|
"header" : EXPECTED_HEADER_CONTENTS,
|
|
"actual-results" : {
|
|
"red.skp": {
|
|
"tiled-images": modified_red_tiles,
|
|
"whole-image": modified_dict(
|
|
RED_WHOLEIMAGE, {"comparisonResult" : "no-comparison"}),
|
|
},
|
|
"green.skp": {
|
|
"tiled-images": GREEN_TILES,
|
|
"whole-image": GREEN_WHOLEIMAGE,
|
|
}
|
|
}
|
|
}
|
|
self._assert_json_contents(output_json_path, expected_summary_dict)
|
|
|
|
def test_untiled(self):
|
|
"""Run without tiles."""
|
|
output_json_path = os.path.join(self._output_dir, 'actuals.json')
|
|
write_path_dir = self.create_empty_dir(
|
|
path=os.path.join(self._output_dir, 'writePath'))
|
|
self._generate_skps()
|
|
expectations_path = self._create_expectations()
|
|
self._run_render_pictures([
|
|
'-r', self._input_skp_dir,
|
|
'--readJsonSummaryPath', expectations_path,
|
|
'--writePath', write_path_dir,
|
|
'--writeJsonSummaryPath', output_json_path])
|
|
expected_summary_dict = {
|
|
"header" : EXPECTED_HEADER_CONTENTS,
|
|
"actual-results" : {
|
|
"red.skp": {
|
|
"whole-image": RED_WHOLEIMAGE,
|
|
},
|
|
"green.skp": {
|
|
"whole-image": GREEN_WHOLEIMAGE,
|
|
}
|
|
}
|
|
}
|
|
self._assert_json_contents(output_json_path, expected_summary_dict)
|
|
self._assert_directory_contents(
|
|
write_path_dir, ['red_skp.png', 'green_skp.png'])
|
|
|
|
def test_untiled_writeChecksumBasedFilenames(self):
|
|
"""Same as test_untiled, but with --writeChecksumBasedFilenames."""
|
|
output_json_path = os.path.join(self._output_dir, 'actuals.json')
|
|
write_path_dir = self.create_empty_dir(
|
|
path=os.path.join(self._output_dir, 'writePath'))
|
|
self._generate_skps()
|
|
self._run_render_pictures(['-r', self._input_skp_dir,
|
|
'--writeChecksumBasedFilenames',
|
|
'--writePath', write_path_dir,
|
|
'--writeJsonSummaryPath', output_json_path])
|
|
expected_summary_dict = {
|
|
"header" : EXPECTED_HEADER_CONTENTS,
|
|
"actual-results" : {
|
|
"red.skp": {
|
|
# Manually verified: 640x400 red rectangle with black border
|
|
"whole-image": {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 11092453015575919668,
|
|
"comparisonResult" : "no-comparison",
|
|
"filepath" : "red_skp/bitmap-64bitMD5_11092453015575919668.png",
|
|
},
|
|
},
|
|
"green.skp": {
|
|
# Manually verified: 640x400 green rectangle with black border
|
|
"whole-image": {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 8891695120562235492,
|
|
"comparisonResult" : "no-comparison",
|
|
"filepath" : "green_skp/bitmap-64bitMD5_8891695120562235492.png",
|
|
},
|
|
}
|
|
}
|
|
}
|
|
self._assert_json_contents(output_json_path, expected_summary_dict)
|
|
self._assert_directory_contents(write_path_dir, ['red_skp', 'green_skp'])
|
|
self._assert_directory_contents(
|
|
os.path.join(write_path_dir, 'red_skp'),
|
|
['bitmap-64bitMD5_11092453015575919668.png'])
|
|
self._assert_directory_contents(
|
|
os.path.join(write_path_dir, 'green_skp'),
|
|
['bitmap-64bitMD5_8891695120562235492.png'])
|
|
|
|
def test_untiled_validate(self):
|
|
"""Same as test_untiled, but with --validate."""
|
|
output_json_path = os.path.join(self._output_dir, 'actuals.json')
|
|
write_path_dir = self.create_empty_dir(
|
|
path=os.path.join(self._output_dir, 'writePath'))
|
|
self._generate_skps()
|
|
expectations_path = self._create_expectations()
|
|
self._run_render_pictures([
|
|
'-r', self._input_skp_dir,
|
|
'--readJsonSummaryPath', expectations_path,
|
|
'--validate',
|
|
'--writePath', write_path_dir,
|
|
'--writeJsonSummaryPath', output_json_path])
|
|
expected_summary_dict = {
|
|
"header" : EXPECTED_HEADER_CONTENTS,
|
|
"actual-results" : {
|
|
"red.skp": {
|
|
"whole-image": RED_WHOLEIMAGE,
|
|
},
|
|
"green.skp": {
|
|
"whole-image": GREEN_WHOLEIMAGE,
|
|
}
|
|
}
|
|
}
|
|
self._assert_json_contents(output_json_path, expected_summary_dict)
|
|
self._assert_directory_contents(
|
|
write_path_dir, ['red_skp.png', 'green_skp.png'])
|
|
|
|
def test_untiled_without_writePath(self):
|
|
"""Same as test_untiled, but without --writePath."""
|
|
output_json_path = os.path.join(self._output_dir, 'actuals.json')
|
|
self._generate_skps()
|
|
expectations_path = self._create_expectations()
|
|
self._run_render_pictures([
|
|
'-r', self._input_skp_dir,
|
|
'--readJsonSummaryPath', expectations_path,
|
|
'--writeJsonSummaryPath', output_json_path])
|
|
expected_summary_dict = {
|
|
"header" : EXPECTED_HEADER_CONTENTS,
|
|
"actual-results" : {
|
|
"red.skp": {
|
|
"whole-image": RED_WHOLEIMAGE,
|
|
},
|
|
"green.skp": {
|
|
"whole-image": GREEN_WHOLEIMAGE,
|
|
}
|
|
}
|
|
}
|
|
self._assert_json_contents(output_json_path, expected_summary_dict)
|
|
|
|
def test_tiled(self):
|
|
"""Generate individual tiles."""
|
|
output_json_path = os.path.join(self._output_dir, 'actuals.json')
|
|
write_path_dir = self.create_empty_dir(
|
|
path=os.path.join(self._output_dir, 'writePath'))
|
|
self._generate_skps()
|
|
expectations_path = self._create_expectations()
|
|
self._run_render_pictures([
|
|
'-r', self._input_skp_dir,
|
|
'--bbh', 'grid', '256', '256',
|
|
'--mode', 'tile', '256', '256',
|
|
'--readJsonSummaryPath', expectations_path,
|
|
'--writePath', write_path_dir,
|
|
'--writeJsonSummaryPath', output_json_path])
|
|
expected_summary_dict = {
|
|
"header" : EXPECTED_HEADER_CONTENTS,
|
|
"actual-results" : {
|
|
"red.skp": {
|
|
"tiled-images": RED_TILES,
|
|
},
|
|
"green.skp": {
|
|
"tiled-images": GREEN_TILES,
|
|
}
|
|
}
|
|
}
|
|
self._assert_json_contents(output_json_path, expected_summary_dict)
|
|
self._assert_directory_contents(
|
|
write_path_dir,
|
|
['red_skp-tile0.png', 'red_skp-tile1.png', 'red_skp-tile2.png',
|
|
'red_skp-tile3.png', 'red_skp-tile4.png', 'red_skp-tile5.png',
|
|
'green_skp-tile0.png', 'green_skp-tile1.png', 'green_skp-tile2.png',
|
|
'green_skp-tile3.png', 'green_skp-tile4.png', 'green_skp-tile5.png',
|
|
])
|
|
|
|
def test_tiled_mismatches(self):
|
|
"""Same as test_tiled, but only write out mismatching images."""
|
|
output_json_path = os.path.join(self._output_dir, 'actuals.json')
|
|
mismatch_path_dir = self.create_empty_dir(
|
|
path=os.path.join(self._output_dir, 'mismatchPath'))
|
|
self._generate_skps()
|
|
expectations_path = self._create_expectations()
|
|
self._run_render_pictures([
|
|
'-r', self._input_skp_dir,
|
|
'--bbh', 'grid', '256', '256',
|
|
'--mode', 'tile', '256', '256',
|
|
'--readJsonSummaryPath', expectations_path,
|
|
'--mismatchPath', mismatch_path_dir,
|
|
'--writeJsonSummaryPath', output_json_path])
|
|
expected_summary_dict = {
|
|
"header" : EXPECTED_HEADER_CONTENTS,
|
|
"actual-results" : {
|
|
"red.skp": {
|
|
"tiled-images": RED_TILES,
|
|
},
|
|
"green.skp": {
|
|
"tiled-images": GREEN_TILES,
|
|
}
|
|
}
|
|
}
|
|
self._assert_json_contents(output_json_path, expected_summary_dict)
|
|
self._assert_directory_contents(
|
|
mismatch_path_dir,
|
|
['red_skp-tile0.png', 'red_skp-tile1.png', 'red_skp-tile2.png',
|
|
'red_skp-tile3.png', 'red_skp-tile4.png', 'red_skp-tile5.png',
|
|
])
|
|
|
|
def test_tiled_writeChecksumBasedFilenames(self):
|
|
"""Same as test_tiled, but with --writeChecksumBasedFilenames."""
|
|
output_json_path = os.path.join(self._output_dir, 'actuals.json')
|
|
write_path_dir = self.create_empty_dir(
|
|
path=os.path.join(self._output_dir, 'writePath'))
|
|
self._generate_skps()
|
|
self._run_render_pictures(['-r', self._input_skp_dir,
|
|
'--bbh', 'grid', '256', '256',
|
|
'--mode', 'tile', '256', '256',
|
|
'--writeChecksumBasedFilenames',
|
|
'--writePath', write_path_dir,
|
|
'--writeJsonSummaryPath', output_json_path])
|
|
expected_summary_dict = {
|
|
"header" : EXPECTED_HEADER_CONTENTS,
|
|
"actual-results" : {
|
|
"red.skp": {
|
|
# Manually verified these 6 images, all 256x256 tiles,
|
|
# consistent with a tiled version of the 640x400 red rect
|
|
# with black borders.
|
|
"tiled-images": [{
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 5815827069051002745,
|
|
"comparisonResult" : "no-comparison",
|
|
"filepath" : "red_skp/bitmap-64bitMD5_5815827069051002745.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 9323613075234140270,
|
|
"comparisonResult" : "no-comparison",
|
|
"filepath" : "red_skp/bitmap-64bitMD5_9323613075234140270.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 16670399404877552232,
|
|
"comparisonResult" : "no-comparison",
|
|
"filepath" : "red_skp/bitmap-64bitMD5_16670399404877552232.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 2507897274083364964,
|
|
"comparisonResult" : "no-comparison",
|
|
"filepath" : "red_skp/bitmap-64bitMD5_2507897274083364964.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 7325267995523877959,
|
|
"comparisonResult" : "no-comparison",
|
|
"filepath" : "red_skp/bitmap-64bitMD5_7325267995523877959.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 2181381724594493116,
|
|
"comparisonResult" : "no-comparison",
|
|
"filepath" : "red_skp/bitmap-64bitMD5_2181381724594493116.png",
|
|
}],
|
|
},
|
|
"green.skp": {
|
|
# Manually verified these 6 images, all 256x256 tiles,
|
|
# consistent with a tiled version of the 640x400 green rect
|
|
# with black borders.
|
|
"tiled-images": [{
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 12587324416545178013,
|
|
"comparisonResult" : "no-comparison",
|
|
"filepath" : "green_skp/bitmap-64bitMD5_12587324416545178013.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 7624374914829746293,
|
|
"comparisonResult" : "no-comparison",
|
|
"filepath" : "green_skp/bitmap-64bitMD5_7624374914829746293.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 5686489729535631913,
|
|
"comparisonResult" : "no-comparison",
|
|
"filepath" : "green_skp/bitmap-64bitMD5_5686489729535631913.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 7980646035555096146,
|
|
"comparisonResult" : "no-comparison",
|
|
"filepath" : "green_skp/bitmap-64bitMD5_7980646035555096146.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 17817086664365875131,
|
|
"comparisonResult" : "no-comparison",
|
|
"filepath" : "green_skp/bitmap-64bitMD5_17817086664365875131.png",
|
|
}, {
|
|
"checksumAlgorithm" : "bitmap-64bitMD5",
|
|
"checksumValue" : 10673669813016809363,
|
|
"comparisonResult" : "no-comparison",
|
|
"filepath" : "green_skp/bitmap-64bitMD5_10673669813016809363.png",
|
|
}],
|
|
}
|
|
}
|
|
}
|
|
self._assert_json_contents(output_json_path, expected_summary_dict)
|
|
self._assert_directory_contents(write_path_dir, ['red_skp', 'green_skp'])
|
|
self._assert_directory_contents(
|
|
os.path.join(write_path_dir, 'red_skp'),
|
|
['bitmap-64bitMD5_5815827069051002745.png',
|
|
'bitmap-64bitMD5_9323613075234140270.png',
|
|
'bitmap-64bitMD5_16670399404877552232.png',
|
|
'bitmap-64bitMD5_2507897274083364964.png',
|
|
'bitmap-64bitMD5_7325267995523877959.png',
|
|
'bitmap-64bitMD5_2181381724594493116.png'])
|
|
self._assert_directory_contents(
|
|
os.path.join(write_path_dir, 'green_skp'),
|
|
['bitmap-64bitMD5_12587324416545178013.png',
|
|
'bitmap-64bitMD5_7624374914829746293.png',
|
|
'bitmap-64bitMD5_5686489729535631913.png',
|
|
'bitmap-64bitMD5_7980646035555096146.png',
|
|
'bitmap-64bitMD5_17817086664365875131.png',
|
|
'bitmap-64bitMD5_10673669813016809363.png'])
|
|
|
|
def _run_render_pictures(self, args):
|
|
binary = self.find_path_to_program('render_pictures')
|
|
return self.run_command([binary,
|
|
'--clone', '1',
|
|
'--config', '8888',
|
|
] + args)
|
|
|
|
def _create_expectations(self, missing_some_images=False,
|
|
rel_path='expectations.json'):
|
|
"""Creates expectations JSON file within self._expectations_dir .
|
|
|
|
Args:
|
|
missing_some_images: (bool) whether to remove expectations for a subset
|
|
of the images
|
|
rel_path: (string) relative path within self._expectations_dir to write
|
|
the expectations into
|
|
|
|
Returns: full path to the expectations file created.
|
|
"""
|
|
expectations_dict = {
|
|
"header" : EXPECTED_HEADER_CONTENTS,
|
|
"expected-results" : {
|
|
# red.skp: these should fail the comparison
|
|
"red.skp": {
|
|
"tiled-images": modified_list_of_dicts(
|
|
RED_TILES, {'checksumValue': 11111}),
|
|
"whole-image": modified_dict(
|
|
RED_WHOLEIMAGE, {'checksumValue': 22222}),
|
|
},
|
|
# green.skp: these should pass the comparison
|
|
"green.skp": {
|
|
"tiled-images": GREEN_TILES,
|
|
"whole-image": GREEN_WHOLEIMAGE,
|
|
}
|
|
}
|
|
}
|
|
if missing_some_images:
|
|
del expectations_dict['expected-results']['red.skp']['whole-image']
|
|
del expectations_dict['expected-results']['red.skp']['tiled-images'][-1]
|
|
path = os.path.join(self._expectations_dir, rel_path)
|
|
with open(path, 'w') as fh:
|
|
json.dump(expectations_dict, fh)
|
|
return path
|
|
|
|
def _generate_skps(self):
|
|
"""Runs the skpmaker binary to generate files in self._input_skp_dir."""
|
|
self._run_skpmaker(
|
|
output_path=os.path.join(self._input_skp_dir, 'red.skp'), red=255)
|
|
self._run_skpmaker(
|
|
output_path=os.path.join(self._input_skp_dir, 'green.skp'), green=255)
|
|
|
|
def _run_skpmaker(self, output_path, red=0, green=0, blue=0,
|
|
width=640, height=400):
|
|
"""Runs the skpmaker binary to generate SKP with known characteristics.
|
|
|
|
Args:
|
|
output_path: Filepath to write the SKP into.
|
|
red: Value of red color channel in image, 0-255.
|
|
green: Value of green color channel in image, 0-255.
|
|
blue: Value of blue color channel in image, 0-255.
|
|
width: Width of canvas to create.
|
|
height: Height of canvas to create.
|
|
"""
|
|
binary = self.find_path_to_program('skpmaker')
|
|
return self.run_command([binary,
|
|
'--red', str(red),
|
|
'--green', str(green),
|
|
'--blue', str(blue),
|
|
'--width', str(width),
|
|
'--height', str(height),
|
|
'--writePath', str(output_path),
|
|
])
|
|
|
|
def _assert_directory_contents(self, dir_path, expected_filenames):
|
|
"""Asserts that files found in a dir are identical to expected_filenames.
|
|
|
|
Args:
|
|
dir_path: Path to a directory on local disk.
|
|
expected_filenames: Set containing the expected filenames within the dir.
|
|
|
|
Raises:
|
|
AssertionError: contents of the directory are not identical to
|
|
expected_filenames.
|
|
"""
|
|
self.assertEqual(set(os.listdir(dir_path)), set(expected_filenames))
|
|
|
|
def _assert_json_contents(self, json_path, expected_dict):
|
|
"""Asserts that contents of a JSON file are identical to expected_dict.
|
|
|
|
Args:
|
|
json_path: Path to a JSON file.
|
|
expected_dict: Dictionary indicating the expected contents of the JSON
|
|
file.
|
|
|
|
Raises:
|
|
AssertionError: contents of the JSON file are not identical to
|
|
expected_dict.
|
|
"""
|
|
prettyprinted_expected_dict = json.dumps(expected_dict, sort_keys=True,
|
|
indent=2)
|
|
with open(json_path, 'r') as fh:
|
|
prettyprinted_json_dict = json.dumps(json.load(fh), sort_keys=True,
|
|
indent=2)
|
|
self.assertMultiLineEqual(prettyprinted_expected_dict,
|
|
prettyprinted_json_dict)
|
|
|
|
|
|
def main():
|
|
base_unittest.main(RenderPicturesTest)
|
|
|
|
|
|
if __name__ == '__main__':
|
|
main()
|