skia2/tools/skqp/README_ALGORITHM.md
Hal Canary ac7f23c807 SkQP: refatctor C++ bits.
* C++ code moved into tools/skqp/src/.
  * State held with single SkQP class.
  * gmkb functions moved to skqp_model.{h,cpp}
  * model no longer knows about report format.
  * skqp_main and skqp_lib no longer have globals
  * jni code has fewer globals.
  * skqp_main no longer uses googletest.
  * AssetMng returns SkData, not a SkStream.
  * Add jitter tool.
  * dump GPU information into grdump.txt
  * JUnit puts report in directory with timestamp.
  * Document SkQP Render Test Algorithm.
  * GPU driver correctness workarounds always off
  * cut_release tool for assembling models
  * make_rendertests_list.py to help cut_release
  * make_gmkb.go emits a list of models

CQ_INCLUDE_TRYBOTS=skia.primary:Build-Debian9-Clang-x86-devrel-Android_SKQP

Change-Id: I7d4f0c24592b1f64be0088578a3f1a0bc366dd4d
Reviewed-on: https://skia-review.googlesource.com/c/110420
Reviewed-by: Hal Canary <halcanary@google.com>
Commit-Queue: Hal Canary <halcanary@google.com>
2018-11-30 18:38:00 +00:00

97 lines
3.9 KiB
Markdown

SkQP Render Test Algorithm
==========================
The following is a description of the render test validation algorithm that
will be used by the version of SkQP that will be released for Android Q-release.
There is a global macro constant: `SK_SKQP_GLOBAL_ERROR_TOLERANCE`, which
reflects the `gn` variable `skia_skqp_global_error_tolerance`. This is usually
set to 8.
First, look for a file named `skqp/rendertests.txt` in the
`platform_tools/android/apps/skqp/src/main/assets` directory. The format of
this file is: each line contains one render test name, followed by a comma,
followed by an integer. The integer is the `passing_threshold` for that test.
For each test, we have a `max_image` and a `min_image`. These are PNG-encoded
images stored in SkQP's APK's asset directory (in the paths `gmkb/${TEST}/min.png`
and `gmkb/${TEST}/max.png`).
The test input is a rendered image. This will be produced by running one of
the render tests against the either the `vk` (Vulkan) or `gles` (OpenGL ES)
Skia backend.
Here is psuedocode for the error calculation:
function calculate_pixel_error(pixel_value, pixel_max, pixel_min):
pixel_error = 0
for color_channel in { red, green, blue, alpha }:
value = get_color(pixel_value, color_channel)
v_max = get_color(pixel_max, color_channel)
v_min = get_color(pixel_min, color_channel)
if value > v_max:
channel_error = value - v_max
elif value < v_min:
channel_error = v_min - value
else:
channel_error = 0
pixel_error = max(pixel_error, channel_error)
return max(0, pixel_error - SK_SKQP_GLOBAL_ERROR_TOLERANCE);
function get_error(rendered_image, max_image, min_image):
assert(dimensions(rendered_image) == dimensions(max_image))
assert(dimensions(rendered_image) == dimensions(min_image))
max_error = 0
bad_pixels = 0
total_error = 0
error_image = allocate_bitmap(dimensions(rendered_image))
for xy in list_all_pixel_coordinates(rendered_image):
pixel_error = calculate_pixel_error(rendered_image(xy),
max_image(xy),
min_image(xy))
if pixel_error > 0:
for neighboring_xy in find_neighbors(xy):
if not inside(neighboring_xy, dimensions(rendered_image)):
continue
pixel_error = min(pixel_error,
calculate_pixel_error(rendered_image(xy),
max_image(neighboring_xy),
min_image(neighboring_xy)))
if pixel_error > 0:
max_error = max(max_error, pixel_error)
bad_pixels += 1
total_error += pixel_error
error_image(xy) = linear_interpolation(black, red, pixel_error)
else:
error_image(xy) = white
return ((total_error, max_error, bad_pixels), error_image)
For each render test, there is a threshold value for `total_error`, :
`passing_threshold`.
If `passing_threshold >= 0 && total_error > passing_threshold`, then the test
is a failure and is included in the report. if `passing_threshold == -1`, then
the test always passes, but we do execute the test to verify that the driver
does not crash.
We generate a report with the following information for each test:
backend_name,render_test_name,max_error,bad_pixels,total_error
in CSV format in the file `out.csv`. A HTML report of just the failing tests
is written to the file `report.html`. This version includes four images for
each test: `rendered_image`, `max_image`, `min_image`, and `error_image`, as
well as the three metrics: `max_error`, `bad_pixels`, and `total_error`.