a0ba612bb2
To get the gold images out of the browser tests, this adds testReporter.js and pathkit_aggregator.go. testReporter bundles up the output as a base64 encoded PNG and sends it over the local network to pathkit_aggregator. pathkit_aggregator will keep a list of test results reported in this way and write the PNGs to /OUT of the container (which is the swarming output directory). Finally, after all the tests are run, the helper script "test_pathkit.sh" makes a POST request that creates the JSON file that gold expects (following the schema https://github.com/google/skia-buildbot/blob/master/golden/docs/INGESTION.md) pathkit_aggregator takes many command line arguments which control the keys that Gold needs in order to ingest and handle the data. Of note, this creates a new set (i.e. source_type) of gold images called "pathkit", which will distinguish it from "gm", "image", etc. There will be at least 2 sub-sets of "pathkit" images, "canvas" and "svg", (representing the 2 output types of PathKit). This CL doesn't quite handle SVG yet, as it needs a way to convert SVG to PNG in the browser and will be addressed in a follow up CL. A "standard" gm is sized at 600x600. This was arbitrarily picked. Note that the functions in testReporter.js return Promises based on the fetch requests to post the data. This eliminates the race condition between the /report_gold_data and /dump_json since running the karma tests won't return until all reports are done. Other changes of note: - Adds go to karma-chrome-tests container. - renames recipe_modules/build/wasm.py -> pathkit.py to be consistent with the name of test_pathkit.py and make for easier grepping. - Increases the JS test timeout to 10s (up from 5) to hopefully avoid the flakes seen in the Debug Test. Bug: skia:8216 Change-Id: Ic2cad54f3d19cc16601cf2e9a87798db1e6887a2 Reviewed-on: https://skia-review.googlesource.com/147042 Reviewed-by: Stephan Altmueller <stephana@google.com> |
||
---|---|---|
.. | ||
android_compile.expected | ||
bookmaker.expected | ||
calmbench.expected | ||
check_generated_files.expected | ||
compile.expected | ||
compute_test.expected | ||
ct_skps.expected | ||
housekeeper.expected | ||
infra.expected | ||
perf.expected | ||
recreate_skps.expected | ||
skpbench.expected | ||
skqp_test.expected | ||
test_pathkit.expected | ||
test.expected | ||
upload_calmbench_results.expected | ||
upload_coverage_results.expected | ||
upload_dm_results.expected | ||
upload_nano_results.expected | ||
upload_skiaserve.expected | ||
android_compile.py | ||
bookmaker.py | ||
calmbench.py | ||
check_generated_files.py | ||
compile.py | ||
compute_test.py | ||
ct_skps.py | ||
housekeeper.py | ||
infra.py | ||
perf.py | ||
README.md | ||
recreate_skps.py | ||
skpbench.py | ||
skqp_test.py | ||
test_pathkit.py | ||
test.py | ||
upload_calmbench_results.py | ||
upload_coverage_results.py | ||
upload_dm_results.py | ||
upload_nano_results.py | ||
upload_skiaserve.py |
Skia Recipes
These are the top-level scripts which run inside of Swarming tasks to perform all of Skia's automated testing.
To run a recipe locally:
$ python infra/bots/recipes.py run --workdir=/tmp/<workdir> <recipe name without .py> key1=value1 key2=value2 ...
Each recipe may have its own required properties which must be entered as key/value pairs in the command.
When you change a recipe, you generally need to re-train the simulation test:
$ python infra/bots/recipes.py test train
Or:
$ cd infra/bots; make train
The test generates expectations files for the tests contained within each recipe which illustrate which steps would run, given a particular set of inputs. Pay attention to the diffs in these files when making changes to ensure that your change has the intended effect.