e3bb3bce3e
A new command-line flag --outResultsFile takes the filename to write the JSON into. The human readable output is the same as before with one exception. Previously DEBUG would be printed if in debug mode, or nothing would be printed if in release mode. Now this is reported as a named option: build=DEBUG or build=RELEASE R=djsollen@google.com, bsalomon@google.com Author: jcgregorio@google.com Review URL: https://codereview.chromium.org/83863002 git-svn-id: http://skia.googlecode.com/svn/trunk@12465 2bbb7eff-a529-9590-31e7-b0007b416f81
2386 lines
97 KiB
C++
2386 lines
97 KiB
C++
/*
|
|
* Copyright 2011 Google Inc.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license that can be
|
|
* found in the LICENSE file.
|
|
*/
|
|
|
|
/*
|
|
* Code for the "gm" (Golden Master) rendering comparison tool.
|
|
*
|
|
* If you make changes to this, re-run the self-tests at gm/tests/run.sh
|
|
* to make sure they still pass... you may need to change the expected
|
|
* results of the self-test.
|
|
*/
|
|
|
|
#include "gm.h"
|
|
#include "gm_error.h"
|
|
#include "gm_expectations.h"
|
|
#include "system_preferences.h"
|
|
#include "SkBitmap.h"
|
|
#include "SkColorPriv.h"
|
|
#include "SkCommandLineFlags.h"
|
|
#include "SkData.h"
|
|
#include "SkDeferredCanvas.h"
|
|
#include "SkDevice.h"
|
|
#include "SkDocument.h"
|
|
#include "SkDrawFilter.h"
|
|
#include "SkForceLinking.h"
|
|
#include "SkGPipe.h"
|
|
#include "SkGraphics.h"
|
|
#include "SkImageDecoder.h"
|
|
#include "SkImageEncoder.h"
|
|
#include "SkJSONCPP.h"
|
|
#include "SkOSFile.h"
|
|
#include "SkPDFRasterizer.h"
|
|
#include "SkPicture.h"
|
|
#include "SkRefCnt.h"
|
|
#include "SkScalar.h"
|
|
#include "SkStream.h"
|
|
#include "SkString.h"
|
|
#include "SkTArray.h"
|
|
#include "SkTDict.h"
|
|
#include "SkTileGridPicture.h"
|
|
#include "SamplePipeControllers.h"
|
|
|
|
#ifdef SK_DEBUG
|
|
static const bool kDebugOnly = true;
|
|
#else
|
|
static const bool kDebugOnly = false;
|
|
#endif
|
|
|
|
__SK_FORCE_IMAGE_DECODER_LINKING;
|
|
|
|
#if SK_SUPPORT_GPU
|
|
#include "GrContextFactory.h"
|
|
#include "SkGpuDevice.h"
|
|
typedef GrContextFactory::GLContextType GLContextType;
|
|
#define DEFAULT_CACHE_VALUE -1
|
|
static int gGpuCacheSizeBytes;
|
|
static int gGpuCacheSizeCount;
|
|
#else
|
|
class GrContextFactory;
|
|
class GrContext;
|
|
class GrSurface;
|
|
typedef int GLContextType;
|
|
#endif
|
|
|
|
#define DEBUGFAIL_SEE_STDERR SkDEBUGFAIL("see stderr for message")
|
|
|
|
extern bool gSkSuppressFontCachePurgeSpew;
|
|
DECLARE_bool(useDocumentInsteadOfDevice);
|
|
|
|
#ifdef SK_SUPPORT_PDF
|
|
#include "SkPDFDevice.h"
|
|
#include "SkPDFDocument.h"
|
|
#endif
|
|
|
|
// Until we resolve http://code.google.com/p/skia/issues/detail?id=455 ,
|
|
// stop writing out XPS-format image baselines in gm.
|
|
#undef SK_SUPPORT_XPS
|
|
#ifdef SK_SUPPORT_XPS
|
|
#include "SkXPSDevice.h"
|
|
#endif
|
|
|
|
#ifdef SK_BUILD_FOR_MAC
|
|
#include "SkCGUtils.h"
|
|
#endif
|
|
|
|
using namespace skiagm;
|
|
|
|
class Iter {
|
|
public:
|
|
Iter() {
|
|
this->reset();
|
|
}
|
|
|
|
void reset() {
|
|
fReg = GMRegistry::Head();
|
|
}
|
|
|
|
GM* next() {
|
|
if (fReg) {
|
|
GMRegistry::Factory fact = fReg->factory();
|
|
fReg = fReg->next();
|
|
return fact(0);
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static int Count() {
|
|
const GMRegistry* reg = GMRegistry::Head();
|
|
int count = 0;
|
|
while (reg) {
|
|
count += 1;
|
|
reg = reg->next();
|
|
}
|
|
return count;
|
|
}
|
|
|
|
private:
|
|
const GMRegistry* fReg;
|
|
};
|
|
|
|
// TODO(epoger): Right now, various places in this code assume that all the
|
|
// image files read/written by GM use this file extension.
|
|
// Search for references to this constant to find these assumptions.
|
|
const static char kPNG_FileExtension[] = "png";
|
|
|
|
enum Backend {
|
|
kRaster_Backend,
|
|
kGPU_Backend,
|
|
kPDF_Backend,
|
|
kXPS_Backend,
|
|
};
|
|
|
|
enum BbhType {
|
|
kNone_BbhType,
|
|
kRTree_BbhType,
|
|
kTileGrid_BbhType,
|
|
};
|
|
|
|
enum ConfigFlags {
|
|
kNone_ConfigFlag = 0x0,
|
|
/* Write GM images if a write path is provided. */
|
|
kWrite_ConfigFlag = 0x1,
|
|
/* Read reference GM images if a read path is provided. */
|
|
kRead_ConfigFlag = 0x2,
|
|
kRW_ConfigFlag = (kWrite_ConfigFlag | kRead_ConfigFlag),
|
|
};
|
|
|
|
struct ConfigData {
|
|
SkBitmap::Config fConfig;
|
|
Backend fBackend;
|
|
GLContextType fGLContextType; // GPU backend only
|
|
int fSampleCnt; // GPU backend only
|
|
ConfigFlags fFlags;
|
|
const char* fName;
|
|
bool fRunByDefault;
|
|
};
|
|
|
|
struct PDFRasterizerData {
|
|
bool (*fRasterizerFunction)(SkStream*, SkBitmap*);
|
|
const char* fName;
|
|
bool fRunByDefault;
|
|
};
|
|
|
|
class BWTextDrawFilter : public SkDrawFilter {
|
|
public:
|
|
virtual bool filter(SkPaint*, Type) SK_OVERRIDE;
|
|
};
|
|
bool BWTextDrawFilter::filter(SkPaint* p, Type t) {
|
|
if (kText_Type == t) {
|
|
p->setAntiAlias(false);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
struct PipeFlagComboData {
|
|
const char* name;
|
|
uint32_t flags;
|
|
};
|
|
|
|
static PipeFlagComboData gPipeWritingFlagCombos[] = {
|
|
{ "", 0 },
|
|
{ " cross-process", SkGPipeWriter::kCrossProcess_Flag },
|
|
{ " cross-process, shared address", SkGPipeWriter::kCrossProcess_Flag
|
|
| SkGPipeWriter::kSharedAddressSpace_Flag }
|
|
};
|
|
|
|
static SkData* encode_to_dct_data(size_t* pixelRefOffset, const SkBitmap& bitmap);
|
|
DECLARE_int32(pdfRasterDpi);
|
|
|
|
const static ErrorCombination kDefaultIgnorableErrorTypes = ErrorCombination()
|
|
.plus(kMissingExpectations_ErrorType)
|
|
.plus(kIntentionallySkipped_ErrorType);
|
|
|
|
class GMMain {
|
|
public:
|
|
GMMain() : fUseFileHierarchy(false), fWriteChecksumBasedFilenames(false),
|
|
fIgnorableErrorTypes(kDefaultIgnorableErrorTypes),
|
|
fMismatchPath(NULL), fMissingExpectationsPath(NULL), fTestsRun(0),
|
|
fRenderModesEncountered(1) {}
|
|
|
|
/**
|
|
* Assemble shortNamePlusConfig from (surprise!) shortName and configName.
|
|
*
|
|
* The method for doing so depends on whether we are using hierarchical naming.
|
|
* For example, shortName "selftest1" and configName "8888" could be assembled into
|
|
* either "selftest1_8888" or "8888/selftest1".
|
|
*/
|
|
SkString make_shortname_plus_config(const char *shortName, const char *configName) {
|
|
SkString name;
|
|
if (0 == strlen(configName)) {
|
|
name.append(shortName);
|
|
} else if (fUseFileHierarchy) {
|
|
name.appendf("%s%c%s", configName, SkPATH_SEPARATOR, shortName);
|
|
} else {
|
|
name.appendf("%s_%s", shortName, configName);
|
|
}
|
|
return name;
|
|
}
|
|
|
|
/**
|
|
* Assemble filename, suitable for writing out the results of a particular test.
|
|
*/
|
|
SkString make_filename(const char *path,
|
|
const char *shortName,
|
|
const char *configName,
|
|
const char *renderModeDescriptor,
|
|
const char *suffix) {
|
|
SkString filename = make_shortname_plus_config(shortName, configName);
|
|
filename.append(renderModeDescriptor);
|
|
filename.appendUnichar('.');
|
|
filename.append(suffix);
|
|
return SkOSPath::SkPathJoin(path, filename.c_str());
|
|
}
|
|
|
|
/**
|
|
* Assemble filename suitable for writing out an SkBitmap.
|
|
*/
|
|
SkString make_bitmap_filename(const char *path,
|
|
const char *shortName,
|
|
const char *configName,
|
|
const char *renderModeDescriptor,
|
|
const GmResultDigest &bitmapDigest) {
|
|
if (fWriteChecksumBasedFilenames) {
|
|
SkString filename;
|
|
filename.append(bitmapDigest.getHashType());
|
|
filename.appendUnichar('_');
|
|
filename.append(shortName);
|
|
filename.appendUnichar('_');
|
|
filename.append(bitmapDigest.getDigestValue());
|
|
filename.appendUnichar('.');
|
|
filename.append(kPNG_FileExtension);
|
|
return SkOSPath::SkPathJoin(path, filename.c_str());
|
|
} else {
|
|
return make_filename(path, shortName, configName, renderModeDescriptor,
|
|
kPNG_FileExtension);
|
|
}
|
|
}
|
|
|
|
/* since PNG insists on unpremultiplying our alpha, we take no
|
|
precision chances and force all pixels to be 100% opaque,
|
|
otherwise on compare we may not get a perfect match.
|
|
*/
|
|
static void force_all_opaque(const SkBitmap& bitmap) {
|
|
SkBitmap::Config config = bitmap.config();
|
|
switch (config) {
|
|
case SkBitmap::kARGB_8888_Config:
|
|
force_all_opaque_8888(bitmap);
|
|
break;
|
|
case SkBitmap::kRGB_565_Config:
|
|
// nothing to do here; 565 bitmaps are inherently opaque
|
|
break;
|
|
default:
|
|
gm_fprintf(stderr, "unsupported bitmap config %d\n", config);
|
|
DEBUGFAIL_SEE_STDERR;
|
|
}
|
|
}
|
|
|
|
static void force_all_opaque_8888(const SkBitmap& bitmap) {
|
|
SkAutoLockPixels lock(bitmap);
|
|
for (int y = 0; y < bitmap.height(); y++) {
|
|
for (int x = 0; x < bitmap.width(); x++) {
|
|
*bitmap.getAddr32(x, y) |= (SK_A32_MASK << SK_A32_SHIFT);
|
|
}
|
|
}
|
|
}
|
|
|
|
static ErrorCombination write_bitmap(const SkString& path, const SkBitmap& bitmap) {
|
|
// TODO(epoger): Now that we have removed force_all_opaque()
|
|
// from this method, we should be able to get rid of the
|
|
// transformation to 8888 format also.
|
|
SkBitmap copy;
|
|
bitmap.copyTo(©, SkBitmap::kARGB_8888_Config);
|
|
if (!SkImageEncoder::EncodeFile(path.c_str(), copy,
|
|
SkImageEncoder::kPNG_Type,
|
|
100)) {
|
|
gm_fprintf(stderr, "FAILED to write bitmap: %s\n", path.c_str());
|
|
return ErrorCombination(kWritingReferenceImage_ErrorType);
|
|
}
|
|
return kEmpty_ErrorCombination;
|
|
}
|
|
|
|
/**
|
|
* Add all render modes encountered thus far to the "modes" array.
|
|
*/
|
|
void GetRenderModesEncountered(SkTArray<SkString> &modes) {
|
|
SkTDict<int>::Iter iter(this->fRenderModesEncountered);
|
|
const char* mode;
|
|
while ((mode = iter.next(NULL)) != NULL) {
|
|
SkString modeAsString = SkString(mode);
|
|
// TODO(epoger): It seems a bit silly that all of these modes were
|
|
// recorded with a leading "-" which we have to remove here
|
|
// (except for mode "", which means plain old original mode).
|
|
// But that's how renderModeDescriptor has been passed into
|
|
// compare_test_results_to_reference_bitmap() historically,
|
|
// and changing that now may affect other parts of our code.
|
|
if (modeAsString.startsWith("-")) {
|
|
modeAsString.remove(0, 1);
|
|
modes.push_back(modeAsString);
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Returns true if failures on this test should be ignored.
|
|
*/
|
|
bool ShouldIgnoreTest(const SkString &name) const {
|
|
for (int i = 0; i < fIgnorableTestSubstrings.count(); i++) {
|
|
if (name.contains(fIgnorableTestSubstrings[i].c_str())) {
|
|
return true;
|
|
}
|
|
}
|
|
return false;
|
|
}
|
|
|
|
/**
|
|
* Records the results of this test in fTestsRun and fFailedTests.
|
|
*
|
|
* We even record successes, and errors that we regard as
|
|
* "ignorable"; we can filter them out later.
|
|
*/
|
|
void RecordTestResults(const ErrorCombination& errorCombination,
|
|
const SkString& shortNamePlusConfig,
|
|
const char renderModeDescriptor []) {
|
|
// Things to do regardless of errorCombination.
|
|
fTestsRun++;
|
|
int renderModeCount = 0;
|
|
this->fRenderModesEncountered.find(renderModeDescriptor, &renderModeCount);
|
|
renderModeCount++;
|
|
this->fRenderModesEncountered.set(renderModeDescriptor, renderModeCount);
|
|
|
|
if (errorCombination.isEmpty()) {
|
|
return;
|
|
}
|
|
|
|
// Things to do only if there is some error condition.
|
|
SkString fullName = shortNamePlusConfig;
|
|
fullName.append(renderModeDescriptor);
|
|
for (int typeInt = 0; typeInt <= kLast_ErrorType; typeInt++) {
|
|
ErrorType type = static_cast<ErrorType>(typeInt);
|
|
if (errorCombination.includes(type)) {
|
|
fFailedTests[type].push_back(fullName);
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Return the number of significant (non-ignorable) errors we have
|
|
* encountered so far.
|
|
*/
|
|
int NumSignificantErrors() {
|
|
int significantErrors = 0;
|
|
for (int typeInt = 0; typeInt <= kLast_ErrorType; typeInt++) {
|
|
ErrorType type = static_cast<ErrorType>(typeInt);
|
|
if (!fIgnorableErrorTypes.includes(type)) {
|
|
significantErrors += fFailedTests[type].count();
|
|
}
|
|
}
|
|
return significantErrors;
|
|
}
|
|
|
|
/**
|
|
* Display the summary of results with this ErrorType.
|
|
*
|
|
* @param type which ErrorType
|
|
* @param verbose whether to be all verbose about it
|
|
*/
|
|
void DisplayResultTypeSummary(ErrorType type, bool verbose) {
|
|
bool isIgnorableType = fIgnorableErrorTypes.includes(type);
|
|
|
|
SkString line;
|
|
if (isIgnorableType) {
|
|
line.append("[ ] ");
|
|
} else {
|
|
line.append("[*] ");
|
|
}
|
|
|
|
SkTArray<SkString> *failedTestsOfThisType = &fFailedTests[type];
|
|
int count = failedTestsOfThisType->count();
|
|
line.appendf("%d %s", count, getErrorTypeName(type));
|
|
if (!isIgnorableType || verbose) {
|
|
line.append(":");
|
|
for (int i = 0; i < count; ++i) {
|
|
line.append(" ");
|
|
line.append((*failedTestsOfThisType)[i]);
|
|
}
|
|
}
|
|
gm_fprintf(stdout, "%s\n", line.c_str());
|
|
}
|
|
|
|
/**
|
|
* List contents of fFailedTests to stdout.
|
|
*
|
|
* @param verbose whether to be all verbose about it
|
|
*/
|
|
void ListErrors(bool verbose) {
|
|
// First, print a single summary line.
|
|
SkString summary;
|
|
summary.appendf("Ran %d tests:", fTestsRun);
|
|
for (int typeInt = 0; typeInt <= kLast_ErrorType; typeInt++) {
|
|
ErrorType type = static_cast<ErrorType>(typeInt);
|
|
summary.appendf(" %s=%d", getErrorTypeName(type), fFailedTests[type].count());
|
|
}
|
|
gm_fprintf(stdout, "%s\n", summary.c_str());
|
|
|
|
// Now, for each failure type, list the tests that failed that way.
|
|
for (int typeInt = 0; typeInt <= kLast_ErrorType; typeInt++) {
|
|
this->DisplayResultTypeSummary(static_cast<ErrorType>(typeInt), verbose);
|
|
}
|
|
gm_fprintf(stdout, "(results marked with [*] will cause nonzero return value)\n");
|
|
}
|
|
|
|
static ErrorCombination write_document(const SkString& path, SkStreamAsset* asset) {
|
|
SkFILEWStream stream(path.c_str());
|
|
if (!stream.writeStream(asset, asset->getLength())) {
|
|
gm_fprintf(stderr, "FAILED to write document: %s\n", path.c_str());
|
|
return ErrorCombination(kWritingReferenceImage_ErrorType);
|
|
}
|
|
return kEmpty_ErrorCombination;
|
|
}
|
|
|
|
/**
|
|
* Prepare an SkBitmap to render a GM into.
|
|
*
|
|
* After you've rendered the GM into the SkBitmap, you must call
|
|
* complete_bitmap()!
|
|
*
|
|
* @todo thudson 22 April 2011 - could refactor this to take in
|
|
* a factory to generate the context, always call readPixels()
|
|
* (logically a noop for rasters, if wasted time), and thus collapse the
|
|
* GPU special case and also let this be used for SkPicture testing.
|
|
*/
|
|
static void setup_bitmap(const ConfigData& gRec, SkISize& size,
|
|
SkBitmap* bitmap) {
|
|
bitmap->setConfig(gRec.fConfig, size.width(), size.height());
|
|
bitmap->allocPixels();
|
|
bitmap->eraseColor(SK_ColorTRANSPARENT);
|
|
}
|
|
|
|
/**
|
|
* Any finalization steps we need to perform on the SkBitmap after
|
|
* we have rendered the GM into it.
|
|
*
|
|
* It's too bad that we are throwing away alpha channel data
|
|
* we could otherwise be examining, but this had always been happening
|
|
* before... it was buried within the compare() method at
|
|
* https://code.google.com/p/skia/source/browse/trunk/gm/gmmain.cpp?r=7289#305 .
|
|
*
|
|
* Apparently we need this, at least for bitmaps that are either:
|
|
* (a) destined to be written out as PNG files, or
|
|
* (b) compared against bitmaps read in from PNG files
|
|
* for the reasons described just above the force_all_opaque() method.
|
|
*
|
|
* Neglecting to do this led to the difficult-to-diagnose
|
|
* http://code.google.com/p/skia/issues/detail?id=1079 ('gm generating
|
|
* spurious pixel_error messages as of r7258')
|
|
*
|
|
* TODO(epoger): Come up with a better solution that allows us to
|
|
* compare full pixel data, including alpha channel, while still being
|
|
* robust in the face of transformations to/from PNG files.
|
|
* Options include:
|
|
*
|
|
* 1. Continue to call force_all_opaque(), but ONLY for bitmaps that
|
|
* will be written to, or compared against, PNG files.
|
|
* PRO: Preserve/compare alpha channel info for the non-PNG cases
|
|
* (comparing different renderModes in-memory)
|
|
* CON: The bitmaps (and hash digests) for these non-PNG cases would be
|
|
* different than those for the PNG-compared cases, and in the
|
|
* case of a failed renderMode comparison, how would we write the
|
|
* image to disk for examination?
|
|
*
|
|
* 2. Always compute image hash digests from PNG format (either
|
|
* directly from the the bytes of a PNG file, or capturing the
|
|
* bytes we would have written to disk if we were writing the
|
|
* bitmap out as a PNG).
|
|
* PRO: I think this would allow us to never force opaque, and to
|
|
* the extent that alpha channel data can be preserved in a PNG
|
|
* file, we could observe it.
|
|
* CON: If we read a bitmap from disk, we need to take its hash digest
|
|
* from the source PNG (we can't compute it from the bitmap we
|
|
* read out of the PNG, because we will have already premultiplied
|
|
* the alpha).
|
|
* CON: Seems wasteful to convert a bitmap to PNG format just to take
|
|
* its hash digest. (Although we're wasting lots of effort already
|
|
* calling force_all_opaque().)
|
|
*
|
|
* 3. Make the alpha premultiply/unpremultiply routines 100% consistent,
|
|
* so we can transform images back and forth without fear of off-by-one
|
|
* errors.
|
|
* CON: Math is hard.
|
|
*
|
|
* 4. Perform a "close enough" comparison of bitmaps (+/- 1 bit in each
|
|
* channel), rather than demanding absolute equality.
|
|
* CON: Can't do this with hash digests.
|
|
*/
|
|
static void complete_bitmap(SkBitmap* bitmap) {
|
|
force_all_opaque(*bitmap);
|
|
}
|
|
|
|
static void installFilter(SkCanvas* canvas);
|
|
|
|
static void invokeGM(GM* gm, SkCanvas* canvas, bool isPDF, bool isDeferred) {
|
|
SkAutoCanvasRestore acr(canvas, true);
|
|
|
|
if (!isPDF) {
|
|
canvas->concat(gm->getInitialTransform());
|
|
}
|
|
installFilter(canvas);
|
|
gm->setCanvasIsDeferred(isDeferred);
|
|
gm->draw(canvas);
|
|
canvas->setDrawFilter(NULL);
|
|
}
|
|
|
|
static ErrorCombination generate_image(GM* gm, const ConfigData& gRec,
|
|
GrSurface* gpuTarget,
|
|
SkBitmap* bitmap,
|
|
bool deferred) {
|
|
SkISize size (gm->getISize());
|
|
setup_bitmap(gRec, size, bitmap);
|
|
|
|
SkAutoTUnref<SkCanvas> canvas;
|
|
|
|
if (gRec.fBackend == kRaster_Backend) {
|
|
SkAutoTUnref<SkBaseDevice> device(SkNEW_ARGS(SkBitmapDevice, (*bitmap)));
|
|
if (deferred) {
|
|
canvas.reset(SkDeferredCanvas::Create(device));
|
|
} else {
|
|
canvas.reset(SkNEW_ARGS(SkCanvas, (device)));
|
|
}
|
|
invokeGM(gm, canvas, false, deferred);
|
|
canvas->flush();
|
|
}
|
|
#if SK_SUPPORT_GPU
|
|
else { // GPU
|
|
SkAutoTUnref<SkBaseDevice> device(SkGpuDevice::Create(gpuTarget));
|
|
if (deferred) {
|
|
canvas.reset(SkDeferredCanvas::Create(device));
|
|
} else {
|
|
canvas.reset(SkNEW_ARGS(SkCanvas, (device)));
|
|
}
|
|
invokeGM(gm, canvas, false, deferred);
|
|
// the device is as large as the current rendertarget, so
|
|
// we explicitly only readback the amount we expect (in
|
|
// size) overwrite our previous allocation
|
|
bitmap->setConfig(SkBitmap::kARGB_8888_Config, size.fWidth,
|
|
size.fHeight);
|
|
canvas->readPixels(bitmap, 0, 0);
|
|
}
|
|
#endif
|
|
complete_bitmap(bitmap);
|
|
return kEmpty_ErrorCombination;
|
|
}
|
|
|
|
static void generate_image_from_picture(GM* gm, const ConfigData& gRec,
|
|
SkPicture* pict, SkBitmap* bitmap,
|
|
SkScalar scale = SK_Scalar1,
|
|
bool tile = false) {
|
|
SkISize size = gm->getISize();
|
|
setup_bitmap(gRec, size, bitmap);
|
|
|
|
if (tile) {
|
|
// Generate the result image by rendering to tiles and accumulating
|
|
// the results in 'bitmap'
|
|
|
|
// This 16x16 tiling matches the settings applied to 'pict' in
|
|
// 'generate_new_picture'
|
|
SkISize tileSize = SkISize::Make(16, 16);
|
|
|
|
SkBitmap tileBM;
|
|
setup_bitmap(gRec, tileSize, &tileBM);
|
|
SkCanvas tileCanvas(tileBM);
|
|
installFilter(&tileCanvas);
|
|
|
|
SkCanvas bmpCanvas(*bitmap);
|
|
SkPaint bmpPaint;
|
|
bmpPaint.setXfermodeMode(SkXfermode::kSrc_Mode);
|
|
|
|
for (int yTile = 0; yTile < (size.height()+15)/16; ++yTile) {
|
|
for (int xTile = 0; xTile < (size.width()+15)/16; ++xTile) {
|
|
int saveCount = tileCanvas.save();
|
|
SkMatrix mat(tileCanvas.getTotalMatrix());
|
|
mat.postTranslate(SkIntToScalar(-xTile*tileSize.width()),
|
|
SkIntToScalar(-yTile*tileSize.height()));
|
|
tileCanvas.setMatrix(mat);
|
|
pict->draw(&tileCanvas);
|
|
tileCanvas.flush();
|
|
tileCanvas.restoreToCount(saveCount);
|
|
bmpCanvas.drawBitmap(tileBM,
|
|
SkIntToScalar(xTile * tileSize.width()),
|
|
SkIntToScalar(yTile * tileSize.height()),
|
|
&bmpPaint);
|
|
}
|
|
}
|
|
} else {
|
|
SkCanvas canvas(*bitmap);
|
|
installFilter(&canvas);
|
|
canvas.scale(scale, scale);
|
|
canvas.drawPicture(*pict);
|
|
complete_bitmap(bitmap);
|
|
}
|
|
}
|
|
|
|
static bool generate_pdf(GM* gm, SkDynamicMemoryWStream& pdf) {
|
|
#ifdef SK_SUPPORT_PDF
|
|
SkMatrix initialTransform = gm->getInitialTransform();
|
|
if (FLAGS_useDocumentInsteadOfDevice) {
|
|
SkISize pageISize = gm->getISize();
|
|
SkAutoTUnref<SkDocument> pdfDoc(
|
|
SkDocument::CreatePDF(&pdf, NULL,
|
|
encode_to_dct_data,
|
|
SkIntToScalar(FLAGS_pdfRasterDpi)));
|
|
|
|
if (!pdfDoc.get()) {
|
|
return false;
|
|
}
|
|
|
|
SkCanvas* canvas = NULL;
|
|
canvas = pdfDoc->beginPage(SkIntToScalar(pageISize.width()),
|
|
SkIntToScalar(pageISize.height()));
|
|
canvas->concat(initialTransform);
|
|
|
|
invokeGM(gm, canvas, true, false);
|
|
|
|
return pdfDoc->close();
|
|
} else {
|
|
SkISize pageSize = gm->getISize();
|
|
SkPDFDevice* dev = NULL;
|
|
if (initialTransform.isIdentity()) {
|
|
dev = new SkPDFDevice(pageSize, pageSize, initialTransform);
|
|
} else {
|
|
SkRect content = SkRect::MakeWH(SkIntToScalar(pageSize.width()),
|
|
SkIntToScalar(pageSize.height()));
|
|
initialTransform.mapRect(&content);
|
|
content.intersect(0, 0, SkIntToScalar(pageSize.width()),
|
|
SkIntToScalar(pageSize.height()));
|
|
SkISize contentSize =
|
|
SkISize::Make(SkScalarRoundToInt(content.width()),
|
|
SkScalarRoundToInt(content.height()));
|
|
dev = new SkPDFDevice(pageSize, contentSize, initialTransform);
|
|
}
|
|
dev->setDCTEncoder(encode_to_dct_data);
|
|
dev->setRasterDpi(SkIntToScalar(FLAGS_pdfRasterDpi));
|
|
SkAutoUnref aur(dev);
|
|
SkCanvas c(dev);
|
|
invokeGM(gm, &c, true, false);
|
|
SkPDFDocument doc;
|
|
doc.appendPage(dev);
|
|
doc.emitPDF(&pdf);
|
|
}
|
|
#endif // SK_SUPPORT_PDF
|
|
return true; // Do not report failure if pdf is not supported.
|
|
}
|
|
|
|
static void generate_xps(GM* gm, SkDynamicMemoryWStream& xps) {
|
|
#ifdef SK_SUPPORT_XPS
|
|
SkISize size = gm->getISize();
|
|
|
|
SkSize trimSize = SkSize::Make(SkIntToScalar(size.width()),
|
|
SkIntToScalar(size.height()));
|
|
static const SkScalar inchesPerMeter = SkScalarDiv(10000, 254);
|
|
static const SkScalar upm = 72 * inchesPerMeter;
|
|
SkVector unitsPerMeter = SkPoint::Make(upm, upm);
|
|
static const SkScalar ppm = 200 * inchesPerMeter;
|
|
SkVector pixelsPerMeter = SkPoint::Make(ppm, ppm);
|
|
|
|
SkXPSDevice* dev = new SkXPSDevice();
|
|
SkAutoUnref aur(dev);
|
|
|
|
SkCanvas c(dev);
|
|
dev->beginPortfolio(&xps);
|
|
dev->beginSheet(unitsPerMeter, pixelsPerMeter, trimSize);
|
|
invokeGM(gm, &c, false, false);
|
|
dev->endSheet();
|
|
dev->endPortfolio();
|
|
|
|
#endif
|
|
}
|
|
|
|
/**
|
|
* Log more detail about the mistmatch between expectedBitmap and
|
|
* actualBitmap.
|
|
*/
|
|
void report_bitmap_diffs(const SkBitmap& expectedBitmap, const SkBitmap& actualBitmap,
|
|
const char *testName) {
|
|
const int expectedWidth = expectedBitmap.width();
|
|
const int expectedHeight = expectedBitmap.height();
|
|
const int width = actualBitmap.width();
|
|
const int height = actualBitmap.height();
|
|
if ((expectedWidth != width) || (expectedHeight != height)) {
|
|
gm_fprintf(stderr, "---- %s: dimension mismatch --"
|
|
" expected [%d %d], actual [%d %d]\n",
|
|
testName, expectedWidth, expectedHeight, width, height);
|
|
return;
|
|
}
|
|
|
|
if ((SkBitmap::kARGB_8888_Config != expectedBitmap.config()) ||
|
|
(SkBitmap::kARGB_8888_Config != actualBitmap.config())) {
|
|
gm_fprintf(stderr, "---- %s: not computing max per-channel"
|
|
" pixel mismatch because non-8888\n", testName);
|
|
return;
|
|
}
|
|
|
|
SkAutoLockPixels alp0(expectedBitmap);
|
|
SkAutoLockPixels alp1(actualBitmap);
|
|
int errR = 0;
|
|
int errG = 0;
|
|
int errB = 0;
|
|
int errA = 0;
|
|
int differingPixels = 0;
|
|
|
|
for (int y = 0; y < height; ++y) {
|
|
const SkPMColor* expectedPixelPtr = expectedBitmap.getAddr32(0, y);
|
|
const SkPMColor* actualPixelPtr = actualBitmap.getAddr32(0, y);
|
|
for (int x = 0; x < width; ++x) {
|
|
SkPMColor expectedPixel = *expectedPixelPtr++;
|
|
SkPMColor actualPixel = *actualPixelPtr++;
|
|
if (expectedPixel != actualPixel) {
|
|
differingPixels++;
|
|
errR = SkMax32(errR, SkAbs32((int)SkGetPackedR32(expectedPixel) -
|
|
(int)SkGetPackedR32(actualPixel)));
|
|
errG = SkMax32(errG, SkAbs32((int)SkGetPackedG32(expectedPixel) -
|
|
(int)SkGetPackedG32(actualPixel)));
|
|
errB = SkMax32(errB, SkAbs32((int)SkGetPackedB32(expectedPixel) -
|
|
(int)SkGetPackedB32(actualPixel)));
|
|
errA = SkMax32(errA, SkAbs32((int)SkGetPackedA32(expectedPixel) -
|
|
(int)SkGetPackedA32(actualPixel)));
|
|
}
|
|
}
|
|
}
|
|
gm_fprintf(stderr, "---- %s: %d (of %d) differing pixels,"
|
|
" max per-channel mismatch R=%d G=%d B=%d A=%d\n",
|
|
testName, differingPixels, width*height, errR, errG, errB, errA);
|
|
}
|
|
|
|
/**
|
|
* Compares actual hash digest to expectations, returning the set of errors
|
|
* (if any) that we saw along the way.
|
|
*
|
|
* If fMismatchPath has been set, and there are pixel diffs, then the
|
|
* actual bitmap will be written out to a file within fMismatchPath.
|
|
* And similarly for fMissingExpectationsPath...
|
|
*
|
|
* @param expectations what expectations to compare actualBitmap against
|
|
* @param actualBitmapAndDigest the SkBitmap we actually generated, and its GmResultDigest
|
|
* @param shortName name of test, e.g. "selftest1"
|
|
* @param configName name of config, e.g. "8888"
|
|
* @param renderModeDescriptor e.g., "-rtree", "-deferred"
|
|
* @param addToJsonSummary whether to add these results (both actual and
|
|
* expected) to the JSON summary. Regardless of this setting, if
|
|
* we find an image mismatch in this test, we will write these
|
|
* results to the JSON summary. (This is so that we will always
|
|
* report errors across rendering modes, such as pipe vs tiled.
|
|
* See https://codereview.chromium.org/13650002/ )
|
|
*/
|
|
ErrorCombination compare_to_expectations(Expectations expectations,
|
|
const BitmapAndDigest& actualBitmapAndDigest,
|
|
const char *shortName, const char *configName,
|
|
const char *renderModeDescriptor,
|
|
bool addToJsonSummary) {
|
|
ErrorCombination errors;
|
|
SkString shortNamePlusConfig = make_shortname_plus_config(shortName, configName);
|
|
SkString completeNameString(shortNamePlusConfig);
|
|
completeNameString.append(renderModeDescriptor);
|
|
completeNameString.append(".");
|
|
completeNameString.append(kPNG_FileExtension);
|
|
const char* completeName = completeNameString.c_str();
|
|
|
|
if (expectations.empty()) {
|
|
errors.add(kMissingExpectations_ErrorType);
|
|
|
|
// Write out the "actuals" for any tests without expectations, if we have
|
|
// been directed to do so.
|
|
if (fMissingExpectationsPath) {
|
|
SkString path = make_bitmap_filename(fMissingExpectationsPath, shortName,
|
|
configName, renderModeDescriptor,
|
|
actualBitmapAndDigest.fDigest);
|
|
write_bitmap(path, actualBitmapAndDigest.fBitmap);
|
|
}
|
|
|
|
} else if (!expectations.match(actualBitmapAndDigest.fDigest)) {
|
|
addToJsonSummary = true;
|
|
// The error mode we record depends on whether this was running
|
|
// in a non-standard renderMode.
|
|
if ('\0' == *renderModeDescriptor) {
|
|
errors.add(kExpectationsMismatch_ErrorType);
|
|
} else {
|
|
errors.add(kRenderModeMismatch_ErrorType);
|
|
}
|
|
|
|
// Write out the "actuals" for any mismatches, if we have
|
|
// been directed to do so.
|
|
if (fMismatchPath) {
|
|
SkString path = make_bitmap_filename(fMismatchPath, shortName, configName,
|
|
renderModeDescriptor,
|
|
actualBitmapAndDigest.fDigest);
|
|
write_bitmap(path, actualBitmapAndDigest.fBitmap);
|
|
}
|
|
|
|
// If we have access to a single expected bitmap, log more
|
|
// detail about the mismatch.
|
|
const SkBitmap *expectedBitmapPtr = expectations.asBitmap();
|
|
if (NULL != expectedBitmapPtr) {
|
|
report_bitmap_diffs(*expectedBitmapPtr, actualBitmapAndDigest.fBitmap,
|
|
completeName);
|
|
}
|
|
}
|
|
|
|
if (addToJsonSummary) {
|
|
add_actual_results_to_json_summary(completeName, actualBitmapAndDigest.fDigest, errors,
|
|
expectations.ignoreFailure());
|
|
add_expected_results_to_json_summary(completeName, expectations);
|
|
}
|
|
|
|
return errors;
|
|
}
|
|
|
|
/**
|
|
* Add this result to the appropriate JSON collection of actual results (but just ONE),
|
|
* depending on errors encountered.
|
|
*/
|
|
void add_actual_results_to_json_summary(const char testName[],
|
|
const GmResultDigest &actualResultDigest,
|
|
ErrorCombination errors,
|
|
bool ignoreFailure) {
|
|
Json::Value jsonActualResults = actualResultDigest.asJsonTypeValuePair();
|
|
Json::Value *resultCollection = NULL;
|
|
|
|
if (errors.isEmpty()) {
|
|
resultCollection = &this->fJsonActualResults_Succeeded;
|
|
} else if (errors.includes(kRenderModeMismatch_ErrorType)) {
|
|
resultCollection = &this->fJsonActualResults_Failed;
|
|
} else if (errors.includes(kExpectationsMismatch_ErrorType)) {
|
|
if (ignoreFailure) {
|
|
resultCollection = &this->fJsonActualResults_FailureIgnored;
|
|
} else {
|
|
resultCollection = &this->fJsonActualResults_Failed;
|
|
}
|
|
} else if (errors.includes(kMissingExpectations_ErrorType)) {
|
|
// TODO: What about the case where there IS an expected
|
|
// image hash digest, but that gm test doesn't actually
|
|
// run? For now, those cases will always be ignored,
|
|
// because gm only looks at expectations that correspond
|
|
// to gm tests that were actually run.
|
|
//
|
|
// Once we have the ability to express expectations as a
|
|
// JSON file, we should fix this (and add a test case for
|
|
// which an expectation is given but the test is never
|
|
// run).
|
|
resultCollection = &this->fJsonActualResults_NoComparison;
|
|
}
|
|
|
|
// If none of the above cases match, we don't add it to ANY tally of actual results.
|
|
if (resultCollection) {
|
|
(*resultCollection)[testName] = jsonActualResults;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Add this test to the JSON collection of expected results.
|
|
*/
|
|
void add_expected_results_to_json_summary(const char testName[],
|
|
Expectations expectations) {
|
|
this->fJsonExpectedResults[testName] = expectations.asJsonValue();
|
|
}
|
|
|
|
/**
|
|
* Compare actualBitmap to expectations stored in this->fExpectationsSource.
|
|
*
|
|
* @param gm which test generated the actualBitmap
|
|
* @param gRec
|
|
* @param configName The config name to look for in the expectation file.
|
|
* @param actualBitmapAndDigest ptr to bitmap generated by this run, or NULL
|
|
* if we don't have a usable bitmap representation
|
|
*/
|
|
ErrorCombination compare_test_results_to_stored_expectations(
|
|
GM* gm, const ConfigData& gRec, const char* configName,
|
|
const BitmapAndDigest* actualBitmapAndDigest) {
|
|
|
|
SkString shortNamePlusConfig = make_shortname_plus_config(gm->shortName(), configName);
|
|
|
|
ErrorCombination errors;
|
|
|
|
if (NULL == actualBitmapAndDigest) {
|
|
// Note that we intentionally skipped validating the results for
|
|
// this test, because we don't know how to generate an SkBitmap
|
|
// version of the output.
|
|
errors.add(ErrorCombination(kIntentionallySkipped_ErrorType));
|
|
} else if (!(gRec.fFlags & kWrite_ConfigFlag)) {
|
|
// We don't record the results for this test or compare them
|
|
// against any expectations, because the output image isn't
|
|
// meaningful.
|
|
// See https://code.google.com/p/skia/issues/detail?id=1410 ('some
|
|
// GM result images not available for download from Google Storage')
|
|
errors.add(ErrorCombination(kIntentionallySkipped_ErrorType));
|
|
} else {
|
|
ExpectationsSource *expectationsSource = this->fExpectationsSource.get();
|
|
SkString nameWithExtension(shortNamePlusConfig);
|
|
nameWithExtension.append(".");
|
|
nameWithExtension.append(kPNG_FileExtension);
|
|
|
|
if (expectationsSource && (gRec.fFlags & kRead_ConfigFlag)) {
|
|
/*
|
|
* Get the expected results for this test, as one or more allowed
|
|
* hash digests. The current implementation of expectationsSource
|
|
* get this by computing the hash digest of a single PNG file on disk.
|
|
*
|
|
* TODO(epoger): This relies on the fact that
|
|
* force_all_opaque() was called on the bitmap before it
|
|
* was written to disk as a PNG in the first place. If
|
|
* not, the hash digest returned here may not match the
|
|
* hash digest of actualBitmap, which *has* been run through
|
|
* force_all_opaque().
|
|
* See comments above complete_bitmap() for more detail.
|
|
*/
|
|
Expectations expectations = expectationsSource->get(nameWithExtension.c_str());
|
|
if (this->ShouldIgnoreTest(shortNamePlusConfig)) {
|
|
expectations.setIgnoreFailure(true);
|
|
}
|
|
errors.add(compare_to_expectations(expectations, *actualBitmapAndDigest,
|
|
gm->shortName(), configName, "", true));
|
|
} else {
|
|
// If we are running without expectations, we still want to
|
|
// record the actual results.
|
|
add_actual_results_to_json_summary(nameWithExtension.c_str(),
|
|
actualBitmapAndDigest->fDigest,
|
|
ErrorCombination(kMissingExpectations_ErrorType),
|
|
false);
|
|
errors.add(ErrorCombination(kMissingExpectations_ErrorType));
|
|
}
|
|
}
|
|
return errors;
|
|
}
|
|
|
|
/**
|
|
* Compare actualBitmap to referenceBitmap.
|
|
*
|
|
* @param shortName test name, e.g. "selftest1"
|
|
* @param configName configuration name, e.g. "8888"
|
|
* @param renderModeDescriptor
|
|
* @param actualBitmap actual bitmap generated by this run
|
|
* @param referenceBitmap bitmap we expected to be generated
|
|
*/
|
|
ErrorCombination compare_test_results_to_reference_bitmap(
|
|
const char *shortName, const char *configName, const char *renderModeDescriptor,
|
|
SkBitmap& actualBitmap, const SkBitmap* referenceBitmap) {
|
|
|
|
SkASSERT(referenceBitmap);
|
|
Expectations expectations(*referenceBitmap);
|
|
BitmapAndDigest actualBitmapAndDigest(actualBitmap);
|
|
|
|
// TODO: Eliminate RecordTestResults from here.
|
|
// Results recording code for the test_drawing path has been refactored so that
|
|
// RecordTestResults is only called once, at the topmost level. However, the
|
|
// other paths have not yet been refactored, and RecordTestResults has been added
|
|
// here to maintain proper behavior for calls not coming from the test_drawing path.
|
|
ErrorCombination errors;
|
|
errors.add(compare_to_expectations(expectations, actualBitmapAndDigest, shortName,
|
|
configName, renderModeDescriptor, false));
|
|
SkString shortNamePlusConfig = make_shortname_plus_config(shortName, configName);
|
|
RecordTestResults(errors, shortNamePlusConfig, renderModeDescriptor);
|
|
|
|
return errors;
|
|
}
|
|
|
|
static SkPicture* generate_new_picture(GM* gm, BbhType bbhType, uint32_t recordFlags,
|
|
SkScalar scale = SK_Scalar1) {
|
|
// Pictures are refcounted so must be on heap
|
|
SkPicture* pict;
|
|
int width = SkScalarCeilToInt(SkScalarMul(SkIntToScalar(gm->getISize().width()), scale));
|
|
int height = SkScalarCeilToInt(SkScalarMul(SkIntToScalar(gm->getISize().height()), scale));
|
|
|
|
if (kTileGrid_BbhType == bbhType) {
|
|
SkTileGridPicture::TileGridInfo info;
|
|
info.fMargin.setEmpty();
|
|
info.fOffset.setZero();
|
|
info.fTileInterval.set(16, 16);
|
|
pict = new SkTileGridPicture(width, height, info);
|
|
} else {
|
|
pict = new SkPicture;
|
|
}
|
|
if (kNone_BbhType != bbhType) {
|
|
recordFlags |= SkPicture::kOptimizeForClippedPlayback_RecordingFlag;
|
|
}
|
|
SkCanvas* cv = pict->beginRecording(width, height, recordFlags);
|
|
cv->scale(scale, scale);
|
|
invokeGM(gm, cv, false, false);
|
|
pict->endRecording();
|
|
|
|
return pict;
|
|
}
|
|
|
|
static SkData* bitmap_encoder(size_t* pixelRefOffset, const SkBitmap& bm) {
|
|
SkPixelRef* pr = bm.pixelRef();
|
|
if (pr != NULL) {
|
|
SkData* data = pr->refEncodedData();
|
|
if (data != NULL) {
|
|
*pixelRefOffset = bm.pixelRefOffset();
|
|
return data;
|
|
}
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
static SkPicture* stream_to_new_picture(const SkPicture& src) {
|
|
SkDynamicMemoryWStream storage;
|
|
src.serialize(&storage, &bitmap_encoder);
|
|
SkAutoTUnref<SkStreamAsset> pictReadback(storage.detachAsStream());
|
|
SkPicture* retval = SkPicture::CreateFromStream(pictReadback,
|
|
&SkImageDecoder::DecodeMemory);
|
|
return retval;
|
|
}
|
|
|
|
// Test: draw into a bitmap or pdf.
|
|
// Depending on flags, possibly compare to an expected image.
|
|
// If writePath is not NULL, also write images (or documents) to the specified path.
|
|
ErrorCombination test_drawing(GM* gm, const ConfigData& gRec,
|
|
const SkTDArray<const PDFRasterizerData*> &pdfRasterizers,
|
|
const char writePath [],
|
|
GrSurface* gpuTarget,
|
|
SkBitmap* bitmap) {
|
|
ErrorCombination errors;
|
|
SkDynamicMemoryWStream document;
|
|
SkString path;
|
|
|
|
if (gRec.fBackend == kRaster_Backend ||
|
|
gRec.fBackend == kGPU_Backend) {
|
|
// Early exit if we can't generate the image.
|
|
errors.add(generate_image(gm, gRec, gpuTarget, bitmap, false));
|
|
if (!errors.isEmpty()) {
|
|
// TODO: Add a test to exercise what the stdout and
|
|
// JSON look like if we get an "early error" while
|
|
// trying to generate the image.
|
|
return errors;
|
|
}
|
|
BitmapAndDigest bitmapAndDigest(*bitmap);
|
|
errors.add(compare_test_results_to_stored_expectations(
|
|
gm, gRec, gRec.fName, &bitmapAndDigest));
|
|
|
|
if (writePath && (gRec.fFlags & kWrite_ConfigFlag)) {
|
|
path = make_bitmap_filename(writePath, gm->shortName(), gRec.fName,
|
|
"", bitmapAndDigest.fDigest);
|
|
errors.add(write_bitmap(path, bitmapAndDigest.fBitmap));
|
|
}
|
|
} else if (gRec.fBackend == kPDF_Backend) {
|
|
if (!generate_pdf(gm, document)) {
|
|
errors.add(kGeneratePdfFailed_ErrorType);
|
|
} else {
|
|
SkAutoTUnref<SkStreamAsset> documentStream(document.detachAsStream());
|
|
if (writePath && (gRec.fFlags & kWrite_ConfigFlag)) {
|
|
path = make_filename(writePath, gm->shortName(), gRec.fName, "", "pdf");
|
|
errors.add(write_document(path, documentStream));
|
|
}
|
|
|
|
if (!(gm->getFlags() & GM::kSkipPDFRasterization_Flag)) {
|
|
for (int i = 0; i < pdfRasterizers.count(); i++) {
|
|
SkBitmap pdfBitmap;
|
|
SkASSERT(documentStream->rewind());
|
|
bool success = (*pdfRasterizers[i]->fRasterizerFunction)(
|
|
documentStream.get(), &pdfBitmap);
|
|
if (!success) {
|
|
gm_fprintf(stderr, "FAILED to render PDF for %s using renderer %s\n",
|
|
gm->shortName(),
|
|
pdfRasterizers[i]->fName);
|
|
continue;
|
|
}
|
|
|
|
SkString configName(gRec.fName);
|
|
configName.append("-");
|
|
configName.append(pdfRasterizers[i]->fName);
|
|
|
|
BitmapAndDigest bitmapAndDigest(pdfBitmap);
|
|
errors.add(compare_test_results_to_stored_expectations(
|
|
gm, gRec, configName.c_str(), &bitmapAndDigest));
|
|
|
|
if (writePath && (gRec.fFlags & kWrite_ConfigFlag)) {
|
|
path = make_bitmap_filename(writePath, gm->shortName(),
|
|
configName.c_str(),
|
|
"", bitmapAndDigest.fDigest);
|
|
errors.add(write_bitmap(path, bitmapAndDigest.fBitmap));
|
|
}
|
|
}
|
|
} else {
|
|
errors.add(kIntentionallySkipped_ErrorType);
|
|
}
|
|
}
|
|
} else if (gRec.fBackend == kXPS_Backend) {
|
|
generate_xps(gm, document);
|
|
SkAutoTUnref<SkStreamAsset> documentStream(document.detachAsStream());
|
|
|
|
errors.add(compare_test_results_to_stored_expectations(
|
|
gm, gRec, gRec.fName, NULL));
|
|
|
|
if (writePath && (gRec.fFlags & kWrite_ConfigFlag)) {
|
|
path = make_filename(writePath, gm->shortName(), gRec.fName, "", "xps");
|
|
errors.add(write_document(path, documentStream));
|
|
}
|
|
} else {
|
|
SkASSERT(false);
|
|
}
|
|
return errors;
|
|
}
|
|
|
|
ErrorCombination test_deferred_drawing(GM* gm,
|
|
const ConfigData& gRec,
|
|
const SkBitmap& referenceBitmap,
|
|
GrSurface* gpuTarget) {
|
|
if (gRec.fBackend == kRaster_Backend ||
|
|
gRec.fBackend == kGPU_Backend) {
|
|
const char renderModeDescriptor[] = "-deferred";
|
|
SkBitmap bitmap;
|
|
// Early exit if we can't generate the image, but this is
|
|
// expected in some cases, so don't report a test failure.
|
|
ErrorCombination errors = generate_image(gm, gRec, gpuTarget, &bitmap, true);
|
|
// TODO(epoger): This logic is the opposite of what is
|
|
// described above... if we succeeded in generating the
|
|
// -deferred image, we exit early! We should fix this
|
|
// ASAP, because it is hiding -deferred errors... but for
|
|
// now, I'm leaving the logic as it is so that the
|
|
// refactoring change
|
|
// https://codereview.chromium.org/12992003/ is unblocked.
|
|
//
|
|
// Filed as https://code.google.com/p/skia/issues/detail?id=1180
|
|
// ('image-surface gm test is failing in "deferred" mode,
|
|
// and gm is not reporting the failure')
|
|
if (errors.isEmpty()) {
|
|
// TODO(epoger): Report this as a new ErrorType,
|
|
// something like kImageGeneration_ErrorType?
|
|
return kEmpty_ErrorCombination;
|
|
}
|
|
return compare_test_results_to_reference_bitmap(
|
|
gm->shortName(), gRec.fName, renderModeDescriptor, bitmap, &referenceBitmap);
|
|
}
|
|
return kEmpty_ErrorCombination;
|
|
}
|
|
|
|
ErrorCombination test_pipe_playback(GM* gm, const ConfigData& gRec,
|
|
const SkBitmap& referenceBitmap, bool simulateFailure) {
|
|
const SkString shortNamePlusConfig = make_shortname_plus_config(gm->shortName(),
|
|
gRec.fName);
|
|
ErrorCombination errors;
|
|
for (size_t i = 0; i < SK_ARRAY_COUNT(gPipeWritingFlagCombos); ++i) {
|
|
SkString renderModeDescriptor("-pipe");
|
|
renderModeDescriptor.append(gPipeWritingFlagCombos[i].name);
|
|
|
|
if (gm->getFlags() & GM::kSkipPipe_Flag
|
|
|| (gPipeWritingFlagCombos[i].flags == SkGPipeWriter::kCrossProcess_Flag
|
|
&& gm->getFlags() & GM::kSkipPipeCrossProcess_Flag)) {
|
|
RecordTestResults(kIntentionallySkipped_ErrorType, shortNamePlusConfig,
|
|
renderModeDescriptor.c_str());
|
|
errors.add(kIntentionallySkipped_ErrorType);
|
|
} else {
|
|
SkBitmap bitmap;
|
|
SkISize size = gm->getISize();
|
|
setup_bitmap(gRec, size, &bitmap);
|
|
SkCanvas canvas(bitmap);
|
|
installFilter(&canvas);
|
|
// Pass a decoding function so the factory GM (which has an SkBitmap
|
|
// with encoded data) will not fail playback.
|
|
PipeController pipeController(&canvas, &SkImageDecoder::DecodeMemory);
|
|
SkGPipeWriter writer;
|
|
SkCanvas* pipeCanvas = writer.startRecording(&pipeController,
|
|
gPipeWritingFlagCombos[i].flags,
|
|
size.width(), size.height());
|
|
if (!simulateFailure) {
|
|
invokeGM(gm, pipeCanvas, false, false);
|
|
}
|
|
complete_bitmap(&bitmap);
|
|
writer.endRecording();
|
|
errors.add(compare_test_results_to_reference_bitmap(
|
|
gm->shortName(), gRec.fName, renderModeDescriptor.c_str(), bitmap,
|
|
&referenceBitmap));
|
|
if (!errors.isEmpty()) {
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
return errors;
|
|
}
|
|
|
|
ErrorCombination test_tiled_pipe_playback(GM* gm, const ConfigData& gRec,
|
|
const SkBitmap& referenceBitmap) {
|
|
const SkString shortNamePlusConfig = make_shortname_plus_config(gm->shortName(),
|
|
gRec.fName);
|
|
ErrorCombination errors;
|
|
for (size_t i = 0; i < SK_ARRAY_COUNT(gPipeWritingFlagCombos); ++i) {
|
|
SkString renderModeDescriptor("-tiled pipe");
|
|
renderModeDescriptor.append(gPipeWritingFlagCombos[i].name);
|
|
|
|
if ((gm->getFlags() & GM::kSkipPipe_Flag) ||
|
|
(gm->getFlags() & GM::kSkipTiled_Flag)) {
|
|
RecordTestResults(kIntentionallySkipped_ErrorType, shortNamePlusConfig,
|
|
renderModeDescriptor.c_str());
|
|
errors.add(kIntentionallySkipped_ErrorType);
|
|
} else {
|
|
SkBitmap bitmap;
|
|
SkISize size = gm->getISize();
|
|
setup_bitmap(gRec, size, &bitmap);
|
|
SkCanvas canvas(bitmap);
|
|
installFilter(&canvas);
|
|
TiledPipeController pipeController(bitmap, &SkImageDecoder::DecodeMemory);
|
|
SkGPipeWriter writer;
|
|
SkCanvas* pipeCanvas = writer.startRecording(&pipeController,
|
|
gPipeWritingFlagCombos[i].flags,
|
|
size.width(), size.height());
|
|
invokeGM(gm, pipeCanvas, false, false);
|
|
complete_bitmap(&bitmap);
|
|
writer.endRecording();
|
|
errors.add(compare_test_results_to_reference_bitmap(gm->shortName(), gRec.fName,
|
|
renderModeDescriptor.c_str(),
|
|
bitmap, &referenceBitmap));
|
|
if (!errors.isEmpty()) {
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
return errors;
|
|
}
|
|
|
|
//
|
|
// member variables.
|
|
// They are public for now, to allow easier setting by tool_main().
|
|
//
|
|
|
|
bool fUseFileHierarchy, fWriteChecksumBasedFilenames;
|
|
ErrorCombination fIgnorableErrorTypes;
|
|
SkTArray<SkString> fIgnorableTestSubstrings;
|
|
|
|
const char* fMismatchPath;
|
|
const char* fMissingExpectationsPath;
|
|
|
|
// collection of tests that have failed with each ErrorType
|
|
SkTArray<SkString> fFailedTests[kLast_ErrorType+1];
|
|
int fTestsRun;
|
|
SkTDict<int> fRenderModesEncountered;
|
|
|
|
// Where to read expectations (expected image hash digests, etc.) from.
|
|
// If unset, we don't do comparisons.
|
|
SkAutoTUnref<ExpectationsSource> fExpectationsSource;
|
|
|
|
// JSON summaries that we generate as we go (just for output).
|
|
Json::Value fJsonExpectedResults;
|
|
Json::Value fJsonActualResults_Failed;
|
|
Json::Value fJsonActualResults_FailureIgnored;
|
|
Json::Value fJsonActualResults_NoComparison;
|
|
Json::Value fJsonActualResults_Succeeded;
|
|
|
|
}; // end of GMMain class definition
|
|
|
|
#if SK_SUPPORT_GPU
|
|
static const GLContextType kDontCare_GLContextType = GrContextFactory::kNative_GLContextType;
|
|
#else
|
|
static const GLContextType kDontCare_GLContextType = 0;
|
|
#endif
|
|
|
|
static const ConfigData gRec[] = {
|
|
{ SkBitmap::kARGB_8888_Config, kRaster_Backend, kDontCare_GLContextType, 0, kRW_ConfigFlag, "8888", true },
|
|
#if 0 // stop testing this (for now at least) since we want to remove support for it (soon please!!!)
|
|
{ SkBitmap::kARGB_4444_Config, kRaster_Backend, kDontCare_GLContextType, 0, kRW_ConfigFlag, "4444", true },
|
|
#endif
|
|
{ SkBitmap::kRGB_565_Config, kRaster_Backend, kDontCare_GLContextType, 0, kRW_ConfigFlag, "565", true },
|
|
#if SK_SUPPORT_GPU
|
|
{ SkBitmap::kARGB_8888_Config, kGPU_Backend, GrContextFactory::kNative_GLContextType, 0, kRW_ConfigFlag, "gpu", true },
|
|
{ SkBitmap::kARGB_8888_Config, kGPU_Backend, GrContextFactory::kNative_GLContextType, 16, kRW_ConfigFlag, "msaa16", false},
|
|
{ SkBitmap::kARGB_8888_Config, kGPU_Backend, GrContextFactory::kNative_GLContextType, 4, kRW_ConfigFlag, "msaa4", false},
|
|
/* The gpudebug context does not generate meaningful images, so don't record
|
|
* the images it generates! We only run it to look for asserts. */
|
|
{ SkBitmap::kARGB_8888_Config, kGPU_Backend, GrContextFactory::kDebug_GLContextType, 0, kNone_ConfigFlag, "gpudebug", kDebugOnly},
|
|
/* The gpunull context does the least amount of work possible and doesn't
|
|
generate meaninful images, so don't record them!. It can be run to
|
|
isolate the CPU-side processing expense from the GPU-side.
|
|
*/
|
|
{ SkBitmap::kARGB_8888_Config, kGPU_Backend, GrContextFactory::kNull_GLContextType, 0, kNone_ConfigFlag, "gpunull", kDebugOnly},
|
|
#if SK_ANGLE
|
|
{ SkBitmap::kARGB_8888_Config, kGPU_Backend, GrContextFactory::kANGLE_GLContextType, 0, kRW_ConfigFlag, "angle", true },
|
|
{ SkBitmap::kARGB_8888_Config, kGPU_Backend, GrContextFactory::kANGLE_GLContextType, 16, kRW_ConfigFlag, "anglemsaa16", true },
|
|
#endif // SK_ANGLE
|
|
#ifdef SK_MESA
|
|
{ SkBitmap::kARGB_8888_Config, kGPU_Backend, GrContextFactory::kMESA_GLContextType, 0, kRW_ConfigFlag, "mesa", true },
|
|
#endif // SK_MESA
|
|
#endif // SK_SUPPORT_GPU
|
|
#ifdef SK_SUPPORT_XPS
|
|
/* At present we have no way of comparing XPS files (either natively or by converting to PNG). */
|
|
{ SkBitmap::kARGB_8888_Config, kXPS_Backend, kDontCare_GLContextType, 0, kWrite_ConfigFlag, "xps", true },
|
|
#endif // SK_SUPPORT_XPS
|
|
#ifdef SK_SUPPORT_PDF
|
|
{ SkBitmap::kARGB_8888_Config, kPDF_Backend, kDontCare_GLContextType, 0, kRW_ConfigFlag, "pdf", true },
|
|
#endif // SK_SUPPORT_PDF
|
|
};
|
|
|
|
static const PDFRasterizerData kPDFRasterizers[] = {
|
|
#ifdef SK_BUILD_FOR_MAC
|
|
{ &SkPDFDocumentToBitmap, "mac", true },
|
|
#endif
|
|
#ifdef SK_BUILD_POPPLER
|
|
{ &SkPopplerRasterizePDF, "poppler", true },
|
|
#endif
|
|
#ifdef SK_BUILD_NATIVE_PDF_RENDERER
|
|
{ &SkNativeRasterizePDF, "native", true },
|
|
#endif // SK_BUILD_NATIVE_PDF_RENDERER
|
|
};
|
|
|
|
static const char kDefaultsConfigStr[] = "defaults";
|
|
static const char kExcludeConfigChar = '~';
|
|
|
|
static SkString configUsage() {
|
|
SkString result;
|
|
result.appendf("Space delimited list of which configs to run. Possible options: [");
|
|
for (size_t i = 0; i < SK_ARRAY_COUNT(gRec); ++i) {
|
|
SkASSERT(gRec[i].fName != kDefaultsConfigStr);
|
|
if (i > 0) {
|
|
result.append("|");
|
|
}
|
|
result.appendf("%s", gRec[i].fName);
|
|
}
|
|
result.append("]\n");
|
|
result.appendf("The default value is: \"");
|
|
SkString firstDefault;
|
|
SkString allButFirstDefaults;
|
|
SkString nonDefault;
|
|
for (size_t i = 0; i < SK_ARRAY_COUNT(gRec); ++i) {
|
|
if (gRec[i].fRunByDefault) {
|
|
if (i > 0) {
|
|
result.append(" ");
|
|
}
|
|
result.append(gRec[i].fName);
|
|
if (firstDefault.isEmpty()) {
|
|
firstDefault = gRec[i].fName;
|
|
} else {
|
|
if (!allButFirstDefaults.isEmpty()) {
|
|
allButFirstDefaults.append(" ");
|
|
}
|
|
allButFirstDefaults.append(gRec[i].fName);
|
|
}
|
|
} else {
|
|
nonDefault = gRec[i].fName;
|
|
}
|
|
}
|
|
result.append("\"\n");
|
|
result.appendf("\"%s\" evaluates to the default set of configs.\n", kDefaultsConfigStr);
|
|
result.appendf("Prepending \"%c\" on a config name excludes it from the set of configs to run.\n"
|
|
"Exclusions always override inclusions regardless of order.\n",
|
|
kExcludeConfigChar);
|
|
result.appendf("E.g. \"--config %s %c%s %s\" will run these configs:\n\t%s %s",
|
|
kDefaultsConfigStr,
|
|
kExcludeConfigChar,
|
|
firstDefault.c_str(),
|
|
nonDefault.c_str(),
|
|
allButFirstDefaults.c_str(),
|
|
nonDefault.c_str());
|
|
return result;
|
|
}
|
|
|
|
static SkString pdfRasterizerUsage() {
|
|
SkString result;
|
|
result.appendf("Space delimited list of which PDF rasterizers to run. Possible options: [");
|
|
// For this (and further) loops through kPDFRasterizers, there is a typecast to int to avoid
|
|
// the compiler giving an "comparison of unsigned expression < 0 is always false" warning
|
|
// and turning it into a build-breaking error.
|
|
for (int i = 0; i < (int)SK_ARRAY_COUNT(kPDFRasterizers); ++i) {
|
|
if (i > 0) {
|
|
result.append(" ");
|
|
}
|
|
result.append(kPDFRasterizers[i].fName);
|
|
}
|
|
result.append("]\n");
|
|
result.append("The default value is: \"");
|
|
for (int i = 0; i < (int)SK_ARRAY_COUNT(kPDFRasterizers); ++i) {
|
|
if (kPDFRasterizers[i].fRunByDefault) {
|
|
if (i > 0) {
|
|
result.append(" ");
|
|
}
|
|
result.append(kPDFRasterizers[i].fName);
|
|
}
|
|
}
|
|
result.append("\"");
|
|
return result;
|
|
}
|
|
|
|
// Macro magic to convert a numeric preprocessor token into a string.
|
|
// Adapted from http://stackoverflow.com/questions/240353/convert-a-preprocessor-token-to-a-string
|
|
// This should probably be moved into one of our common headers...
|
|
#define TOSTRING_INTERNAL(x) #x
|
|
#define TOSTRING(x) TOSTRING_INTERNAL(x)
|
|
|
|
// Alphabetized ignoring "no" prefix ("readPath", "noreplay", "resourcePath").
|
|
DEFINE_string(config, "", configUsage().c_str());
|
|
DEFINE_string(pdfRasterizers, "default", pdfRasterizerUsage().c_str());
|
|
DEFINE_bool(deferred, false, "Exercise the deferred rendering test pass.");
|
|
DEFINE_string(excludeConfig, "", "Space delimited list of configs to skip.");
|
|
DEFINE_bool(forceBWtext, false, "Disable text anti-aliasing.");
|
|
#if SK_SUPPORT_GPU
|
|
DEFINE_string(gpuCacheSize, "", "<bytes> <count>: Limit the gpu cache to byte size or "
|
|
"object count. " TOSTRING(DEFAULT_CACHE_VALUE) " for either value means "
|
|
"use the default. 0 for either disables the cache.");
|
|
#endif
|
|
DEFINE_bool(hierarchy, false, "Whether to use multilevel directory structure "
|
|
"when reading/writing files.");
|
|
DEFINE_string(ignoreErrorTypes, kDefaultIgnorableErrorTypes.asString(" ").c_str(),
|
|
"Space-separated list of ErrorTypes that should be ignored. If any *other* error "
|
|
"types are encountered, the tool will exit with a nonzero return value.");
|
|
DEFINE_string(ignoreFailuresFile, "", "Path to file containing a list of tests for which we "
|
|
"should ignore failures.\n"
|
|
"The file should list one test per line, except for comment lines starting with #");
|
|
DEFINE_string(match, "", "[~][^]substring[$] [...] of test name to run.\n"
|
|
"Multiple matches may be separated by spaces.\n"
|
|
"~ causes a matching test to always be skipped\n"
|
|
"^ requires the start of the test to match\n"
|
|
"$ requires the end of the test to match\n"
|
|
"^ and $ requires an exact match\n"
|
|
"If a test does not match any list entry,\n"
|
|
"it is skipped unless some list entry starts with ~");
|
|
DEFINE_string(missingExpectationsPath, "", "Write images for tests without expectations "
|
|
"into this directory.");
|
|
DEFINE_string(mismatchPath, "", "Write images for tests that failed due to "
|
|
"pixel mismatches into this directory.");
|
|
DEFINE_string(modulo, "", "[--modulo <remainder> <divisor>]: only run tests for which "
|
|
"testIndex %% divisor == remainder.");
|
|
DEFINE_bool(pipe, false, "Exercise the SkGPipe replay test pass.");
|
|
DEFINE_string2(readPath, r, "", "Read reference images from this dir, and report "
|
|
"any differences between those and the newly generated ones.");
|
|
DEFINE_bool(replay, false, "Exercise the SkPicture replay test pass.");
|
|
#if SK_SUPPORT_GPU
|
|
DEFINE_bool(resetGpuContext, false, "Reset the GrContext prior to running each GM.");
|
|
#endif
|
|
DEFINE_string2(resourcePath, i, "", "Directory that stores image resources.");
|
|
DEFINE_bool(rtree, false, "Exercise the R-Tree variant of SkPicture test pass.");
|
|
DEFINE_bool(serialize, false, "Exercise the SkPicture serialization & deserialization test pass.");
|
|
DEFINE_bool(simulatePipePlaybackFailure, false, "Simulate a rendering failure in pipe mode only.");
|
|
DEFINE_bool(tiledPipe, false, "Exercise tiled SkGPipe replay.");
|
|
DEFINE_bool(tileGrid, false, "Exercise the tile grid variant of SkPicture.");
|
|
DEFINE_string(tileGridReplayScales, "", "Space separated list of floating-point scale "
|
|
"factors to be used for tileGrid playback testing. Default value: 1.0");
|
|
DEFINE_bool2(verbose, v, false, "Give more detail (e.g. list all GMs run, more info about "
|
|
"each test).");
|
|
DEFINE_bool(writeChecksumBasedFilenames, false, "When writing out actual images, use checksum-"
|
|
"based filenames, as rebaseline.py will use when downloading them from Google Storage");
|
|
DEFINE_string(writeJsonSummaryPath, "", "Write a JSON-formatted result summary to this file.");
|
|
DEFINE_string2(writePath, w, "", "Write rendered images into this directory.");
|
|
DEFINE_string2(writePicturePath, p, "", "Write .skp files into this directory.");
|
|
DEFINE_int32(pdfJpegQuality, -1, "Encodes images in JPEG at quality level N, "
|
|
"which can be in range 0-100). N = -1 will disable JPEG compression. "
|
|
"Default is N = 100, maximum quality.");
|
|
// TODO(edisonn): pass a matrix instead of forcePerspectiveMatrix
|
|
// Either the 9 numbers defining the matrix
|
|
// or probably more readable would be to replace it with a set of a few predicates
|
|
// Like --prerotate 100 200 10 --posttranslate 10, 10
|
|
// Probably define spacial names like centerx, centery, top, bottom, left, right
|
|
// then we can write something reabable like --rotate centerx centery 90
|
|
DEFINE_bool(forcePerspectiveMatrix, false, "Force a perspective matrix.");
|
|
DEFINE_bool(useDocumentInsteadOfDevice, false, "Use SkDocument::CreateFoo instead of SkFooDevice.");
|
|
DEFINE_int32(pdfRasterDpi, 72, "Scale at which at which the non suported "
|
|
"features in PDF are rasterized. Must be be in range 0-10000. "
|
|
"Default is 72. N = 0 will disable rasterizing features like "
|
|
"text shadows or perspective bitmaps.");
|
|
static SkData* encode_to_dct_data(size_t* pixelRefOffset, const SkBitmap& bitmap) {
|
|
// Filter output of warnings that JPEG is not available for the image.
|
|
if (bitmap.width() >= 65500 || bitmap.height() >= 65500) return NULL;
|
|
if (FLAGS_pdfJpegQuality == -1) return NULL;
|
|
|
|
SkBitmap bm = bitmap;
|
|
#if defined(SK_BUILD_FOR_MAC)
|
|
// Workaround bug #1043 where bitmaps with referenced pixels cause
|
|
// CGImageDestinationFinalize to crash
|
|
SkBitmap copy;
|
|
bitmap.deepCopyTo(©, bitmap.config());
|
|
bm = copy;
|
|
#endif
|
|
|
|
SkPixelRef* pr = bm.pixelRef();
|
|
if (pr != NULL) {
|
|
SkData* data = pr->refEncodedData();
|
|
if (data != NULL) {
|
|
*pixelRefOffset = bm.pixelRefOffset();
|
|
return data;
|
|
}
|
|
}
|
|
|
|
*pixelRefOffset = 0;
|
|
return SkImageEncoder::EncodeData(bm,
|
|
SkImageEncoder::kJPEG_Type,
|
|
FLAGS_pdfJpegQuality);
|
|
}
|
|
|
|
static int findConfig(const char config[]) {
|
|
for (size_t i = 0; i < SK_ARRAY_COUNT(gRec); i++) {
|
|
if (!strcmp(config, gRec[i].fName)) {
|
|
return (int) i;
|
|
}
|
|
}
|
|
return -1;
|
|
}
|
|
|
|
static const PDFRasterizerData* findPDFRasterizer(const char rasterizer[]) {
|
|
for (int i = 0; i < (int)SK_ARRAY_COUNT(kPDFRasterizers); i++) {
|
|
if (!strcmp(rasterizer, kPDFRasterizers[i].fName)) {
|
|
return &kPDFRasterizers[i];
|
|
}
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
template <typename T> void appendUnique(SkTDArray<T>* array, const T& value) {
|
|
int index = array->find(value);
|
|
if (index < 0) {
|
|
*array->append() = value;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Run this test in a number of different configs (8888, 565, PDF,
|
|
* etc.), confirming that the resulting bitmaps match expectations
|
|
* (which may be different for each config).
|
|
*
|
|
* Returns all errors encountered while doing so.
|
|
*/
|
|
ErrorCombination run_multiple_configs(GMMain &gmmain, GM *gm,
|
|
const SkTDArray<size_t> &configs,
|
|
const SkTDArray<const PDFRasterizerData*> &pdfRasterizers,
|
|
GrContextFactory *grFactory);
|
|
ErrorCombination run_multiple_configs(GMMain &gmmain, GM *gm,
|
|
const SkTDArray<size_t> &configs,
|
|
const SkTDArray<const PDFRasterizerData*> &pdfRasterizers,
|
|
GrContextFactory *grFactory) {
|
|
const char renderModeDescriptor[] = "";
|
|
ErrorCombination errorsForAllConfigs;
|
|
uint32_t gmFlags = gm->getFlags();
|
|
|
|
for (int i = 0; i < configs.count(); i++) {
|
|
ConfigData config = gRec[configs[i]];
|
|
const SkString shortNamePlusConfig = gmmain.make_shortname_plus_config(gm->shortName(),
|
|
config.fName);
|
|
|
|
// Skip any tests that we don't even need to try.
|
|
// If any of these were skipped on a per-GM basis, record them as
|
|
// kIntentionallySkipped.
|
|
if (kPDF_Backend == config.fBackend) {
|
|
if (gmFlags & GM::kSkipPDF_Flag) {
|
|
gmmain.RecordTestResults(kIntentionallySkipped_ErrorType, shortNamePlusConfig,
|
|
renderModeDescriptor);
|
|
errorsForAllConfigs.add(kIntentionallySkipped_ErrorType);
|
|
continue;
|
|
}
|
|
}
|
|
if ((gmFlags & GM::kSkip565_Flag) &&
|
|
(kRaster_Backend == config.fBackend) &&
|
|
(SkBitmap::kRGB_565_Config == config.fConfig)) {
|
|
gmmain.RecordTestResults(kIntentionallySkipped_ErrorType, shortNamePlusConfig,
|
|
renderModeDescriptor);
|
|
errorsForAllConfigs.add(kIntentionallySkipped_ErrorType);
|
|
continue;
|
|
}
|
|
if (((gmFlags & GM::kSkipGPU_Flag) && kGPU_Backend == config.fBackend) ||
|
|
((gmFlags & GM::kGPUOnly_Flag) && kGPU_Backend != config.fBackend)) {
|
|
gmmain.RecordTestResults(kIntentionallySkipped_ErrorType, shortNamePlusConfig,
|
|
renderModeDescriptor);
|
|
errorsForAllConfigs.add(kIntentionallySkipped_ErrorType);
|
|
continue;
|
|
}
|
|
|
|
// Now we know that we want to run this test and record its
|
|
// success or failure.
|
|
ErrorCombination errorsForThisConfig;
|
|
GrSurface* gpuTarget = NULL;
|
|
#if SK_SUPPORT_GPU
|
|
SkAutoTUnref<GrSurface> auGpuTarget;
|
|
if ((errorsForThisConfig.isEmpty()) && (kGPU_Backend == config.fBackend)) {
|
|
if (FLAGS_resetGpuContext) {
|
|
grFactory->destroyContexts();
|
|
}
|
|
GrContext* gr = grFactory->get(config.fGLContextType);
|
|
bool grSuccess = false;
|
|
if (gr) {
|
|
// create a render target to back the device
|
|
GrTextureDesc desc;
|
|
desc.fConfig = kSkia8888_GrPixelConfig;
|
|
desc.fFlags = kRenderTarget_GrTextureFlagBit;
|
|
desc.fWidth = gm->getISize().width();
|
|
desc.fHeight = gm->getISize().height();
|
|
desc.fSampleCnt = config.fSampleCnt;
|
|
auGpuTarget.reset(gr->createUncachedTexture(desc, NULL, 0));
|
|
if (NULL != auGpuTarget) {
|
|
gpuTarget = auGpuTarget;
|
|
grSuccess = true;
|
|
// Set the user specified cache limits if non-default.
|
|
size_t bytes;
|
|
int count;
|
|
gr->getTextureCacheLimits(&count, &bytes);
|
|
if (DEFAULT_CACHE_VALUE != gGpuCacheSizeBytes) {
|
|
bytes = static_cast<size_t>(gGpuCacheSizeBytes);
|
|
}
|
|
if (DEFAULT_CACHE_VALUE != gGpuCacheSizeCount) {
|
|
count = gGpuCacheSizeCount;
|
|
}
|
|
gr->setTextureCacheLimits(count, bytes);
|
|
}
|
|
}
|
|
if (!grSuccess) {
|
|
errorsForThisConfig.add(kNoGpuContext_ErrorType);
|
|
}
|
|
}
|
|
#endif
|
|
|
|
SkBitmap comparisonBitmap;
|
|
|
|
const char* writePath;
|
|
if (FLAGS_writePath.count() == 1) {
|
|
writePath = FLAGS_writePath[0];
|
|
} else {
|
|
writePath = NULL;
|
|
}
|
|
|
|
if (errorsForThisConfig.isEmpty()) {
|
|
errorsForThisConfig.add(gmmain.test_drawing(gm, config, pdfRasterizers,
|
|
writePath, gpuTarget,
|
|
&comparisonBitmap));
|
|
gmmain.RecordTestResults(errorsForThisConfig, shortNamePlusConfig, "");
|
|
}
|
|
|
|
if (FLAGS_deferred && errorsForThisConfig.isEmpty() &&
|
|
(kGPU_Backend == config.fBackend || kRaster_Backend == config.fBackend)) {
|
|
errorsForThisConfig.add(gmmain.test_deferred_drawing(gm, config, comparisonBitmap,
|
|
gpuTarget));
|
|
}
|
|
|
|
errorsForAllConfigs.add(errorsForThisConfig);
|
|
}
|
|
return errorsForAllConfigs;
|
|
}
|
|
|
|
/**
|
|
* Run this test in a number of different drawing modes (pipe,
|
|
* deferred, tiled, etc.), confirming that the resulting bitmaps all
|
|
* *exactly* match comparisonBitmap.
|
|
*
|
|
* Returns all errors encountered while doing so.
|
|
*/
|
|
ErrorCombination run_multiple_modes(GMMain &gmmain, GM *gm, const ConfigData &compareConfig,
|
|
const SkBitmap &comparisonBitmap,
|
|
const SkTDArray<SkScalar> &tileGridReplayScales);
|
|
ErrorCombination run_multiple_modes(GMMain &gmmain, GM *gm, const ConfigData &compareConfig,
|
|
const SkBitmap &comparisonBitmap,
|
|
const SkTDArray<SkScalar> &tileGridReplayScales) {
|
|
ErrorCombination errorsForAllModes;
|
|
uint32_t gmFlags = gm->getFlags();
|
|
const SkString shortNamePlusConfig = gmmain.make_shortname_plus_config(gm->shortName(),
|
|
compareConfig.fName);
|
|
|
|
SkPicture* pict = gmmain.generate_new_picture(gm, kNone_BbhType, 0);
|
|
SkAutoUnref aur(pict);
|
|
if (FLAGS_replay) {
|
|
const char renderModeDescriptor[] = "-replay";
|
|
if (gmFlags & GM::kSkipPicture_Flag) {
|
|
gmmain.RecordTestResults(kIntentionallySkipped_ErrorType, shortNamePlusConfig,
|
|
renderModeDescriptor);
|
|
errorsForAllModes.add(kIntentionallySkipped_ErrorType);
|
|
} else {
|
|
SkBitmap bitmap;
|
|
gmmain.generate_image_from_picture(gm, compareConfig, pict, &bitmap);
|
|
errorsForAllModes.add(gmmain.compare_test_results_to_reference_bitmap(
|
|
gm->shortName(), compareConfig.fName, renderModeDescriptor, bitmap,
|
|
&comparisonBitmap));
|
|
}
|
|
}
|
|
|
|
if (FLAGS_serialize) {
|
|
const char renderModeDescriptor[] = "-serialize";
|
|
if (gmFlags & GM::kSkipPicture_Flag) {
|
|
gmmain.RecordTestResults(kIntentionallySkipped_ErrorType, shortNamePlusConfig,
|
|
renderModeDescriptor);
|
|
errorsForAllModes.add(kIntentionallySkipped_ErrorType);
|
|
} else {
|
|
SkPicture* repict = gmmain.stream_to_new_picture(*pict);
|
|
SkAutoUnref aurr(repict);
|
|
SkBitmap bitmap;
|
|
gmmain.generate_image_from_picture(gm, compareConfig, repict, &bitmap);
|
|
errorsForAllModes.add(gmmain.compare_test_results_to_reference_bitmap(
|
|
gm->shortName(), compareConfig.fName, renderModeDescriptor, bitmap,
|
|
&comparisonBitmap));
|
|
}
|
|
}
|
|
|
|
if ((1 == FLAGS_writePicturePath.count()) &&
|
|
!(gmFlags & GM::kSkipPicture_Flag)) {
|
|
const char* pictureSuffix = "skp";
|
|
// TODO(epoger): Make sure this still works even though the
|
|
// filename now contains the config name (it used to contain
|
|
// just the shortName). I think this is actually an
|
|
// *improvement*, because now runs with different configs will
|
|
// write out their SkPictures to separate files rather than
|
|
// overwriting each other. But we should make sure it doesn't
|
|
// break anybody.
|
|
SkString path = gmmain.make_filename(FLAGS_writePicturePath[0], gm->shortName(),
|
|
compareConfig.fName, "", pictureSuffix);
|
|
SkFILEWStream stream(path.c_str());
|
|
pict->serialize(&stream);
|
|
}
|
|
|
|
if (FLAGS_rtree) {
|
|
const char renderModeDescriptor[] = "-rtree";
|
|
if (gmFlags & GM::kSkipPicture_Flag) {
|
|
gmmain.RecordTestResults(kIntentionallySkipped_ErrorType, shortNamePlusConfig,
|
|
renderModeDescriptor);
|
|
errorsForAllModes.add(kIntentionallySkipped_ErrorType);
|
|
} else {
|
|
SkPicture* pict = gmmain.generate_new_picture(
|
|
gm, kRTree_BbhType, SkPicture::kOptimizeForClippedPlayback_RecordingFlag);
|
|
SkAutoUnref aur(pict);
|
|
SkBitmap bitmap;
|
|
gmmain.generate_image_from_picture(gm, compareConfig, pict, &bitmap);
|
|
errorsForAllModes.add(gmmain.compare_test_results_to_reference_bitmap(
|
|
gm->shortName(), compareConfig.fName, renderModeDescriptor, bitmap,
|
|
&comparisonBitmap));
|
|
}
|
|
}
|
|
|
|
if (FLAGS_tileGrid) {
|
|
for(int scaleIndex = 0; scaleIndex < tileGridReplayScales.count(); ++scaleIndex) {
|
|
SkScalar replayScale = tileGridReplayScales[scaleIndex];
|
|
SkString renderModeDescriptor("-tilegrid");
|
|
if (SK_Scalar1 != replayScale) {
|
|
renderModeDescriptor += "-scale-";
|
|
renderModeDescriptor.appendScalar(replayScale);
|
|
}
|
|
|
|
if ((gmFlags & GM::kSkipPicture_Flag) ||
|
|
((gmFlags & GM::kSkipScaledReplay_Flag) && replayScale != 1)) {
|
|
gmmain.RecordTestResults(kIntentionallySkipped_ErrorType, shortNamePlusConfig,
|
|
renderModeDescriptor.c_str());
|
|
errorsForAllModes.add(kIntentionallySkipped_ErrorType);
|
|
} else {
|
|
// We record with the reciprocal scale to obtain a replay
|
|
// result that can be validated against comparisonBitmap.
|
|
SkScalar recordScale = SkScalarInvert(replayScale);
|
|
SkPicture* pict = gmmain.generate_new_picture(
|
|
gm, kTileGrid_BbhType, SkPicture::kUsePathBoundsForClip_RecordingFlag,
|
|
recordScale);
|
|
SkAutoUnref aur(pict);
|
|
SkBitmap bitmap;
|
|
// We cannot yet pass 'true' to generate_image_from_picture to
|
|
// perform actual tiled rendering (see Issue 1198 -
|
|
// https://code.google.com/p/skia/issues/detail?id=1198)
|
|
gmmain.generate_image_from_picture(gm, compareConfig, pict, &bitmap,
|
|
replayScale /*, true */);
|
|
errorsForAllModes.add(gmmain.compare_test_results_to_reference_bitmap(
|
|
gm->shortName(), compareConfig.fName, renderModeDescriptor.c_str(), bitmap,
|
|
&comparisonBitmap));
|
|
}
|
|
}
|
|
}
|
|
|
|
// run the pipe centric GM steps
|
|
if (FLAGS_pipe) {
|
|
errorsForAllModes.add(gmmain.test_pipe_playback(gm, compareConfig, comparisonBitmap,
|
|
FLAGS_simulatePipePlaybackFailure));
|
|
if (FLAGS_tiledPipe) {
|
|
errorsForAllModes.add(gmmain.test_tiled_pipe_playback(gm, compareConfig,
|
|
comparisonBitmap));
|
|
}
|
|
}
|
|
return errorsForAllModes;
|
|
}
|
|
|
|
/**
|
|
* Read individual lines from a file, pushing them into the given array.
|
|
*
|
|
* @param filename path to the file to read
|
|
* @param lines array of strings to add the lines to
|
|
* @returns true if able to read lines from the file
|
|
*/
|
|
static bool read_lines_from_file(const char* filename, SkTArray<SkString> &lines) {
|
|
SkAutoTUnref<SkStream> streamWrapper(SkStream::NewFromFile(filename));
|
|
SkStream *stream = streamWrapper.get();
|
|
if (!stream) {
|
|
gm_fprintf(stderr, "unable to read file '%s'\n", filename);
|
|
return false;
|
|
}
|
|
|
|
char c;
|
|
SkString line;
|
|
while (1 == stream->read(&c, 1)) {
|
|
// If we hit either CR or LF, we've completed a line.
|
|
//
|
|
// TODO: If the file uses both CR and LF, this will return an extra blank
|
|
// line for each line of the file. Which is OK for current purposes...
|
|
//
|
|
// TODO: Does this properly handle unicode? It doesn't matter for
|
|
// current purposes...
|
|
if ((c == 0x0d) || (c == 0x0a)) {
|
|
lines.push_back(line);
|
|
line.reset();
|
|
} else {
|
|
line.append(&c, 1);
|
|
}
|
|
}
|
|
lines.push_back(line);
|
|
return true;
|
|
}
|
|
|
|
/**
|
|
* Return a list of all entries in an array of strings as a single string
|
|
* of this form:
|
|
* "item1", "item2", "item3"
|
|
*/
|
|
SkString list_all(const SkTArray<SkString> &stringArray);
|
|
SkString list_all(const SkTArray<SkString> &stringArray) {
|
|
SkString total;
|
|
for (int i = 0; i < stringArray.count(); i++) {
|
|
if (i > 0) {
|
|
total.append(", ");
|
|
}
|
|
total.append("\"");
|
|
total.append(stringArray[i]);
|
|
total.append("\"");
|
|
}
|
|
return total;
|
|
}
|
|
|
|
/**
|
|
* Return a list of configuration names, as a single string of this form:
|
|
* "item1", "item2", "item3"
|
|
*
|
|
* @param configs configurations, as a list of indices into gRec
|
|
*/
|
|
SkString list_all_config_names(const SkTDArray<size_t> &configs);
|
|
SkString list_all_config_names(const SkTDArray<size_t> &configs) {
|
|
SkString total;
|
|
for (int i = 0; i < configs.count(); i++) {
|
|
if (i > 0) {
|
|
total.append(", ");
|
|
}
|
|
total.append("\"");
|
|
total.append(gRec[configs[i]].fName);
|
|
total.append("\"");
|
|
}
|
|
return total;
|
|
}
|
|
|
|
static bool prepare_subdirectories(const char *root, bool useFileHierarchy,
|
|
const SkTDArray<size_t> &configs,
|
|
const SkTDArray<const PDFRasterizerData*>& pdfRasterizers) {
|
|
if (!sk_mkdir(root)) {
|
|
return false;
|
|
}
|
|
if (useFileHierarchy) {
|
|
for (int i = 0; i < configs.count(); i++) {
|
|
ConfigData config = gRec[configs[i]];
|
|
SkString subdir;
|
|
subdir.appendf("%s%c%s", root, SkPATH_SEPARATOR, config.fName);
|
|
if (!sk_mkdir(subdir.c_str())) {
|
|
return false;
|
|
}
|
|
|
|
if (config.fBackend == kPDF_Backend) {
|
|
for (int j = 0; j < pdfRasterizers.count(); j++) {
|
|
SkString pdfSubdir = subdir;
|
|
pdfSubdir.appendf("-%s", pdfRasterizers[j]->fName);
|
|
if (!sk_mkdir(pdfSubdir.c_str())) {
|
|
return false;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static bool parse_flags_configs(SkTDArray<size_t>* outConfigs,
|
|
GrContextFactory* grFactory) {
|
|
SkTDArray<size_t> excludeConfigs;
|
|
|
|
for (int i = 0; i < FLAGS_config.count(); i++) {
|
|
const char* config = FLAGS_config[i];
|
|
bool exclude = false;
|
|
if (*config == kExcludeConfigChar) {
|
|
exclude = true;
|
|
config += 1;
|
|
}
|
|
int index = findConfig(config);
|
|
if (index >= 0) {
|
|
if (exclude) {
|
|
*excludeConfigs.append() = index;
|
|
} else {
|
|
appendUnique<size_t>(outConfigs, index);
|
|
}
|
|
} else if (0 == strcmp(kDefaultsConfigStr, config)) {
|
|
if (exclude) {
|
|
gm_fprintf(stderr, "%c%s is not allowed.\n",
|
|
kExcludeConfigChar, kDefaultsConfigStr);
|
|
return false;
|
|
}
|
|
for (size_t c = 0; c < SK_ARRAY_COUNT(gRec); ++c) {
|
|
if (gRec[c].fRunByDefault) {
|
|
appendUnique<size_t>(outConfigs, c);
|
|
}
|
|
}
|
|
} else {
|
|
gm_fprintf(stderr, "unrecognized config %s\n", config);
|
|
return false;
|
|
}
|
|
}
|
|
|
|
for (int i = 0; i < FLAGS_excludeConfig.count(); i++) {
|
|
int index = findConfig(FLAGS_excludeConfig[i]);
|
|
if (index >= 0) {
|
|
*excludeConfigs.append() = index;
|
|
} else {
|
|
gm_fprintf(stderr, "unrecognized excludeConfig %s\n", FLAGS_excludeConfig[i]);
|
|
return false;
|
|
}
|
|
}
|
|
|
|
if (outConfigs->count() == 0) {
|
|
// if no config is specified by user, add the defaults
|
|
for (size_t i = 0; i < SK_ARRAY_COUNT(gRec); ++i) {
|
|
if (gRec[i].fRunByDefault) {
|
|
*outConfigs->append() = i;
|
|
}
|
|
}
|
|
}
|
|
// now remove any explicitly excluded configs
|
|
for (int i = 0; i < excludeConfigs.count(); ++i) {
|
|
int index = outConfigs->find(excludeConfigs[i]);
|
|
if (index >= 0) {
|
|
outConfigs->remove(index);
|
|
// now assert that there was only one copy in configs[]
|
|
SkASSERT(outConfigs->find(excludeConfigs[i]) < 0);
|
|
}
|
|
}
|
|
|
|
#if SK_SUPPORT_GPU
|
|
SkASSERT(grFactory != NULL);
|
|
for (int i = 0; i < outConfigs->count(); ++i) {
|
|
size_t index = (*outConfigs)[i];
|
|
if (kGPU_Backend == gRec[index].fBackend) {
|
|
GrContext* ctx = grFactory->get(gRec[index].fGLContextType);
|
|
if (NULL == ctx) {
|
|
gm_fprintf(stderr, "GrContext could not be created for config %s."
|
|
" Config will be skipped.\n", gRec[index].fName);
|
|
outConfigs->remove(i);
|
|
--i;
|
|
continue;
|
|
}
|
|
if (gRec[index].fSampleCnt > ctx->getMaxSampleCount()) {
|
|
gm_fprintf(stderr, "Sample count (%d) of config %s is not supported."
|
|
" Config will be skipped.\n",
|
|
gRec[index].fSampleCnt, gRec[index].fName);
|
|
outConfigs->remove(i);
|
|
--i;
|
|
}
|
|
}
|
|
}
|
|
#endif
|
|
|
|
if (outConfigs->isEmpty()) {
|
|
gm_fprintf(stderr, "No configs to run.");
|
|
return false;
|
|
}
|
|
|
|
// now show the user the set of configs that will be run.
|
|
SkString configStr("These configs will be run:");
|
|
// show the user the config that will run.
|
|
for (int i = 0; i < outConfigs->count(); ++i) {
|
|
configStr.appendf(" %s", gRec[(*outConfigs)[i]].fName);
|
|
}
|
|
gm_fprintf(stdout, "%s\n", configStr.c_str());
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool parse_flags_pdf_rasterizers(const SkTDArray<size_t>& configs,
|
|
SkTDArray<const PDFRasterizerData*>* outRasterizers) {
|
|
// No need to run this check (and display the PDF rasterizers message)
|
|
// if no PDF backends are in the configs.
|
|
bool configHasPDF = false;
|
|
for (int i = 0; i < configs.count(); i++) {
|
|
if (gRec[configs[i]].fBackend == kPDF_Backend) {
|
|
configHasPDF = true;
|
|
break;
|
|
}
|
|
}
|
|
if (!configHasPDF) {
|
|
return true;
|
|
}
|
|
|
|
if (FLAGS_pdfRasterizers.count() == 1 &&
|
|
!strcmp(FLAGS_pdfRasterizers[0], "default")) {
|
|
for (int i = 0; i < (int)SK_ARRAY_COUNT(kPDFRasterizers); ++i) {
|
|
if (kPDFRasterizers[i].fRunByDefault) {
|
|
*outRasterizers->append() = &kPDFRasterizers[i];
|
|
}
|
|
}
|
|
} else {
|
|
for (int i = 0; i < FLAGS_pdfRasterizers.count(); i++) {
|
|
const char* rasterizer = FLAGS_pdfRasterizers[i];
|
|
const PDFRasterizerData* rasterizerPtr =
|
|
findPDFRasterizer(rasterizer);
|
|
if (rasterizerPtr == NULL) {
|
|
gm_fprintf(stderr, "unrecognized rasterizer %s\n", rasterizer);
|
|
return false;
|
|
}
|
|
appendUnique<const PDFRasterizerData*>(outRasterizers,
|
|
rasterizerPtr);
|
|
}
|
|
}
|
|
|
|
// now show the user the set of configs that will be run.
|
|
SkString configStr("These PDF rasterizers will be run:");
|
|
// show the user the config that will run.
|
|
for (int i = 0; i < outRasterizers->count(); ++i) {
|
|
configStr.appendf(" %s", (*outRasterizers)[i]->fName);
|
|
}
|
|
gm_fprintf(stdout, "%s\n", configStr.c_str());
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool parse_flags_ignore_error_types(ErrorCombination* outErrorTypes) {
|
|
if (FLAGS_ignoreErrorTypes.count() > 0) {
|
|
*outErrorTypes = ErrorCombination();
|
|
for (int i = 0; i < FLAGS_ignoreErrorTypes.count(); i++) {
|
|
ErrorType type;
|
|
const char *name = FLAGS_ignoreErrorTypes[i];
|
|
if (!getErrorTypeByName(name, &type)) {
|
|
gm_fprintf(stderr, "cannot find ErrorType with name '%s'\n", name);
|
|
return false;
|
|
} else {
|
|
outErrorTypes->add(type);
|
|
}
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
/**
|
|
* Replace contents of ignoreTestSubstrings with a list of testname/config substrings, indicating
|
|
* which tests' failures should be ignored.
|
|
*/
|
|
static bool parse_flags_ignore_tests(SkTArray<SkString> &ignoreTestSubstrings) {
|
|
ignoreTestSubstrings.reset();
|
|
|
|
// Parse --ignoreFailuresFile
|
|
for (int i = 0; i < FLAGS_ignoreFailuresFile.count(); i++) {
|
|
SkTArray<SkString> linesFromFile;
|
|
if (!read_lines_from_file(FLAGS_ignoreFailuresFile[i], linesFromFile)) {
|
|
return false;
|
|
} else {
|
|
for (int j = 0; j < linesFromFile.count(); j++) {
|
|
SkString thisLine = linesFromFile[j];
|
|
if (thisLine.isEmpty() || thisLine.startsWith('#')) {
|
|
// skip this line
|
|
} else {
|
|
ignoreTestSubstrings.push_back(thisLine);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static bool parse_flags_modulo(int* moduloRemainder, int* moduloDivisor) {
|
|
if (FLAGS_modulo.count() == 2) {
|
|
*moduloRemainder = atoi(FLAGS_modulo[0]);
|
|
*moduloDivisor = atoi(FLAGS_modulo[1]);
|
|
if (*moduloRemainder < 0 || *moduloDivisor <= 0 ||
|
|
*moduloRemainder >= *moduloDivisor) {
|
|
gm_fprintf(stderr, "invalid modulo values.");
|
|
return false;
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
#if SK_SUPPORT_GPU
|
|
static bool parse_flags_gpu_cache(int* sizeBytes, int* sizeCount) {
|
|
if (FLAGS_gpuCacheSize.count() > 0) {
|
|
if (FLAGS_gpuCacheSize.count() != 2) {
|
|
gm_fprintf(stderr, "--gpuCacheSize requires two arguments\n");
|
|
return false;
|
|
}
|
|
*sizeBytes = atoi(FLAGS_gpuCacheSize[0]);
|
|
*sizeCount = atoi(FLAGS_gpuCacheSize[1]);
|
|
} else {
|
|
*sizeBytes = DEFAULT_CACHE_VALUE;
|
|
*sizeCount = DEFAULT_CACHE_VALUE;
|
|
}
|
|
return true;
|
|
}
|
|
#endif
|
|
|
|
static bool parse_flags_tile_grid_replay_scales(SkTDArray<SkScalar>* outScales) {
|
|
*outScales->append() = SK_Scalar1; // By default only test at scale 1.0
|
|
if (FLAGS_tileGridReplayScales.count() > 0) {
|
|
outScales->reset();
|
|
for (int i = 0; i < FLAGS_tileGridReplayScales.count(); i++) {
|
|
double val = atof(FLAGS_tileGridReplayScales[i]);
|
|
if (0 < val) {
|
|
*outScales->append() = SkDoubleToScalar(val);
|
|
}
|
|
}
|
|
if (0 == outScales->count()) {
|
|
// Should have at least one scale
|
|
gm_fprintf(stderr, "--tileGridReplayScales requires at least one scale.\n");
|
|
return false;
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static bool parse_flags_gmmain_paths(GMMain* gmmain) {
|
|
gmmain->fUseFileHierarchy = FLAGS_hierarchy;
|
|
gmmain->fWriteChecksumBasedFilenames = FLAGS_writeChecksumBasedFilenames;
|
|
|
|
if (FLAGS_mismatchPath.count() == 1) {
|
|
gmmain->fMismatchPath = FLAGS_mismatchPath[0];
|
|
}
|
|
|
|
if (FLAGS_missingExpectationsPath.count() == 1) {
|
|
gmmain->fMissingExpectationsPath = FLAGS_missingExpectationsPath[0];
|
|
}
|
|
|
|
if (FLAGS_readPath.count() == 1) {
|
|
const char* readPath = FLAGS_readPath[0];
|
|
if (!sk_exists(readPath)) {
|
|
gm_fprintf(stderr, "readPath %s does not exist!\n", readPath);
|
|
return false;
|
|
}
|
|
if (sk_isdir(readPath)) {
|
|
if (FLAGS_verbose) {
|
|
gm_fprintf(stdout, "reading from %s\n", readPath);
|
|
}
|
|
gmmain->fExpectationsSource.reset(SkNEW_ARGS(
|
|
IndividualImageExpectationsSource, (readPath)));
|
|
} else {
|
|
if (FLAGS_verbose) {
|
|
gm_fprintf(stdout, "reading expectations from JSON summary file %s\n", readPath);
|
|
}
|
|
gmmain->fExpectationsSource.reset(SkNEW_ARGS(JsonExpectationsSource, (readPath)));
|
|
}
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static bool parse_flags_resource_path() {
|
|
if (FLAGS_resourcePath.count() == 1) {
|
|
GM::SetResourcePath(FLAGS_resourcePath[0]);
|
|
}
|
|
return true;
|
|
}
|
|
|
|
static bool parse_flags_jpeg_quality() {
|
|
if (FLAGS_pdfJpegQuality < -1 || FLAGS_pdfJpegQuality > 100) {
|
|
gm_fprintf(stderr, "%s\n", "pdfJpegQuality must be in [-1 .. 100] range.");
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
int tool_main(int argc, char** argv);
|
|
int tool_main(int argc, char** argv) {
|
|
|
|
#if SK_ENABLE_INST_COUNT
|
|
gPrintInstCount = true;
|
|
#endif
|
|
|
|
SkGraphics::Init();
|
|
// we don't need to see this during a run
|
|
gSkSuppressFontCachePurgeSpew = true;
|
|
|
|
setSystemPreferences();
|
|
GMMain gmmain;
|
|
|
|
SkString usage;
|
|
usage.printf("Run the golden master tests.\n");
|
|
SkCommandLineFlags::SetUsage(usage.c_str());
|
|
SkCommandLineFlags::Parse(argc, argv);
|
|
|
|
SkTDArray<size_t> configs;
|
|
|
|
int moduloRemainder = -1;
|
|
int moduloDivisor = -1;
|
|
SkTDArray<const PDFRasterizerData*> pdfRasterizers;
|
|
SkTDArray<SkScalar> tileGridReplayScales;
|
|
#if SK_SUPPORT_GPU
|
|
GrContextFactory* grFactory = new GrContextFactory;
|
|
#else
|
|
GrContextFactory* grFactory = NULL;
|
|
#endif
|
|
|
|
if (!parse_flags_modulo(&moduloRemainder, &moduloDivisor) ||
|
|
!parse_flags_ignore_error_types(&gmmain.fIgnorableErrorTypes) ||
|
|
!parse_flags_ignore_tests(gmmain.fIgnorableTestSubstrings) ||
|
|
#if SK_SUPPORT_GPU
|
|
!parse_flags_gpu_cache(&gGpuCacheSizeBytes, &gGpuCacheSizeCount) ||
|
|
#endif
|
|
!parse_flags_tile_grid_replay_scales(&tileGridReplayScales) ||
|
|
!parse_flags_resource_path() ||
|
|
!parse_flags_jpeg_quality() ||
|
|
!parse_flags_configs(&configs, grFactory) ||
|
|
!parse_flags_pdf_rasterizers(configs, &pdfRasterizers) ||
|
|
!parse_flags_gmmain_paths(&gmmain)) {
|
|
return -1;
|
|
}
|
|
|
|
if (FLAGS_verbose) {
|
|
if (FLAGS_writePath.count() == 1) {
|
|
gm_fprintf(stdout, "writing to %s\n", FLAGS_writePath[0]);
|
|
}
|
|
if (NULL != gmmain.fMismatchPath) {
|
|
gm_fprintf(stdout, "writing mismatches to %s\n", gmmain.fMismatchPath);
|
|
}
|
|
if (NULL != gmmain.fMissingExpectationsPath) {
|
|
gm_fprintf(stdout, "writing images without expectations to %s\n",
|
|
gmmain.fMissingExpectationsPath);
|
|
}
|
|
if (FLAGS_writePicturePath.count() == 1) {
|
|
gm_fprintf(stdout, "writing pictures to %s\n", FLAGS_writePicturePath[0]);
|
|
}
|
|
if (FLAGS_resourcePath.count() == 1) {
|
|
gm_fprintf(stdout, "reading resources from %s\n", FLAGS_resourcePath[0]);
|
|
}
|
|
}
|
|
|
|
int gmsRun = 0;
|
|
int gmIndex = -1;
|
|
SkString moduloStr;
|
|
|
|
// If we will be writing out files, prepare subdirectories.
|
|
if (FLAGS_writePath.count() == 1) {
|
|
if (!prepare_subdirectories(FLAGS_writePath[0], gmmain.fUseFileHierarchy,
|
|
configs, pdfRasterizers)) {
|
|
return -1;
|
|
}
|
|
}
|
|
if (NULL != gmmain.fMismatchPath) {
|
|
if (!prepare_subdirectories(gmmain.fMismatchPath, gmmain.fUseFileHierarchy,
|
|
configs, pdfRasterizers)) {
|
|
return -1;
|
|
}
|
|
}
|
|
if (NULL != gmmain.fMissingExpectationsPath) {
|
|
if (!prepare_subdirectories(gmmain.fMissingExpectationsPath, gmmain.fUseFileHierarchy,
|
|
configs, pdfRasterizers)) {
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
Iter iter;
|
|
GM* gm;
|
|
while ((gm = iter.next()) != NULL) {
|
|
if (FLAGS_forcePerspectiveMatrix) {
|
|
SkMatrix perspective;
|
|
perspective.setIdentity();
|
|
perspective.setPerspY(SkScalarDiv(SK_Scalar1, SkIntToScalar(1000)));
|
|
perspective.setSkewX(SkScalarDiv(SkIntToScalar(8),
|
|
SkIntToScalar(25)));
|
|
|
|
gm->setStarterMatrix(perspective);
|
|
}
|
|
SkAutoTDelete<GM> adgm(gm);
|
|
++gmIndex;
|
|
if (moduloRemainder >= 0) {
|
|
if ((gmIndex % moduloDivisor) != moduloRemainder) {
|
|
continue;
|
|
}
|
|
moduloStr.printf("[%d.%d] ", gmIndex, moduloDivisor);
|
|
}
|
|
|
|
const char* shortName = gm->shortName();
|
|
|
|
if (SkCommandLineFlags::ShouldSkip(FLAGS_match, shortName)) {
|
|
continue;
|
|
}
|
|
|
|
gmsRun++;
|
|
SkISize size = gm->getISize();
|
|
gm_fprintf(stdout, "%sdrawing... %s [%d %d]\n", moduloStr.c_str(), shortName,
|
|
size.width(), size.height());
|
|
|
|
run_multiple_configs(gmmain, gm, configs, pdfRasterizers, grFactory);
|
|
|
|
SkBitmap comparisonBitmap;
|
|
const ConfigData compareConfig =
|
|
{ SkBitmap::kARGB_8888_Config, kRaster_Backend, kDontCare_GLContextType, 0, kRW_ConfigFlag, "comparison", false };
|
|
gmmain.generate_image(gm, compareConfig, NULL, &comparisonBitmap, false);
|
|
|
|
// TODO(epoger): only run this if gmmain.generate_image() succeeded?
|
|
// Otherwise, what are we comparing against?
|
|
run_multiple_modes(gmmain, gm, compareConfig, comparisonBitmap, tileGridReplayScales);
|
|
}
|
|
|
|
SkTArray<SkString> modes;
|
|
gmmain.GetRenderModesEncountered(modes);
|
|
bool reportError = false;
|
|
if (gmmain.NumSignificantErrors() > 0) {
|
|
reportError = true;
|
|
}
|
|
int expectedNumberOfTests = gmsRun * (configs.count() + modes.count());
|
|
|
|
// Output summary to stdout.
|
|
if (FLAGS_verbose) {
|
|
gm_fprintf(stdout, "Ran %d GMs\n", gmsRun);
|
|
gm_fprintf(stdout, "... over %2d configs [%s]\n", configs.count(),
|
|
list_all_config_names(configs).c_str());
|
|
gm_fprintf(stdout, "... and %2d modes [%s]\n", modes.count(), list_all(modes).c_str());
|
|
gm_fprintf(stdout, "... so there should be a total of %d tests.\n", expectedNumberOfTests);
|
|
}
|
|
gmmain.ListErrors(FLAGS_verbose);
|
|
|
|
// TODO(epoger): Enable this check for Android, too, once we resolve
|
|
// https://code.google.com/p/skia/issues/detail?id=1222
|
|
// ('GM is unexpectedly skipping tests on Android')
|
|
#ifndef SK_BUILD_FOR_ANDROID
|
|
if (expectedNumberOfTests != gmmain.fTestsRun) {
|
|
gm_fprintf(stderr, "expected %d tests, but ran or skipped %d tests\n",
|
|
expectedNumberOfTests, gmmain.fTestsRun);
|
|
reportError = true;
|
|
}
|
|
#endif
|
|
|
|
if (FLAGS_writeJsonSummaryPath.count() == 1) {
|
|
Json::Value root = CreateJsonTree(
|
|
gmmain.fJsonExpectedResults,
|
|
gmmain.fJsonActualResults_Failed, gmmain.fJsonActualResults_FailureIgnored,
|
|
gmmain.fJsonActualResults_NoComparison, gmmain.fJsonActualResults_Succeeded);
|
|
std::string jsonStdString = root.toStyledString();
|
|
SkFILEWStream stream(FLAGS_writeJsonSummaryPath[0]);
|
|
stream.write(jsonStdString.c_str(), jsonStdString.length());
|
|
}
|
|
|
|
#if SK_SUPPORT_GPU
|
|
|
|
#if GR_CACHE_STATS
|
|
for (int i = 0; i < configs.count(); i++) {
|
|
ConfigData config = gRec[configs[i]];
|
|
|
|
if (FLAGS_verbose && (kGPU_Backend == config.fBackend)) {
|
|
GrContext* gr = grFactory->get(config.fGLContextType);
|
|
|
|
gm_fprintf(stdout, "config: %s %x\n", config.fName, gr);
|
|
gr->printCacheStats();
|
|
}
|
|
}
|
|
#endif
|
|
|
|
delete grFactory;
|
|
#endif
|
|
SkGraphics::Term();
|
|
|
|
return (reportError) ? -1 : 0;
|
|
}
|
|
|
|
void GMMain::installFilter(SkCanvas* canvas) {
|
|
if (FLAGS_forceBWtext) {
|
|
canvas->setDrawFilter(SkNEW(BWTextDrawFilter))->unref();
|
|
}
|
|
}
|
|
|
|
#if !defined(SK_BUILD_FOR_IOS) && !defined(SK_BUILD_FOR_NACL)
|
|
int main(int argc, char * const argv[]) {
|
|
return tool_main(argc, (char**) argv);
|
|
}
|
|
#endif
|