6ce447a4ec
From now on, sample counts always refer to the number of actual color samples, and render targets don't have separate color and stencil sample counts. If mixed samples support is available when making a "GrAAType::kCoverage" draw, then an op may attach and use a mixed sampled stencil buffer internally. But this will all be invisible to the client. After this CL, we temporarily won't have a mode to use nvpr with mixed samples. That will soon be fixed by a follow-on CL that enables nvpr with mixed samples in the normal "gl" and "gles" configs. Bug: skia: Change-Id: I1cb8277f0d2d0d371f24bb9f39cd473ed5c5c83b Reviewed-on: https://skia-review.googlesource.com/c/skia/+/221878 Commit-Queue: Chris Dalton <csmartdalton@google.com> Reviewed-by: Brian Salomon <bsalomon@google.com>
905 lines
37 KiB
C++
905 lines
37 KiB
C++
/*
|
|
* Copyright 2017 Google Inc.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license that can be
|
|
* found in the LICENSE file.
|
|
*/
|
|
|
|
#include "include/core/SkTypes.h"
|
|
#include "tests/Test.h"
|
|
|
|
#include "include/core/SkMatrix.h"
|
|
#include "include/core/SkRect.h"
|
|
#include "include/gpu/GrTexture.h"
|
|
#include "include/gpu/mock/GrMockTypes.h"
|
|
#include "include/private/GrRecordingContext.h"
|
|
#include "src/core/SkExchange.h"
|
|
#include "src/core/SkPathPriv.h"
|
|
#include "src/gpu/GrClip.h"
|
|
#include "src/gpu/GrContextPriv.h"
|
|
#include "src/gpu/GrDrawingManager.h"
|
|
#include "src/gpu/GrPaint.h"
|
|
#include "src/gpu/GrPathRenderer.h"
|
|
#include "src/gpu/GrRecordingContextPriv.h"
|
|
#include "src/gpu/GrRenderTargetContext.h"
|
|
#include "src/gpu/GrRenderTargetContextPriv.h"
|
|
#include "src/gpu/ccpr/GrCCPathCache.h"
|
|
#include "src/gpu/ccpr/GrCoverageCountingPathRenderer.h"
|
|
#include "src/gpu/geometry/GrShape.h"
|
|
#include "tools/ToolUtils.h"
|
|
|
|
#include <cmath>
|
|
|
|
static constexpr int kCanvasSize = 100;
|
|
|
|
class CCPRClip : public GrClip {
|
|
public:
|
|
CCPRClip(GrCoverageCountingPathRenderer* ccpr, const SkPath& path) : fCCPR(ccpr), fPath(path) {}
|
|
|
|
private:
|
|
bool apply(GrRecordingContext* context, GrRenderTargetContext* rtc, bool useHWAA,
|
|
bool hasUserStencilSettings, GrAppliedClip* out, SkRect* bounds) const override {
|
|
out->addCoverageFP(fCCPR->makeClipProcessor(rtc->priv().testingOnly_getOpListID(), fPath,
|
|
SkIRect::MakeWH(rtc->width(), rtc->height()),
|
|
rtc->width(), rtc->height(),
|
|
*context->priv().caps()));
|
|
return true;
|
|
}
|
|
bool quickContains(const SkRect&) const final { return false; }
|
|
bool isRRect(const SkRect& rtBounds, SkRRect* rr, GrAA*) const final { return false; }
|
|
void getConservativeBounds(int width, int height, SkIRect* rect, bool* iior) const final {
|
|
rect->set(0, 0, width, height);
|
|
if (iior) {
|
|
*iior = false;
|
|
}
|
|
}
|
|
GrCoverageCountingPathRenderer* const fCCPR;
|
|
const SkPath fPath;
|
|
};
|
|
|
|
class CCPRPathDrawer {
|
|
public:
|
|
CCPRPathDrawer(sk_sp<GrContext> ctx, skiatest::Reporter* reporter, bool doStroke)
|
|
: fCtx(ctx)
|
|
, fCCPR(fCtx->priv().drawingManager()->getCoverageCountingPathRenderer())
|
|
, fRTC(fCtx->priv().makeDeferredRenderTargetContext(
|
|
ctx->priv().caps()->getBackendFormatFromColorType(kRGBA_8888_SkColorType),
|
|
SkBackingFit::kExact, kCanvasSize, kCanvasSize, kRGBA_8888_GrPixelConfig,
|
|
nullptr))
|
|
, fDoStroke(doStroke) {
|
|
if (!fCCPR) {
|
|
ERRORF(reporter, "ccpr not enabled in GrContext for ccpr tests");
|
|
}
|
|
if (!fRTC) {
|
|
ERRORF(reporter, "failed to create GrRenderTargetContext for ccpr tests");
|
|
}
|
|
}
|
|
|
|
GrContext* ctx() const { return fCtx.get(); }
|
|
GrCoverageCountingPathRenderer* ccpr() const { return fCCPR; }
|
|
|
|
bool valid() const { return fCCPR && fRTC; }
|
|
void clear() const { fRTC->clear(nullptr, SK_PMColor4fTRANSPARENT,
|
|
GrRenderTargetContext::CanClearFullscreen::kYes); }
|
|
void destroyGrContext() {
|
|
SkASSERT(fRTC->unique());
|
|
SkASSERT(fCtx->unique());
|
|
fRTC.reset();
|
|
fCCPR = nullptr;
|
|
fCtx.reset();
|
|
}
|
|
|
|
void drawPath(const SkPath& path, const SkMatrix& matrix = SkMatrix::I()) const {
|
|
SkASSERT(this->valid());
|
|
|
|
GrPaint paint;
|
|
paint.setColor4f({ 0, 1, 0, 1 });
|
|
|
|
GrNoClip noClip;
|
|
SkIRect clipBounds = SkIRect::MakeWH(kCanvasSize, kCanvasSize);
|
|
|
|
GrShape shape;
|
|
if (!fDoStroke) {
|
|
shape = GrShape(path);
|
|
} else {
|
|
// Use hairlines for now, since they are the only stroke type that doesn't require a
|
|
// rigid-body transform. The CCPR stroke code makes no distinction between hairlines
|
|
// and regular strokes other than how it decides the device-space stroke width.
|
|
SkStrokeRec stroke(SkStrokeRec::kHairline_InitStyle);
|
|
stroke.setStrokeParams(SkPaint::kRound_Cap, SkPaint::kMiter_Join, 4);
|
|
shape = GrShape(path, GrStyle(stroke, nullptr));
|
|
}
|
|
|
|
fCCPR->testingOnly_drawPathDirectly({
|
|
fCtx.get(), std::move(paint), &GrUserStencilSettings::kUnused, fRTC.get(), &noClip,
|
|
&clipBounds, &matrix, &shape, GrAAType::kCoverage, false});
|
|
}
|
|
|
|
void clipFullscreenRect(SkPath clipPath, SkPMColor4f color = { 0, 1, 0, 1 }) {
|
|
SkASSERT(this->valid());
|
|
|
|
GrPaint paint;
|
|
paint.setColor4f(color);
|
|
|
|
fRTC->drawRect(CCPRClip(fCCPR, clipPath), std::move(paint), GrAA::kYes, SkMatrix::I(),
|
|
SkRect::MakeIWH(kCanvasSize, kCanvasSize));
|
|
}
|
|
|
|
void flush() const {
|
|
SkASSERT(this->valid());
|
|
fCtx->flush();
|
|
}
|
|
|
|
private:
|
|
sk_sp<GrContext> fCtx;
|
|
GrCoverageCountingPathRenderer* fCCPR;
|
|
sk_sp<GrRenderTargetContext> fRTC;
|
|
const bool fDoStroke;
|
|
};
|
|
|
|
class CCPRTest {
|
|
public:
|
|
void run(skiatest::Reporter* reporter, bool doStroke) {
|
|
GrMockOptions mockOptions;
|
|
mockOptions.fInstanceAttribSupport = true;
|
|
mockOptions.fHalfFloatVertexAttributeSupport = true;
|
|
mockOptions.fMapBufferFlags = GrCaps::kCanMap_MapFlag;
|
|
mockOptions.fConfigOptions[kAlpha_half_GrPixelConfig].fRenderability =
|
|
GrMockOptions::ConfigOptions::Renderability::kNonMSAA;
|
|
mockOptions.fConfigOptions[kAlpha_half_GrPixelConfig].fTexturable = true;
|
|
mockOptions.fConfigOptions[kAlpha_8_GrPixelConfig].fRenderability =
|
|
GrMockOptions::ConfigOptions::Renderability::kNonMSAA;
|
|
mockOptions.fConfigOptions[kAlpha_8_GrPixelConfig].fTexturable = true;
|
|
mockOptions.fGeometryShaderSupport = true;
|
|
mockOptions.fIntegerSupport = true;
|
|
mockOptions.fFlatInterpolationSupport = true;
|
|
|
|
GrContextOptions ctxOptions;
|
|
ctxOptions.fDisableCoverageCountingPaths = false;
|
|
ctxOptions.fAllowPathMaskCaching = false;
|
|
ctxOptions.fGpuPathRenderers = GpuPathRenderers::kCoverageCounting;
|
|
|
|
this->customizeOptions(&mockOptions, &ctxOptions);
|
|
|
|
sk_sp<GrContext> mockContext = GrContext::MakeMock(&mockOptions, ctxOptions);
|
|
if (!mockContext) {
|
|
ERRORF(reporter, "could not create mock context");
|
|
return;
|
|
}
|
|
if (!mockContext->unique()) {
|
|
ERRORF(reporter, "mock context is not unique");
|
|
return;
|
|
}
|
|
|
|
CCPRPathDrawer ccpr(skstd::exchange(mockContext, nullptr), reporter, doStroke);
|
|
if (!ccpr.valid()) {
|
|
return;
|
|
}
|
|
|
|
fPath.moveTo(0, 0);
|
|
fPath.cubicTo(50, 50, 0, 50, 50, 0);
|
|
this->onRun(reporter, ccpr);
|
|
}
|
|
|
|
virtual ~CCPRTest() {}
|
|
|
|
protected:
|
|
virtual void customizeOptions(GrMockOptions*, GrContextOptions*) {}
|
|
virtual void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr) = 0;
|
|
|
|
SkPath fPath;
|
|
};
|
|
|
|
#define DEF_CCPR_TEST(name) \
|
|
DEF_GPUTEST(name, reporter, /* options */) { \
|
|
name test; \
|
|
test.run(reporter, false); \
|
|
test.run(reporter, true); \
|
|
}
|
|
|
|
class CCPR_cleanup : public CCPRTest {
|
|
void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr) override {
|
|
REPORTER_ASSERT(reporter, SkPathPriv::TestingOnly_unique(fPath));
|
|
|
|
// Ensure paths get unreffed.
|
|
for (int i = 0; i < 10; ++i) {
|
|
ccpr.drawPath(fPath);
|
|
}
|
|
REPORTER_ASSERT(reporter, !SkPathPriv::TestingOnly_unique(fPath));
|
|
ccpr.flush();
|
|
REPORTER_ASSERT(reporter, SkPathPriv::TestingOnly_unique(fPath));
|
|
|
|
// Ensure clip paths get unreffed.
|
|
for (int i = 0; i < 10; ++i) {
|
|
ccpr.clipFullscreenRect(fPath);
|
|
}
|
|
REPORTER_ASSERT(reporter, !SkPathPriv::TestingOnly_unique(fPath));
|
|
ccpr.flush();
|
|
REPORTER_ASSERT(reporter, SkPathPriv::TestingOnly_unique(fPath));
|
|
|
|
// Ensure paths get unreffed when we delete the context without flushing.
|
|
for (int i = 0; i < 10; ++i) {
|
|
ccpr.drawPath(fPath);
|
|
ccpr.clipFullscreenRect(fPath);
|
|
}
|
|
REPORTER_ASSERT(reporter, !SkPathPriv::TestingOnly_unique(fPath));
|
|
|
|
ccpr.destroyGrContext();
|
|
REPORTER_ASSERT(reporter, SkPathPriv::TestingOnly_unique(fPath));
|
|
}
|
|
};
|
|
DEF_CCPR_TEST(CCPR_cleanup)
|
|
|
|
class CCPR_cleanupWithTexAllocFail : public CCPR_cleanup {
|
|
void customizeOptions(GrMockOptions* mockOptions, GrContextOptions*) override {
|
|
mockOptions->fFailTextureAllocations = true;
|
|
}
|
|
};
|
|
DEF_CCPR_TEST(CCPR_cleanupWithTexAllocFail)
|
|
|
|
class CCPR_unregisterCulledOps : public CCPRTest {
|
|
void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr) override {
|
|
REPORTER_ASSERT(reporter, SkPathPriv::TestingOnly_unique(fPath));
|
|
|
|
// Ensure Ops get unregistered from CCPR when culled early.
|
|
ccpr.drawPath(fPath);
|
|
REPORTER_ASSERT(reporter, !SkPathPriv::TestingOnly_unique(fPath));
|
|
ccpr.clear(); // Clear should delete the CCPR Op.
|
|
REPORTER_ASSERT(reporter, SkPathPriv::TestingOnly_unique(fPath));
|
|
ccpr.flush(); // Should not crash (DrawPathsOp should have unregistered itself).
|
|
|
|
// Ensure Op unregisters work when we delete the context without flushing.
|
|
ccpr.drawPath(fPath);
|
|
REPORTER_ASSERT(reporter, !SkPathPriv::TestingOnly_unique(fPath));
|
|
ccpr.clear(); // Clear should delete the CCPR DrawPathsOp.
|
|
REPORTER_ASSERT(reporter, SkPathPriv::TestingOnly_unique(fPath));
|
|
ccpr.destroyGrContext(); // Should not crash (DrawPathsOp should have unregistered itself).
|
|
}
|
|
};
|
|
DEF_CCPR_TEST(CCPR_unregisterCulledOps)
|
|
|
|
class CCPR_parseEmptyPath : public CCPRTest {
|
|
void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr) override {
|
|
REPORTER_ASSERT(reporter, SkPathPriv::TestingOnly_unique(fPath));
|
|
|
|
// Make a path large enough that ccpr chooses to crop it by the RT bounds, and ends up with
|
|
// an empty path.
|
|
SkPath largeOutsidePath;
|
|
largeOutsidePath.moveTo(-1e30f, -1e30f);
|
|
largeOutsidePath.lineTo(-1e30f, +1e30f);
|
|
largeOutsidePath.lineTo(-1e10f, +1e30f);
|
|
ccpr.drawPath(largeOutsidePath);
|
|
|
|
// Normally an empty path is culled before reaching ccpr, however we use a back door for
|
|
// testing so this path will make it.
|
|
SkPath emptyPath;
|
|
SkASSERT(emptyPath.isEmpty());
|
|
ccpr.drawPath(emptyPath);
|
|
|
|
// This is the test. It will exercise various internal asserts and verify we do not crash.
|
|
ccpr.flush();
|
|
|
|
// Now try again with clips.
|
|
ccpr.clipFullscreenRect(largeOutsidePath);
|
|
ccpr.clipFullscreenRect(emptyPath);
|
|
ccpr.flush();
|
|
|
|
// ... and both.
|
|
ccpr.drawPath(largeOutsidePath);
|
|
ccpr.clipFullscreenRect(largeOutsidePath);
|
|
ccpr.drawPath(emptyPath);
|
|
ccpr.clipFullscreenRect(emptyPath);
|
|
ccpr.flush();
|
|
}
|
|
};
|
|
DEF_CCPR_TEST(CCPR_parseEmptyPath)
|
|
|
|
static int get_mock_texture_id(const GrTexture* texture) {
|
|
const GrBackendTexture& backingTexture = texture->getBackendTexture();
|
|
SkASSERT(GrBackendApi::kMock == backingTexture.backend());
|
|
|
|
if (!backingTexture.isValid()) {
|
|
return 0;
|
|
}
|
|
|
|
GrMockTextureInfo info;
|
|
backingTexture.getMockTextureInfo(&info);
|
|
return info.fID;
|
|
}
|
|
|
|
// Base class for cache path unit tests.
|
|
class CCPRCacheTest : public CCPRTest {
|
|
protected:
|
|
// Registers as an onFlush callback in order to snag the CCPR per-flush resources and note the
|
|
// texture IDs.
|
|
class RecordLastMockAtlasIDs : public GrOnFlushCallbackObject {
|
|
public:
|
|
RecordLastMockAtlasIDs(sk_sp<GrCoverageCountingPathRenderer> ccpr) : fCCPR(ccpr) {}
|
|
|
|
int lastCopyAtlasID() const { return fLastCopyAtlasID; }
|
|
int lastRenderedAtlasID() const { return fLastRenderedAtlasID; }
|
|
|
|
void preFlush(GrOnFlushResourceProvider*, const uint32_t* opListIDs, int numOpListIDs,
|
|
SkTArray<sk_sp<GrRenderTargetContext>>* out) override {
|
|
fLastRenderedAtlasID = fLastCopyAtlasID = 0;
|
|
|
|
const GrCCPerFlushResources* resources = fCCPR->testingOnly_getCurrentFlushResources();
|
|
if (!resources) {
|
|
return;
|
|
}
|
|
|
|
if (const GrTexture* tex = resources->testingOnly_frontCopyAtlasTexture()) {
|
|
fLastCopyAtlasID = get_mock_texture_id(tex);
|
|
}
|
|
if (const GrTexture* tex = resources->testingOnly_frontRenderedAtlasTexture()) {
|
|
fLastRenderedAtlasID = get_mock_texture_id(tex);
|
|
}
|
|
}
|
|
|
|
void postFlush(GrDeferredUploadToken, const uint32_t*, int) override {}
|
|
|
|
private:
|
|
sk_sp<GrCoverageCountingPathRenderer> fCCPR;
|
|
int fLastCopyAtlasID = 0;
|
|
int fLastRenderedAtlasID = 0;
|
|
};
|
|
|
|
CCPRCacheTest() {
|
|
static constexpr int primes[11] = {2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31};
|
|
|
|
SkRandom rand;
|
|
for (size_t i = 0; i < SK_ARRAY_COUNT(fPaths); ++i) {
|
|
int numPts = rand.nextRangeU(GrShape::kMaxKeyFromDataVerbCnt + 1,
|
|
GrShape::kMaxKeyFromDataVerbCnt * 2);
|
|
int step;
|
|
do {
|
|
step = primes[rand.nextU() % SK_ARRAY_COUNT(primes)];
|
|
} while (step == numPts);
|
|
fPaths[i] = ToolUtils::make_star(SkRect::MakeLTRB(0, 0, 1, 1), numPts, step);
|
|
}
|
|
}
|
|
|
|
void drawPathsAndFlush(CCPRPathDrawer& ccpr, const SkMatrix& m) {
|
|
this->drawPathsAndFlush(ccpr, &m, 1);
|
|
}
|
|
void drawPathsAndFlush(CCPRPathDrawer& ccpr, const SkMatrix* matrices, int numMatrices) {
|
|
// Draw all the paths.
|
|
for (size_t i = 0; i < SK_ARRAY_COUNT(fPaths); ++i) {
|
|
ccpr.drawPath(fPaths[i], matrices[i % numMatrices]);
|
|
}
|
|
// Re-draw a few paths, to test the case where a cache entry is hit more than once in a
|
|
// single flush.
|
|
SkRandom rand;
|
|
int duplicateIndices[10];
|
|
for (size_t i = 0; i < SK_ARRAY_COUNT(duplicateIndices); ++i) {
|
|
duplicateIndices[i] = rand.nextULessThan(SK_ARRAY_COUNT(fPaths));
|
|
}
|
|
for (size_t i = 0; i < SK_ARRAY_COUNT(duplicateIndices); ++i) {
|
|
for (size_t j = 0; j <= i; ++j) {
|
|
int idx = duplicateIndices[j];
|
|
ccpr.drawPath(fPaths[idx], matrices[idx % numMatrices]);
|
|
}
|
|
}
|
|
ccpr.flush();
|
|
}
|
|
|
|
private:
|
|
void customizeOptions(GrMockOptions*, GrContextOptions* ctxOptions) override {
|
|
ctxOptions->fAllowPathMaskCaching = true;
|
|
}
|
|
|
|
void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr) final {
|
|
RecordLastMockAtlasIDs atlasIDRecorder(sk_ref_sp(ccpr.ccpr()));
|
|
ccpr.ctx()->priv().addOnFlushCallbackObject(&atlasIDRecorder);
|
|
|
|
this->onRun(reporter, ccpr, atlasIDRecorder);
|
|
|
|
ccpr.ctx()->priv().testingOnly_flushAndRemoveOnFlushCallbackObject(&atlasIDRecorder);
|
|
}
|
|
|
|
virtual void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr,
|
|
const RecordLastMockAtlasIDs&) = 0;
|
|
|
|
protected:
|
|
SkPath fPaths[350];
|
|
};
|
|
|
|
// Ensures ccpr always reuses the same atlas texture in the animation use case.
|
|
class CCPR_cache_animationAtlasReuse : public CCPRCacheTest {
|
|
void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr,
|
|
const RecordLastMockAtlasIDs& atlasIDRecorder) override {
|
|
SkMatrix m = SkMatrix::MakeTrans(kCanvasSize/2, kCanvasSize/2);
|
|
m.preScale(80, 80);
|
|
m.preTranslate(-.5,-.5);
|
|
this->drawPathsAndFlush(ccpr, m);
|
|
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastRenderedAtlasID());
|
|
const int atlasID = atlasIDRecorder.lastRenderedAtlasID();
|
|
|
|
// Ensures we always reuse the same atlas texture in the animation use case.
|
|
for (int i = 0; i < 12; ++i) {
|
|
// 59 is prime, so we will hit every integer modulo 360 before repeating.
|
|
m.preRotate(59, .5, .5);
|
|
|
|
// Go twice. Paths have to get drawn twice with the same matrix before we cache their
|
|
// atlas. This makes sure that on the subsequent draw, after an atlas has been cached
|
|
// and is then invalidated since the matrix will change, that the same underlying
|
|
// texture object is still reused for the next atlas.
|
|
for (int j = 0; j < 2; ++j) {
|
|
this->drawPathsAndFlush(ccpr, m);
|
|
// Nothing should be copied to an 8-bit atlas after just two draws.
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, atlasIDRecorder.lastRenderedAtlasID() == atlasID);
|
|
}
|
|
}
|
|
|
|
// Do the last draw again. (On draw 3 they should get copied to an 8-bit atlas.)
|
|
this->drawPathsAndFlush(ccpr, m);
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastRenderedAtlasID());
|
|
|
|
// Now double-check that everything continues to hit the cache as expected when the matrix
|
|
// doesn't change.
|
|
for (int i = 0; i < 10; ++i) {
|
|
this->drawPathsAndFlush(ccpr, m);
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastRenderedAtlasID());
|
|
}
|
|
}
|
|
};
|
|
DEF_CCPR_TEST(CCPR_cache_animationAtlasReuse)
|
|
|
|
class CCPR_cache_recycleEntries : public CCPRCacheTest {
|
|
void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr,
|
|
const RecordLastMockAtlasIDs& atlasIDRecorder) override {
|
|
SkMatrix m = SkMatrix::MakeTrans(kCanvasSize/2, kCanvasSize/2);
|
|
m.preScale(80, 80);
|
|
m.preTranslate(-.5,-.5);
|
|
|
|
auto cache = ccpr.ccpr()->testingOnly_getPathCache();
|
|
REPORTER_ASSERT(reporter, cache);
|
|
|
|
const auto& lru = cache->testingOnly_getLRU();
|
|
|
|
SkTArray<const void*> expectedPtrs;
|
|
|
|
// Ensures we always reuse the same atlas texture in the animation use case.
|
|
for (int i = 0; i < 5; ++i) {
|
|
// 59 is prime, so we will hit every integer modulo 360 before repeating.
|
|
m.preRotate(59, .5, .5);
|
|
|
|
// Go twice. Paths have to get drawn twice with the same matrix before we cache their
|
|
// atlas.
|
|
for (int j = 0; j < 2; ++j) {
|
|
this->drawPathsAndFlush(ccpr, m);
|
|
// Nothing should be copied to an 8-bit atlas after just two draws.
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastRenderedAtlasID());
|
|
}
|
|
|
|
int idx = 0;
|
|
for (const GrCCPathCacheEntry* entry : lru) {
|
|
if (0 == i) {
|
|
expectedPtrs.push_back(entry);
|
|
} else {
|
|
// The same pointer should have been recycled for the new matrix.
|
|
REPORTER_ASSERT(reporter, entry == expectedPtrs[idx]);
|
|
}
|
|
++idx;
|
|
}
|
|
}
|
|
}
|
|
};
|
|
DEF_CCPR_TEST(CCPR_cache_recycleEntries)
|
|
|
|
// Ensures mostly-visible paths get their full mask cached.
|
|
class CCPR_cache_mostlyVisible : public CCPRCacheTest {
|
|
void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr,
|
|
const RecordLastMockAtlasIDs& atlasIDRecorder) override {
|
|
SkMatrix matrices[3] = {
|
|
SkMatrix::MakeScale(kCanvasSize/2, kCanvasSize/2), // Fully visible.
|
|
SkMatrix::MakeScale(kCanvasSize * 1.25, kCanvasSize * 1.25), // Mostly visible.
|
|
SkMatrix::MakeScale(kCanvasSize * 1.5, kCanvasSize * 1.5), // Mostly NOT visible.
|
|
};
|
|
|
|
for (int i = 0; i < 10; ++i) {
|
|
this->drawPathsAndFlush(ccpr, matrices, 3);
|
|
if (2 == i) {
|
|
// The mostly-visible paths should still get cached.
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastCopyAtlasID());
|
|
} else {
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
}
|
|
// Ensure mostly NOT-visible paths never get cached.
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastRenderedAtlasID());
|
|
}
|
|
|
|
// Clear the path cache.
|
|
this->drawPathsAndFlush(ccpr, SkMatrix::I());
|
|
|
|
// Now only draw the fully/mostly visible ones.
|
|
for (int i = 0; i < 2; ++i) {
|
|
this->drawPathsAndFlush(ccpr, matrices, 2);
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastRenderedAtlasID());
|
|
}
|
|
|
|
// On draw 3 they should get copied to an 8-bit atlas.
|
|
this->drawPathsAndFlush(ccpr, matrices, 2);
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastRenderedAtlasID());
|
|
|
|
for (int i = 0; i < 10; ++i) {
|
|
this->drawPathsAndFlush(ccpr, matrices, 2);
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastRenderedAtlasID());
|
|
}
|
|
|
|
// Draw a different part of the path to ensure the full mask was cached.
|
|
matrices[1].postTranslate(SkScalarFloorToInt(kCanvasSize * -.25f),
|
|
SkScalarFloorToInt(kCanvasSize * -.25f));
|
|
for (int i = 0; i < 10; ++i) {
|
|
this->drawPathsAndFlush(ccpr, matrices, 2);
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastRenderedAtlasID());
|
|
}
|
|
}
|
|
};
|
|
DEF_CCPR_TEST(CCPR_cache_mostlyVisible)
|
|
|
|
// Ensures GrContext::performDeferredCleanup works.
|
|
class CCPR_cache_deferredCleanup : public CCPRCacheTest {
|
|
void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr,
|
|
const RecordLastMockAtlasIDs& atlasIDRecorder) override {
|
|
SkMatrix m = SkMatrix::MakeScale(20, 20);
|
|
int lastRenderedAtlasID = 0;
|
|
|
|
for (int i = 0; i < 5; ++i) {
|
|
this->drawPathsAndFlush(ccpr, m);
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastRenderedAtlasID());
|
|
int renderedAtlasID = atlasIDRecorder.lastRenderedAtlasID();
|
|
REPORTER_ASSERT(reporter, renderedAtlasID != lastRenderedAtlasID);
|
|
lastRenderedAtlasID = renderedAtlasID;
|
|
|
|
this->drawPathsAndFlush(ccpr, m);
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, lastRenderedAtlasID == atlasIDRecorder.lastRenderedAtlasID());
|
|
|
|
// On draw 3 they should get copied to an 8-bit atlas.
|
|
this->drawPathsAndFlush(ccpr, m);
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastRenderedAtlasID());
|
|
|
|
for (int i = 0; i < 10; ++i) {
|
|
this->drawPathsAndFlush(ccpr, m);
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastRenderedAtlasID());
|
|
}
|
|
|
|
ccpr.ctx()->performDeferredCleanup(std::chrono::milliseconds(0));
|
|
}
|
|
}
|
|
};
|
|
DEF_CCPR_TEST(CCPR_cache_deferredCleanup)
|
|
|
|
// Verifies the cache/hash table internals.
|
|
class CCPR_cache_hashTable : public CCPRCacheTest {
|
|
void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr,
|
|
const RecordLastMockAtlasIDs& atlasIDRecorder) override {
|
|
using CoverageType = GrCCAtlas::CoverageType;
|
|
SkMatrix m = SkMatrix::MakeScale(20, 20);
|
|
|
|
for (int i = 0; i < 5; ++i) {
|
|
this->drawPathsAndFlush(ccpr, m);
|
|
if (2 == i) {
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastCopyAtlasID());
|
|
} else {
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
}
|
|
if (i < 2) {
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastRenderedAtlasID());
|
|
} else {
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastRenderedAtlasID());
|
|
}
|
|
|
|
auto cache = ccpr.ccpr()->testingOnly_getPathCache();
|
|
REPORTER_ASSERT(reporter, cache);
|
|
|
|
const auto& hash = cache->testingOnly_getHashTable();
|
|
const auto& lru = cache->testingOnly_getLRU();
|
|
int count = 0;
|
|
for (GrCCPathCacheEntry* entry : lru) {
|
|
auto* node = hash.find(entry->cacheKey());
|
|
REPORTER_ASSERT(reporter, node);
|
|
REPORTER_ASSERT(reporter, node->entry() == entry);
|
|
REPORTER_ASSERT(reporter, 0 == entry->testingOnly_peekOnFlushRefCnt());
|
|
REPORTER_ASSERT(reporter, entry->unique());
|
|
if (0 == i) {
|
|
REPORTER_ASSERT(reporter, !entry->cachedAtlas());
|
|
} else {
|
|
const GrCCCachedAtlas* cachedAtlas = entry->cachedAtlas();
|
|
REPORTER_ASSERT(reporter, cachedAtlas);
|
|
if (1 == i) {
|
|
REPORTER_ASSERT(reporter, CoverageType::kFP16_CoverageCount
|
|
== cachedAtlas->coverageType());
|
|
} else {
|
|
REPORTER_ASSERT(reporter, CoverageType::kA8_LiteralCoverage
|
|
== cachedAtlas->coverageType());
|
|
}
|
|
REPORTER_ASSERT(reporter, cachedAtlas->textureKey().isValid());
|
|
// The actual proxy should not be held past the end of a flush.
|
|
REPORTER_ASSERT(reporter, !cachedAtlas->getOnFlushProxy());
|
|
REPORTER_ASSERT(reporter, 0 == cachedAtlas->testingOnly_peekOnFlushRefCnt());
|
|
}
|
|
++count;
|
|
}
|
|
REPORTER_ASSERT(reporter, hash.count() == count);
|
|
}
|
|
}
|
|
};
|
|
DEF_CCPR_TEST(CCPR_cache_hashTable)
|
|
|
|
// Ensures paths get cached even when using a sporadic flushing pattern and drawing out of order
|
|
// (a la Chrome tiles).
|
|
class CCPR_cache_multiFlush : public CCPRCacheTest {
|
|
void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr,
|
|
const RecordLastMockAtlasIDs& atlasIDRecorder) override {
|
|
static constexpr int kNumPaths = SK_ARRAY_COUNT(fPaths);
|
|
static constexpr int kBigPrimes[] = {
|
|
9323, 11059, 22993, 38749, 45127, 53147, 64853, 77969, 83269, 99989};
|
|
|
|
SkRandom rand;
|
|
SkMatrix m = SkMatrix::I();
|
|
|
|
for (size_t i = 0; i < SK_ARRAY_COUNT(kBigPrimes); ++i) {
|
|
int prime = kBigPrimes[i];
|
|
int endPathIdx = (int)rand.nextULessThan(kNumPaths);
|
|
int pathIdx = endPathIdx;
|
|
int nextFlush = rand.nextRangeU(1, 47);
|
|
for (int j = 0; j < kNumPaths; ++j) {
|
|
pathIdx = (pathIdx + prime) % kNumPaths;
|
|
int repeat = rand.nextRangeU(1, 3);
|
|
for (int k = 0; k < repeat; ++k) {
|
|
ccpr.drawPath(fPaths[pathIdx], m);
|
|
}
|
|
if (nextFlush == j) {
|
|
ccpr.flush();
|
|
// The paths are small enough that we should never copy to an A8 atlas.
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
if (i < 2) {
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastRenderedAtlasID());
|
|
} else {
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastRenderedAtlasID());
|
|
}
|
|
nextFlush = SkTMin(j + (int)rand.nextRangeU(1, 29), kNumPaths - 1);
|
|
}
|
|
}
|
|
SkASSERT(endPathIdx == pathIdx % kNumPaths);
|
|
}
|
|
}
|
|
};
|
|
DEF_CCPR_TEST(CCPR_cache_multiFlush)
|
|
|
|
// Ensures a path drawn over mutiple tiles gets cached.
|
|
class CCPR_cache_multiTileCache : public CCPRCacheTest {
|
|
void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr,
|
|
const RecordLastMockAtlasIDs& atlasIDRecorder) override {
|
|
// Make sure a path drawn over 9 tiles gets cached (1 tile out of 9 is >10% visibility).
|
|
const SkMatrix m0 = SkMatrix::MakeScale(kCanvasSize*3, kCanvasSize*3);
|
|
const SkPath p0 = fPaths[0];
|
|
for (int i = 0; i < 9; ++i) {
|
|
static constexpr int kRowOrder[9] = {0,1,1,0,2,2,2,1,0};
|
|
static constexpr int kColumnOrder[9] = {0,0,1,1,0,1,2,2,2};
|
|
|
|
SkMatrix tileM = m0;
|
|
tileM.postTranslate(-kCanvasSize * kColumnOrder[i], -kCanvasSize * kRowOrder[i]);
|
|
ccpr.drawPath(p0, tileM);
|
|
ccpr.flush();
|
|
if (i < 5) {
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastRenderedAtlasID());
|
|
} else if (5 == i) {
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastRenderedAtlasID());
|
|
} else {
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastRenderedAtlasID());
|
|
}
|
|
}
|
|
|
|
// Now make sure paths don't get cached when visibility is <10% for every draw (12 tiles).
|
|
const SkMatrix m1 = SkMatrix::MakeScale(kCanvasSize*4, kCanvasSize*3);
|
|
const SkPath p1 = fPaths[1];
|
|
for (int row = 0; row < 3; ++row) {
|
|
for (int col = 0; col < 4; ++col) {
|
|
SkMatrix tileM = m1;
|
|
tileM.postTranslate(-kCanvasSize * col, -kCanvasSize * row);
|
|
ccpr.drawPath(p1, tileM);
|
|
ccpr.flush();
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastRenderedAtlasID());
|
|
}
|
|
}
|
|
|
|
// Double-check the cache is still intact.
|
|
ccpr.drawPath(p0, m0);
|
|
ccpr.flush();
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastRenderedAtlasID());
|
|
|
|
ccpr.drawPath(p1, m1);
|
|
ccpr.flush();
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastRenderedAtlasID());
|
|
}
|
|
};
|
|
DEF_CCPR_TEST(CCPR_cache_multiTileCache)
|
|
|
|
// This test exercises CCPR's cache capabilities by drawing many paths with two different
|
|
// transformation matrices. We then vary the matrices independently by whole and partial pixels,
|
|
// and verify the caching behaved as expected.
|
|
class CCPR_cache_partialInvalidate : public CCPRCacheTest {
|
|
void customizeOptions(GrMockOptions*, GrContextOptions* ctxOptions) override {
|
|
ctxOptions->fAllowPathMaskCaching = true;
|
|
}
|
|
|
|
static constexpr int kPathSize = 4;
|
|
|
|
void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr,
|
|
const RecordLastMockAtlasIDs& atlasIDRecorder) override {
|
|
SkMatrix matrices[2] = {
|
|
SkMatrix::MakeTrans(5, 5),
|
|
SkMatrix::MakeTrans(kCanvasSize - kPathSize - 5, kCanvasSize - kPathSize - 5)
|
|
};
|
|
matrices[0].preScale(kPathSize, kPathSize);
|
|
matrices[1].preScale(kPathSize, kPathSize);
|
|
|
|
int firstAtlasID = 0;
|
|
|
|
for (int iterIdx = 0; iterIdx < 4*3*2; ++iterIdx) {
|
|
this->drawPathsAndFlush(ccpr, matrices, 2);
|
|
|
|
if (0 == iterIdx) {
|
|
// First iteration: just note the ID of the stashed atlas and continue.
|
|
firstAtlasID = atlasIDRecorder.lastRenderedAtlasID();
|
|
REPORTER_ASSERT(reporter, 0 != firstAtlasID);
|
|
continue;
|
|
}
|
|
|
|
int testIdx = (iterIdx/2) % 3;
|
|
int repetitionIdx = iterIdx % 2;
|
|
switch (testIdx) {
|
|
case 0:
|
|
if (0 == repetitionIdx) {
|
|
// This is the big test. New paths were drawn twice last round. On hit 2
|
|
// (last time), 'firstAtlasID' was cached as a 16-bit atlas. Now, on hit 3,
|
|
// these paths should be copied out of 'firstAtlasID', and into an A8 atlas.
|
|
// THEN: we should recycle 'firstAtlasID' and reuse that same texture to
|
|
// render the new masks.
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastCopyAtlasID());
|
|
REPORTER_ASSERT(reporter,
|
|
atlasIDRecorder.lastRenderedAtlasID() == firstAtlasID);
|
|
} else {
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
// This is hit 2 for the new masks. Next time they will be copied to an A8
|
|
// atlas.
|
|
REPORTER_ASSERT(reporter,
|
|
atlasIDRecorder.lastRenderedAtlasID() == firstAtlasID);
|
|
}
|
|
|
|
if (1 == repetitionIdx) {
|
|
// Integer translates: all path masks stay valid.
|
|
matrices[0].preTranslate(-1, -1);
|
|
matrices[1].preTranslate(1, 1);
|
|
}
|
|
break;
|
|
|
|
case 1:
|
|
if (0 == repetitionIdx) {
|
|
// New paths were drawn twice last round. The third hit (now) they should be
|
|
// copied to an A8 atlas.
|
|
REPORTER_ASSERT(reporter, 0 != atlasIDRecorder.lastCopyAtlasID());
|
|
} else {
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
}
|
|
|
|
// This draw should have gotten 100% cache hits; we only did integer translates
|
|
// last time (or none if it was the first flush). Therefore, everything should
|
|
// have been cached.
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastRenderedAtlasID());
|
|
|
|
if (1 == repetitionIdx) {
|
|
// Invalidate even path masks.
|
|
matrices[0].preTranslate(1.6f, 1.4f);
|
|
}
|
|
break;
|
|
|
|
case 2:
|
|
// No new masks to copy from last time; it had 100% cache hits.
|
|
REPORTER_ASSERT(reporter, 0 == atlasIDRecorder.lastCopyAtlasID());
|
|
|
|
// Even path masks were invalidated last iteration by a subpixel translate.
|
|
// They should have been re-rendered this time in the original 'firstAtlasID'
|
|
// texture.
|
|
REPORTER_ASSERT(reporter,
|
|
atlasIDRecorder.lastRenderedAtlasID() == firstAtlasID);
|
|
|
|
if (1 == repetitionIdx) {
|
|
// Invalidate odd path masks.
|
|
matrices[1].preTranslate(-1.4f, -1.6f);
|
|
}
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
};
|
|
DEF_CCPR_TEST(CCPR_cache_partialInvalidate)
|
|
|
|
class CCPR_unrefPerOpListPathsBeforeOps : public CCPRTest {
|
|
void onRun(skiatest::Reporter* reporter, CCPRPathDrawer& ccpr) override {
|
|
REPORTER_ASSERT(reporter, SkPathPriv::TestingOnly_unique(fPath));
|
|
for (int i = 0; i < 10000; ++i) {
|
|
// Draw enough paths to make the arena allocator hit the heap.
|
|
ccpr.drawPath(fPath);
|
|
}
|
|
|
|
// Unref the GrCCPerOpListPaths object.
|
|
auto perOpListPathsMap = ccpr.ccpr()->detachPendingPaths();
|
|
perOpListPathsMap.clear();
|
|
|
|
// Now delete the Op and all its draws.
|
|
REPORTER_ASSERT(reporter, !SkPathPriv::TestingOnly_unique(fPath));
|
|
ccpr.flush();
|
|
REPORTER_ASSERT(reporter, SkPathPriv::TestingOnly_unique(fPath));
|
|
}
|
|
};
|
|
DEF_CCPR_TEST(CCPR_unrefPerOpListPathsBeforeOps)
|
|
|
|
class CCPRRenderingTest {
|
|
public:
|
|
void run(skiatest::Reporter* reporter, GrContext* ctx, bool doStroke) const {
|
|
if (!ctx->priv().drawingManager()->getCoverageCountingPathRenderer()) {
|
|
return; // CCPR is not enabled on this GPU.
|
|
}
|
|
CCPRPathDrawer ccpr(sk_ref_sp(ctx), reporter, doStroke);
|
|
if (!ccpr.valid()) {
|
|
return;
|
|
}
|
|
this->onRun(reporter, ccpr);
|
|
}
|
|
|
|
virtual ~CCPRRenderingTest() {}
|
|
|
|
protected:
|
|
virtual void onRun(skiatest::Reporter* reporter, const CCPRPathDrawer& ccpr) const = 0;
|
|
};
|
|
|
|
#define DEF_CCPR_RENDERING_TEST(name) \
|
|
DEF_GPUTEST_FOR_RENDERING_CONTEXTS(name, reporter, ctxInfo) { \
|
|
name test; \
|
|
test.run(reporter, ctxInfo.grContext(), false); \
|
|
test.run(reporter, ctxInfo.grContext(), true); \
|
|
}
|
|
|
|
class CCPR_busyPath : public CCPRRenderingTest {
|
|
void onRun(skiatest::Reporter* reporter, const CCPRPathDrawer& ccpr) const override {
|
|
static constexpr int kNumBusyVerbs = 1 << 17;
|
|
ccpr.clear();
|
|
SkPath busyPath;
|
|
busyPath.moveTo(0, 0); // top left
|
|
busyPath.lineTo(kCanvasSize, kCanvasSize); // bottom right
|
|
for (int i = 2; i < kNumBusyVerbs; ++i) {
|
|
float offset = i * ((float)kCanvasSize / kNumBusyVerbs);
|
|
busyPath.lineTo(kCanvasSize - offset, kCanvasSize + offset); // offscreen
|
|
}
|
|
ccpr.drawPath(busyPath);
|
|
|
|
ccpr.flush(); // If this doesn't crash, the test passed.
|
|
// If it does, maybe fiddle with fMaxInstancesPerDrawArraysWithoutCrashing in
|
|
// your platform's GrGLCaps.
|
|
}
|
|
};
|
|
DEF_CCPR_RENDERING_TEST(CCPR_busyPath)
|