d3895d867c
Centralize the calculation to SkGlyphCacheCommon. This allows this function to be used with NO_GPU. In addition, this was used in the last remaining function in GrTest. That function was used in a single place. I made the function a static and remove GrTest.h. This had massive knock-on effects. Change-Id: I80f874a988f9af4383a83acf7c273d23b8d67c22 Reviewed-on: https://skia-review.googlesource.com/151480 Reviewed-by: Jim Van Verth <jvanverth@google.com> Commit-Queue: Herb Derby <herb@google.com>
276 lines
11 KiB
C++
276 lines
11 KiB
C++
/*
|
|
* Copyright 2017 Google Inc.
|
|
*
|
|
* Use of this source code is governed by a BSD-style license that can be
|
|
* found in the LICENSE file.
|
|
*/
|
|
|
|
#include "SkTypes.h"
|
|
|
|
#include "Test.h"
|
|
|
|
#include "GrContextPriv.h"
|
|
#include "GrGpu.h"
|
|
#include "GrProxyProvider.h"
|
|
#include "GrResourceAllocator.h"
|
|
#include "GrResourceProvider.h"
|
|
#include "GrSurfaceProxyPriv.h"
|
|
#include "GrTexture.h"
|
|
#include "GrTextureProxy.h"
|
|
#include "GrUninstantiateProxyTracker.h"
|
|
|
|
#include "SkSurface.h"
|
|
|
|
struct ProxyParams {
|
|
int fSize;
|
|
bool fIsRT;
|
|
GrPixelConfig fConfig;
|
|
SkBackingFit fFit;
|
|
int fSampleCnt;
|
|
GrSurfaceOrigin fOrigin;
|
|
// TODO: do we care about mipmapping
|
|
};
|
|
|
|
static GrSurfaceProxy* make_deferred(GrProxyProvider* proxyProvider, const ProxyParams& p) {
|
|
GrSurfaceDesc desc;
|
|
desc.fFlags = p.fIsRT ? kRenderTarget_GrSurfaceFlag : kNone_GrSurfaceFlags;
|
|
desc.fWidth = p.fSize;
|
|
desc.fHeight = p.fSize;
|
|
desc.fConfig = p.fConfig;
|
|
desc.fSampleCnt = p.fSampleCnt;
|
|
|
|
auto tmp = proxyProvider->createProxy(desc, p.fOrigin, p.fFit, SkBudgeted::kNo);
|
|
if (!tmp) {
|
|
return nullptr;
|
|
}
|
|
GrSurfaceProxy* ret = tmp.release();
|
|
|
|
// Add a read to keep the proxy around but unref it so its backing surfaces can be recycled
|
|
ret->addPendingRead();
|
|
ret->unref();
|
|
return ret;
|
|
}
|
|
|
|
static GrSurfaceProxy* make_backend(GrContext* context, const ProxyParams& p,
|
|
GrBackendTexture* backendTex) {
|
|
GrProxyProvider* proxyProvider = context->contextPriv().proxyProvider();
|
|
GrGpu* gpu = context->contextPriv().getGpu();
|
|
|
|
*backendTex = gpu->createTestingOnlyBackendTexture(nullptr, p.fSize, p.fSize,
|
|
p.fConfig, false,
|
|
GrMipMapped::kNo);
|
|
|
|
auto tmp = proxyProvider->wrapBackendTexture(*backendTex, p.fOrigin);
|
|
if (!tmp) {
|
|
return nullptr;
|
|
}
|
|
GrSurfaceProxy* ret = tmp.release();
|
|
|
|
// Add a read to keep the proxy around but unref it so its backing surfaces can be recycled
|
|
ret->addPendingRead();
|
|
ret->unref();
|
|
return ret;
|
|
}
|
|
|
|
static void cleanup_backend(GrContext* context, const GrBackendTexture& backendTex) {
|
|
context->contextPriv().getGpu()->deleteTestingOnlyBackendTexture(backendTex);
|
|
}
|
|
|
|
// Basic test that two proxies with overlapping intervals and compatible descriptors are
|
|
// assigned different GrSurfaces.
|
|
static void overlap_test(skiatest::Reporter* reporter, GrResourceProvider* resourceProvider,
|
|
GrSurfaceProxy* p1, GrSurfaceProxy* p2, bool expectedResult) {
|
|
GrResourceAllocator alloc(resourceProvider);
|
|
|
|
alloc.addInterval(p1, 0, 4);
|
|
alloc.addInterval(p2, 1, 2);
|
|
alloc.markEndOfOpList(0);
|
|
|
|
int startIndex, stopIndex;
|
|
GrResourceAllocator::AssignError error;
|
|
GrUninstantiateProxyTracker uninstantiateTracker;
|
|
alloc.assign(&startIndex, &stopIndex, &uninstantiateTracker, &error);
|
|
REPORTER_ASSERT(reporter, GrResourceAllocator::AssignError::kNoError == error);
|
|
|
|
REPORTER_ASSERT(reporter, p1->peekSurface());
|
|
REPORTER_ASSERT(reporter, p2->peekSurface());
|
|
bool doTheBackingStoresMatch = p1->underlyingUniqueID() == p2->underlyingUniqueID();
|
|
REPORTER_ASSERT(reporter, expectedResult == doTheBackingStoresMatch);
|
|
}
|
|
|
|
// Test various cases when two proxies do not have overlapping intervals.
|
|
// This mainly acts as a test of the ResourceAllocator's free pool.
|
|
static void non_overlap_test(skiatest::Reporter* reporter, GrResourceProvider* resourceProvider,
|
|
GrSurfaceProxy* p1, GrSurfaceProxy* p2,
|
|
bool expectedResult) {
|
|
GrResourceAllocator alloc(resourceProvider);
|
|
|
|
alloc.addInterval(p1, 0, 2);
|
|
alloc.addInterval(p2, 3, 5);
|
|
alloc.markEndOfOpList(0);
|
|
|
|
int startIndex, stopIndex;
|
|
GrResourceAllocator::AssignError error;
|
|
GrUninstantiateProxyTracker uninstantiateTracker;
|
|
alloc.assign(&startIndex, &stopIndex, &uninstantiateTracker, &error);
|
|
REPORTER_ASSERT(reporter, GrResourceAllocator::AssignError::kNoError == error);
|
|
|
|
REPORTER_ASSERT(reporter, p1->peekSurface());
|
|
REPORTER_ASSERT(reporter, p2->peekSurface());
|
|
bool doTheBackingStoresMatch = p1->underlyingUniqueID() == p2->underlyingUniqueID();
|
|
REPORTER_ASSERT(reporter, expectedResult == doTheBackingStoresMatch);
|
|
}
|
|
|
|
bool GrResourceProvider::testingOnly_setExplicitlyAllocateGPUResources(bool newValue) {
|
|
bool oldValue = fExplicitlyAllocateGPUResources;
|
|
fExplicitlyAllocateGPUResources = newValue;
|
|
return oldValue;
|
|
}
|
|
|
|
DEF_GPUTEST_FOR_RENDERING_CONTEXTS(ResourceAllocatorTest, reporter, ctxInfo) {
|
|
GrProxyProvider* proxyProvider = ctxInfo.grContext()->contextPriv().proxyProvider();
|
|
GrResourceProvider* resourceProvider = ctxInfo.grContext()->contextPriv().resourceProvider();
|
|
|
|
bool orig = resourceProvider->testingOnly_setExplicitlyAllocateGPUResources(true);
|
|
|
|
struct TestCase {
|
|
ProxyParams fP1;
|
|
ProxyParams fP2;
|
|
bool fExpectation;
|
|
};
|
|
|
|
constexpr bool kRT = true;
|
|
constexpr bool kNotRT = false;
|
|
|
|
constexpr bool kShare = true;
|
|
constexpr bool kDontShare = false;
|
|
// Non-RT GrSurfaces are never recycled on some platforms.
|
|
bool kConditionallyShare = resourceProvider->caps()->reuseScratchTextures();
|
|
|
|
const GrPixelConfig kRGBA = kRGBA_8888_GrPixelConfig;
|
|
const GrPixelConfig kBGRA = kBGRA_8888_GrPixelConfig;
|
|
|
|
const SkBackingFit kE = SkBackingFit::kExact;
|
|
const SkBackingFit kA = SkBackingFit::kApprox;
|
|
|
|
const GrSurfaceOrigin kTL = kTopLeft_GrSurfaceOrigin;
|
|
const GrSurfaceOrigin kBL = kBottomLeft_GrSurfaceOrigin;
|
|
|
|
//--------------------------------------------------------------------------------------------
|
|
TestCase gOverlappingTests[] = {
|
|
//----------------------------------------------------------------------------------------
|
|
// Two proxies with overlapping intervals and compatible descriptors should never share
|
|
// RT version
|
|
{ { 64, kRT, kRGBA, kA, 0, kTL }, { 64, kRT, kRGBA, kA, 0, kTL }, kDontShare },
|
|
// non-RT version
|
|
{ { 64, kNotRT, kRGBA, kA, 0, kTL }, { 64, kNotRT, kRGBA, kA, 0, kTL }, kDontShare },
|
|
};
|
|
|
|
for (auto test : gOverlappingTests) {
|
|
GrSurfaceProxy* p1 = make_deferred(proxyProvider, test.fP1);
|
|
GrSurfaceProxy* p2 = make_deferred(proxyProvider, test.fP2);
|
|
overlap_test(reporter, resourceProvider, p1, p2, test.fExpectation);
|
|
p1->completedRead();
|
|
p2->completedRead();
|
|
}
|
|
|
|
int k2 = ctxInfo.grContext()->contextPriv().caps()->getRenderTargetSampleCount(2, kRGBA);
|
|
int k4 = ctxInfo.grContext()->contextPriv().caps()->getRenderTargetSampleCount(4, kRGBA);
|
|
|
|
//--------------------------------------------------------------------------------------------
|
|
TestCase gNonOverlappingTests[] = {
|
|
//----------------------------------------------------------------------------------------
|
|
// Two non-overlapping intervals w/ compatible proxies should share
|
|
// both same size & approx
|
|
{ { 64, kRT, kRGBA, kA, 0, kTL }, { 64, kRT, kRGBA, kA, 0, kTL }, kShare },
|
|
{ { 64, kNotRT, kRGBA, kA, 0, kTL }, { 64, kNotRT, kRGBA, kA, 0, kTL }, kConditionallyShare },
|
|
// diffs sizes but still approx
|
|
{ { 64, kRT, kRGBA, kA, 0, kTL }, { 50, kRT, kRGBA, kA, 0, kTL }, kShare },
|
|
{ { 64, kNotRT, kRGBA, kA, 0, kTL }, { 50, kNotRT, kRGBA, kA, 0, kTL }, kConditionallyShare },
|
|
// sames sizes but exact
|
|
{ { 64, kRT, kRGBA, kE, 0, kTL }, { 64, kRT, kRGBA, kE, 0, kTL }, kShare },
|
|
{ { 64, kNotRT, kRGBA, kE, 0, kTL }, { 64, kNotRT, kRGBA, kE, 0, kTL }, kConditionallyShare },
|
|
//----------------------------------------------------------------------------------------
|
|
// Two non-overlapping intervals w/ different exact sizes should not share
|
|
{ { 56, kRT, kRGBA, kE, 0, kTL }, { 54, kRT, kRGBA, kE, 0, kTL }, kDontShare },
|
|
// Two non-overlapping intervals w/ _very different_ approx sizes should not share
|
|
{ { 255, kRT, kRGBA, kA, 0, kTL }, { 127, kRT, kRGBA, kA, 0, kTL }, kDontShare },
|
|
// Two non-overlapping intervals w/ different MSAA sample counts should not share
|
|
{ { 64, kRT, kRGBA, kA, k2, kTL },{ 64, kRT, kRGBA, kA, k4, kTL}, k2 == k4 },
|
|
// Two non-overlapping intervals w/ different configs should not share
|
|
{ { 64, kRT, kRGBA, kA, 0, kTL }, { 64, kRT, kBGRA, kA, 0, kTL }, kDontShare },
|
|
// Two non-overlapping intervals w/ different RT classifications should never share
|
|
{ { 64, kRT, kRGBA, kA, 0, kTL }, { 64, kNotRT, kRGBA, kA, 0, kTL }, kDontShare },
|
|
{ { 64, kNotRT, kRGBA, kA, 0, kTL }, { 64, kRT, kRGBA, kA, 0, kTL }, kDontShare },
|
|
// Two non-overlapping intervals w/ different origins should share
|
|
{ { 64, kRT, kRGBA, kA, 0, kTL }, { 64, kRT, kRGBA, kA, 0, kBL }, kShare },
|
|
};
|
|
|
|
for (auto test : gNonOverlappingTests) {
|
|
GrSurfaceProxy* p1 = make_deferred(proxyProvider, test.fP1);
|
|
GrSurfaceProxy* p2 = make_deferred(proxyProvider, test.fP2);
|
|
|
|
if (!p1 || !p2) {
|
|
continue; // creation can fail (i.e., for msaa4 on iOS)
|
|
}
|
|
|
|
non_overlap_test(reporter, resourceProvider, p1, p2, test.fExpectation);
|
|
|
|
p1->completedRead();
|
|
p2->completedRead();
|
|
}
|
|
|
|
{
|
|
// Wrapped backend textures should never be reused
|
|
TestCase t[1] = {
|
|
{ { 64, kNotRT, kRGBA, kE, 0, kTL }, { 64, kNotRT, kRGBA, kE, 0, kTL }, kDontShare }
|
|
};
|
|
|
|
GrBackendTexture backEndTex;
|
|
GrSurfaceProxy* p1 = make_backend(ctxInfo.grContext(), t[0].fP1, &backEndTex);
|
|
GrSurfaceProxy* p2 = make_deferred(proxyProvider, t[0].fP2);
|
|
|
|
non_overlap_test(reporter, resourceProvider, p1, p2, t[0].fExpectation);
|
|
|
|
p1->completedRead();
|
|
p2->completedRead();
|
|
|
|
cleanup_backend(ctxInfo.grContext(), backEndTex);
|
|
}
|
|
|
|
resourceProvider->testingOnly_setExplicitlyAllocateGPUResources(orig);
|
|
}
|
|
|
|
static void draw(GrContext* context) {
|
|
SkImageInfo ii = SkImageInfo::Make(1024, 1024, kRGBA_8888_SkColorType, kPremul_SkAlphaType);
|
|
|
|
sk_sp<SkSurface> s = SkSurface::MakeRenderTarget(context, SkBudgeted::kYes,
|
|
ii, 1, kTopLeft_GrSurfaceOrigin, nullptr);
|
|
|
|
SkCanvas* c = s->getCanvas();
|
|
|
|
c->clear(SK_ColorBLACK);
|
|
}
|
|
|
|
|
|
DEF_GPUTEST_FOR_RENDERING_CONTEXTS(ResourceAllocatorStressTest, reporter, ctxInfo) {
|
|
GrContext* context = ctxInfo.grContext();
|
|
GrResourceProvider* resourceProvider = ctxInfo.grContext()->contextPriv().resourceProvider();
|
|
|
|
int maxNum;
|
|
size_t maxBytes;
|
|
context->getResourceCacheLimits(&maxNum, &maxBytes);
|
|
|
|
bool orig = resourceProvider->testingOnly_setExplicitlyAllocateGPUResources(true);
|
|
context->setResourceCacheLimits(0, 0); // We'll always be overbudget
|
|
|
|
draw(context);
|
|
draw(context);
|
|
draw(context);
|
|
draw(context);
|
|
context->flush();
|
|
|
|
context->setResourceCacheLimits(maxNum, maxBytes);
|
|
resourceProvider->testingOnly_setExplicitlyAllocateGPUResources(orig);
|
|
}
|