/* * Copyright 2017 Google Inc. * * Use of this source code is governed by a BSD-style license that can be * found in the LICENSE file. */ #include "include/core/SkCanvas.h" #include "include/core/SkSurface.h" #include "include/gpu/GrDirectContext.h" #include "src/core/SkSpan.h" #include "src/gpu/GrDirectContextPriv.h" #include "src/gpu/GrGpu.h" #include "src/gpu/GrProxyProvider.h" #include "src/gpu/GrResourceAllocator.h" #include "src/gpu/GrResourceProviderPriv.h" #include "src/gpu/GrSurfaceProxyPriv.h" #include "src/gpu/GrTexture.h" #include "src/gpu/GrTextureProxy.h" #include "tests/Test.h" #include "tools/gpu/ManagedBackendTexture.h" namespace { struct ProxyParams { int fSize; GrRenderable fRenderable; GrColorType fColorType; SkBackingFit fFit; int fSampleCnt; SkBudgeted fBudgeted; enum Kind { kDeferred, kBackend, kFullyLazy, kLazy, kInstantiated }; Kind fKind; GrUniqueKey fUniqueKey = GrUniqueKey(); // TODO: do we care about mipmapping }; constexpr GrRenderable kRT = GrRenderable::kYes; constexpr GrRenderable kNotRT = GrRenderable::kNo; constexpr GrColorType kRGBA = GrColorType::kRGBA_8888; constexpr GrColorType kAlpha = GrColorType::kAlpha_8; constexpr SkBackingFit kE = SkBackingFit::kExact; constexpr SkBackingFit kA = SkBackingFit::kApprox; constexpr SkBudgeted kNotB = SkBudgeted::kNo; constexpr SkBudgeted kB = SkBudgeted::kYes; constexpr ProxyParams::Kind kDeferred = ProxyParams::Kind::kDeferred; constexpr ProxyParams::Kind kBackend = ProxyParams::Kind::kBackend; constexpr ProxyParams::Kind kInstantiated = ProxyParams::Kind::kInstantiated; constexpr ProxyParams::Kind kLazy = ProxyParams::Kind::kLazy; constexpr ProxyParams::Kind kFullyLazy = ProxyParams::Kind::kFullyLazy; }; static sk_sp make_deferred(GrProxyProvider* proxyProvider, const GrCaps* caps, const ProxyParams& p) { const GrBackendFormat format = caps->getDefaultBackendFormat(p.fColorType, p.fRenderable); return proxyProvider->createProxy(format, {p.fSize, p.fSize}, p.fRenderable, p.fSampleCnt, GrMipmapped::kNo, p.fFit, p.fBudgeted, GrProtected::kNo); } static sk_sp make_backend(GrDirectContext* dContext, const ProxyParams& p) { GrProxyProvider* proxyProvider = dContext->priv().proxyProvider(); SkColorType skColorType = GrColorTypeToSkColorType(p.fColorType); SkASSERT(SkColorType::kUnknown_SkColorType != skColorType); auto mbet = sk_gpu_test::ManagedBackendTexture::MakeWithoutData( dContext, p.fSize, p.fSize, skColorType, GrMipmapped::kNo, GrRenderable::kNo); if (!mbet) { return nullptr; } return proxyProvider->wrapBackendTexture(mbet->texture(), kBorrow_GrWrapOwnership, GrWrapCacheable::kNo, kRead_GrIOType, mbet->refCountedCallback()); } static sk_sp make_fully_lazy(GrProxyProvider* proxyProvider, const GrCaps* caps, const ProxyParams& p) { const GrBackendFormat format = caps->getDefaultBackendFormat(p.fColorType, p.fRenderable); auto cb = [p](GrResourceProvider* provider, const GrSurfaceProxy::LazySurfaceDesc& desc) { auto tex = provider->createTexture({p.fSize, p.fSize}, desc.fFormat, desc.fRenderable, desc.fSampleCnt, desc.fMipmapped, desc.fBudgeted, desc.fProtected); return GrSurfaceProxy::LazyCallbackResult(std::move(tex)); }; return GrProxyProvider::MakeFullyLazyProxy(std::move(cb), format, p.fRenderable, p.fSampleCnt, GrProtected::kNo, *caps, GrSurfaceProxy::UseAllocator::kYes); } static sk_sp make_lazy(GrProxyProvider* proxyProvider, const GrCaps* caps, const ProxyParams& p) { const GrBackendFormat format = caps->getDefaultBackendFormat(p.fColorType, p.fRenderable); auto cb = [](GrResourceProvider* provider, const GrSurfaceProxy::LazySurfaceDesc& desc) { auto tex = provider->createTexture(desc.fDimensions, desc.fFormat, desc.fRenderable, desc.fSampleCnt, desc.fMipmapped, desc.fBudgeted, desc.fProtected); return GrSurfaceProxy::LazyCallbackResult(std::move(tex)); }; return proxyProvider->createLazyProxy(std::move(cb), format, {p.fSize, p.fSize}, GrMipmapped::kNo, GrMipmapStatus::kNotAllocated, GrInternalSurfaceFlags::kNone, p.fFit, p.fBudgeted, GrProtected::kNo, GrSurfaceProxy::UseAllocator::kYes); } static sk_sp make_proxy(GrDirectContext* dContext, const ProxyParams& p) { GrProxyProvider* proxyProvider = dContext->priv().proxyProvider(); const GrCaps* caps = dContext->priv().caps(); sk_sp proxy; switch (p.fKind) { case ProxyParams::kDeferred: proxy = make_deferred(proxyProvider, caps, p); break; case ProxyParams::kBackend: proxy = make_backend(dContext, p); break; case ProxyParams::kFullyLazy: proxy = make_fully_lazy(proxyProvider, caps, p); break; case ProxyParams::kLazy: proxy = make_lazy(proxyProvider, caps, p); break; case ProxyParams::kInstantiated: proxy = make_deferred(proxyProvider, caps, p); if (proxy) { auto surf = proxy->priv().createSurface(dContext->priv().resourceProvider()); proxy->priv().assign(std::move(surf)); } break; } if (proxy && p.fUniqueKey.isValid()) { SkASSERT(proxy->asTextureProxy()); proxyProvider->assignUniqueKeyToProxy(p.fUniqueKey, proxy->asTextureProxy()); } return proxy; } // Basic test that two proxies with overlapping intervals and compatible descriptors are // assigned different GrSurfaces. static void overlap_test(skiatest::Reporter* reporter, GrDirectContext* dContext, sk_sp p1, sk_sp p2, bool expectedResult) { GrResourceAllocator alloc(dContext); alloc.addInterval(p1.get(), 0, 4, GrResourceAllocator::ActualUse::kYes); alloc.incOps(); alloc.addInterval(p2.get(), 1, 2, GrResourceAllocator::ActualUse::kYes); alloc.incOps(); REPORTER_ASSERT(reporter, alloc.planAssignment()); REPORTER_ASSERT(reporter, alloc.makeBudgetHeadroom()); REPORTER_ASSERT(reporter, alloc.assign()); REPORTER_ASSERT(reporter, p1->peekSurface()); REPORTER_ASSERT(reporter, p2->peekSurface()); bool doTheBackingStoresMatch = p1->underlyingUniqueID() == p2->underlyingUniqueID(); REPORTER_ASSERT(reporter, expectedResult == doTheBackingStoresMatch); } // Test various cases when two proxies do not have overlapping intervals. // This mainly acts as a test of the ResourceAllocator's free pool. static void non_overlap_test(skiatest::Reporter* reporter, GrDirectContext* dContext, sk_sp p1, sk_sp p2, bool expectedResult) { GrResourceAllocator alloc(dContext); alloc.incOps(); alloc.incOps(); alloc.incOps(); alloc.incOps(); alloc.incOps(); alloc.incOps(); alloc.addInterval(p1.get(), 0, 2, GrResourceAllocator::ActualUse::kYes); alloc.addInterval(p2.get(), 3, 5, GrResourceAllocator::ActualUse::kYes); REPORTER_ASSERT(reporter, alloc.planAssignment()); REPORTER_ASSERT(reporter, alloc.makeBudgetHeadroom()); REPORTER_ASSERT(reporter, alloc.assign()); REPORTER_ASSERT(reporter, p1->peekSurface()); REPORTER_ASSERT(reporter, p2->peekSurface()); bool doTheBackingStoresMatch = p1->underlyingUniqueID() == p2->underlyingUniqueID(); REPORTER_ASSERT(reporter, expectedResult == doTheBackingStoresMatch); } DEF_GPUTEST_FOR_RENDERING_CONTEXTS(ResourceAllocatorTest, reporter, ctxInfo) { auto dContext = ctxInfo.directContext(); const GrCaps* caps = dContext->priv().caps(); struct TestCase { ProxyParams fP1; ProxyParams fP2; bool fExpectation; }; constexpr bool kShare = true; constexpr bool kDontShare = false; // Non-RT GrSurfaces are never recycled on some platforms. bool kConditionallyShare = caps->reuseScratchTextures(); static const TestCase overlappingTests[] = { // Two proxies with overlapping intervals and compatible descriptors should never share // RT version {{64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, {64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, kDontShare}, // non-RT version {{64, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, {64, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, kDontShare}, }; for (size_t i = 0; i < SK_ARRAY_COUNT(overlappingTests); i++) { const TestCase& test = overlappingTests[i]; sk_sp p1 = make_proxy(dContext, test.fP1); sk_sp p2 = make_proxy(dContext, test.fP2); reporter->push(SkStringPrintf("case %d", SkToInt(i))); overlap_test(reporter, dContext, std::move(p1), std::move(p2), test.fExpectation); reporter->pop(); } auto beFormat = caps->getDefaultBackendFormat(GrColorType::kRGBA_8888, GrRenderable::kYes); int k2 = caps->getRenderTargetSampleCount(2, beFormat); int k4 = caps->getRenderTargetSampleCount(4, beFormat); static const TestCase nonOverlappingTests[] = { // Two non-overlapping intervals w/ compatible proxies should share // both same size & approx {{64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, {64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, kShare}, {{64, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, {64, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, kConditionallyShare}, // diffs sizes but still approx {{64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, {50, kRT, kRGBA, kA, 1, kNotB, kDeferred}, kShare}, {{64, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, {50, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, kConditionallyShare}, // sames sizes but exact {{64, kRT, kRGBA, kE, 1, kNotB, kDeferred}, {64, kRT, kRGBA, kE, 1, kNotB, kDeferred}, kShare}, {{64, kNotRT, kRGBA, kE, 1, kNotB, kDeferred}, {64, kNotRT, kRGBA, kE, 1, kNotB, kDeferred}, kConditionallyShare}, // Two non-overlapping intervals w/ different exact sizes should not share {{56, kRT, kRGBA, kE, 1, kNotB, kDeferred}, {54, kRT, kRGBA, kE, 1, kNotB, kDeferred}, kDontShare}, // Two non-overlapping intervals w/ _very different_ approx sizes should not share {{255, kRT, kRGBA, kA, 1, kNotB, kDeferred}, {127, kRT, kRGBA, kA, 1, kNotB, kDeferred}, kDontShare}, // Two non-overlapping intervals w/ different MSAA sample counts should not share {{64, kRT, kRGBA, kA, k2, kNotB, kDeferred}, {64, kRT, kRGBA, kA, k4, kNotB, kDeferred}, k2 == k4}, // Two non-overlapping intervals w/ different configs should not share {{64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, {64, kRT, kAlpha, kA, 1, kNotB, kDeferred}, kDontShare}, // Two non-overlapping intervals w/ different RT classifications should never share {{64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, {64, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, kDontShare}, {{64, kNotRT, kRGBA, kA, 1, kNotB, kDeferred}, {64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, kDontShare}, // Two non-overlapping intervals w/ different origins should share {{64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, {64, kRT, kRGBA, kA, 1, kNotB, kDeferred}, kShare}, // Wrapped backend textures should never be reused {{64, kNotRT, kRGBA, kE, 1, kNotB, kBackend}, {64, kNotRT, kRGBA, kE, 1, kNotB, kDeferred}, kDontShare} }; for (size_t i = 0; i < SK_ARRAY_COUNT(nonOverlappingTests); i++) { const TestCase& test = nonOverlappingTests[i]; sk_sp p1 = make_proxy(dContext, test.fP1); sk_sp p2 = make_proxy(dContext, test.fP2); if (!p1 || !p2) { continue; // creation can fail (e.g., for msaa4 on iOS) } reporter->push(SkStringPrintf("case %d", SkToInt(i))); non_overlap_test(reporter, dContext, std::move(p1), std::move(p2), test.fExpectation); reporter->pop(); } } static void draw(GrRecordingContext* rContext) { SkImageInfo ii = SkImageInfo::Make(1024, 1024, kRGBA_8888_SkColorType, kPremul_SkAlphaType); sk_sp s = SkSurface::MakeRenderTarget(rContext, SkBudgeted::kYes, ii, 1, kTopLeft_GrSurfaceOrigin, nullptr); SkCanvas* c = s->getCanvas(); c->clear(SK_ColorBLACK); } DEF_GPUTEST_FOR_RENDERING_CONTEXTS(ResourceAllocatorStressTest, reporter, ctxInfo) { auto context = ctxInfo.directContext(); size_t maxBytes = context->getResourceCacheLimit(); context->setResourceCacheLimit(0); // We'll always be overbudget draw(context); draw(context); draw(context); draw(context); context->flushAndSubmit(); context->setResourceCacheLimit(maxBytes); } struct Interval { ProxyParams fParams; int fStart; int fEnd; sk_sp fProxy = nullptr; }; struct TestCase { const char * fName; bool fShouldFit; size_t fBudget; SkTArray fPurgeableResourcesInCache = {}; SkTArray fUnpurgeableResourcesInCache = {}; SkTArray fIntervals; }; static void memory_budget_test(skiatest::Reporter* reporter, GrDirectContext* dContext, const TestCase& test) { // Reset cache. auto cache = dContext->priv().getResourceCache(); cache->releaseAll(); cache->setLimit(test.fBudget); // Add purgeable entries. size_t expectedPurgeableBytes = 0; SkTArray> purgeableSurfaces; for (auto& params : test.fPurgeableResourcesInCache) { SkASSERT(params.fKind == kInstantiated); sk_sp proxy = make_proxy(dContext, params); REPORTER_ASSERT(reporter, proxy->peekSurface()); expectedPurgeableBytes += proxy->gpuMemorySize(); purgeableSurfaces.push_back(sk_ref_sp(proxy->peekSurface())); } purgeableSurfaces.reset(); REPORTER_ASSERT(reporter, expectedPurgeableBytes == cache->getPurgeableBytes(), "%zu", cache->getPurgeableBytes()); // Add unpurgeable entries. size_t expectedUnpurgeableBytes = 0; SkTArray> unpurgeableSurfaces; for (auto& params : test.fUnpurgeableResourcesInCache) { SkASSERT(params.fKind == kInstantiated); sk_sp proxy = make_proxy(dContext, params); REPORTER_ASSERT(reporter, proxy->peekSurface()); expectedUnpurgeableBytes += proxy->gpuMemorySize(); unpurgeableSurfaces.push_back(sk_ref_sp(proxy->peekSurface())); } auto unpurgeableBytes = cache->getBudgetedResourceBytes() - cache->getPurgeableBytes(); REPORTER_ASSERT(reporter, expectedUnpurgeableBytes == unpurgeableBytes, "%zu", unpurgeableBytes); // Add intervals and test. GrResourceAllocator alloc(dContext); for (auto& interval : test.fIntervals) { for (int i = interval.fStart; i <= interval.fEnd; i++) { alloc.incOps(); } alloc.addInterval(interval.fProxy.get(), interval.fStart, interval.fEnd, GrResourceAllocator::ActualUse::kYes); } REPORTER_ASSERT(reporter, alloc.planAssignment()); REPORTER_ASSERT(reporter, alloc.makeBudgetHeadroom() == test.fShouldFit); } DEF_GPUTEST_FOR_RENDERING_CONTEXTS(ResourceAllocatorMemoryBudgetTest, reporter, ctxInfo) { auto dContext = ctxInfo.directContext(); constexpr bool kUnder = true; constexpr bool kOver = false; constexpr size_t kRGBA64Bytes = 4 * 64 * 64; const ProxyParams kProxy64 = {64, kRT, kRGBA, kE, 1, kB, kDeferred}; const ProxyParams kProxy64NotBudgeted = {64, kRT, kRGBA, kE, 1, kNotB, kDeferred}; const ProxyParams kProxy64Lazy = {64, kRT, kRGBA, kE, 1, kB, kLazy}; const ProxyParams kProxy64FullyLazy = {64, kRT, kRGBA, kE, 1, kB, kFullyLazy}; const ProxyParams kProxy32Instantiated = {32, kRT, kRGBA, kE, 1, kB, kInstantiated}; const ProxyParams kProxy64Instantiated = {64, kRT, kRGBA, kE, 1, kB, kInstantiated}; TestCase tests[] = { {"empty DAG", kUnder, 0, {}, {}, {}}, {"unbudgeted", kUnder, 0, {}, {}, {{kProxy64NotBudgeted, 0, 2}}}, {"basic", kUnder, kRGBA64Bytes, {}, {}, {{kProxy64, 0, 2}}}, {"basic, over", kOver, kRGBA64Bytes - 1, {}, {}, {{kProxy64, 0, 2}}}, {"shared", kUnder, kRGBA64Bytes, {}, {}, { {kProxy64, 0, 2}, {kProxy64, 3, 5}, }}, {"retrieved from cache", kUnder, kRGBA64Bytes, /* purgeable */{kProxy64Instantiated}, /* unpurgeable */{}, { {kProxy64, 0, 2} }}, {"purge 4", kUnder, kRGBA64Bytes, /* purgeable */{ kProxy32Instantiated, kProxy32Instantiated, kProxy32Instantiated, kProxy32Instantiated }, /* unpurgeable */{}, { {kProxy64, 0, 2} }}, {"dont purge what we've reserved", kOver, kRGBA64Bytes, /* purgeable */{kProxy64Instantiated}, /* unpurgeable */{}, { {kProxy64, 0, 2}, {kProxy64, 1, 3} }}, {"unpurgeable", kOver, kRGBA64Bytes, /* purgeable */{}, /* unpurgeable */{kProxy64Instantiated}, { {kProxy64, 0, 2} }}, {"lazy", kUnder, kRGBA64Bytes, /* purgeable */{}, /* unpurgeable */{}, { {kProxy64Lazy, 0, 2} }}, {"lazy, over", kOver, kRGBA64Bytes - 1, /* purgeable */{}, /* unpurgeable */{}, { {kProxy64Lazy, 0, 2} }}, {"fully-lazy", kUnder, kRGBA64Bytes, /* purgeable */{}, /* unpurgeable */{}, { {kProxy64FullyLazy, 0, 2} }}, {"fully-lazy, over", kOver, kRGBA64Bytes - 1, /* purgeable */{}, /* unpurgeable */{}, { {kProxy64FullyLazy, 0, 2} }}, }; SkString match(""); for (size_t i = 0; i < SK_ARRAY_COUNT(tests); i++) { TestCase& test = tests[i]; if (match.isEmpty() || match == SkString(test.fName)) { // Create proxies for (Interval& interval : test.fIntervals) { interval.fProxy = make_proxy(dContext, interval.fParams); } reporter->push(SkString(test.fName)); memory_budget_test(reporter, dContext, test); reporter->pop(); } } }