| /* |
| * Copyright 2018 Google Inc. |
| * |
| * Use of this source code is governed by a BSD-style license that can be |
| * found in the LICENSE file. |
| */ |
| #include "include/gpu/GrDirectContext.h" |
| |
| #include "include/core/SkImageInfo.h" |
| #include "include/core/SkPixmap.h" |
| #include "include/core/SkSize.h" |
| #include "include/core/SkSurface.h" |
| #include "include/core/SkTextureCompressionType.h" |
| #include "include/core/SkTraceMemoryDump.h" |
| #include "include/gpu/GpuTypes.h" |
| #include "include/gpu/GrBackendSemaphore.h" |
| #include "include/gpu/GrBackendSurface.h" |
| #include "include/gpu/GrContextThreadSafeProxy.h" |
| #include "include/private/base/SingleOwner.h" |
| #include "include/private/base/SkTArray.h" |
| #include "include/private/base/SkTemplates.h" |
| #include "include/private/gpu/ganesh/GrTypesPriv.h" |
| #include "src/base/SkAutoMalloc.h" |
| #include "src/core/SkCompressedDataUtils.h" |
| #include "src/core/SkMipmap.h" |
| #include "src/core/SkTaskGroup.h" |
| #include "src/core/SkTraceEvent.h" |
| #include "src/gpu/DataUtils.h" |
| #include "src/gpu/GpuTypesPriv.h" |
| #include "src/gpu/RefCntedCallback.h" |
| #include "src/gpu/Swizzle.h" |
| #include "src/gpu/ganesh/Device.h" |
| #include "src/gpu/ganesh/GrBackendUtils.h" |
| #include "src/gpu/ganesh/GrCaps.h" |
| #include "src/gpu/ganesh/GrClientMappedBufferManager.h" |
| #include "src/gpu/ganesh/GrColorInfo.h" |
| #include "src/gpu/ganesh/GrContextThreadSafeProxyPriv.h" |
| #include "src/gpu/ganesh/GrDirectContextPriv.h" |
| #include "src/gpu/ganesh/GrDrawOpAtlas.h" |
| #include "src/gpu/ganesh/GrDrawingManager.h" |
| #include "src/gpu/ganesh/GrGpu.h" |
| #include "src/gpu/ganesh/GrPixmap.h" |
| #include "src/gpu/ganesh/GrProxyProvider.h" |
| #include "src/gpu/ganesh/GrRenderTargetProxy.h" |
| #include "src/gpu/ganesh/GrResourceCache.h" |
| #include "src/gpu/ganesh/GrResourceProvider.h" |
| #include "src/gpu/ganesh/GrSemaphore.h" // IWYU pragma: keep |
| #include "src/gpu/ganesh/GrShaderCaps.h" |
| #include "src/gpu/ganesh/GrSurfaceProxy.h" |
| #include "src/gpu/ganesh/GrSurfaceProxyView.h" |
| #include "src/gpu/ganesh/GrThreadSafePipelineBuilder.h" // IWYU pragma: keep |
| #include "src/gpu/ganesh/SurfaceContext.h" |
| #include "src/gpu/ganesh/image/SkImage_GaneshBase.h" |
| #include "src/gpu/ganesh/mock/GrMockGpu.h" |
| #include "src/gpu/ganesh/ops/SmallPathAtlasMgr.h" |
| #include "src/gpu/ganesh/surface/SkSurface_Ganesh.h" |
| #include "src/gpu/ganesh/text/GrAtlasManager.h" |
| #include "src/image/SkImage_Base.h" |
| #include "src/image/SkSurface_Base.h" |
| #include "src/text/gpu/StrikeCache.h" |
| #include "src/text/gpu/TextBlobRedrawCoordinator.h" |
| |
| #include <array> |
| #include <atomic> |
| #include <forward_list> |
| #include <memory> |
| #include <utility> |
| |
| #ifdef SK_DIRECT3D |
| #include "src/gpu/ganesh/d3d/GrD3DGpu.h" |
| #endif |
| |
| using namespace skia_private; |
| |
| #define ASSERT_SINGLE_OWNER SKGPU_ASSERT_SINGLE_OWNER(this->singleOwner()) |
| |
| using StrikeCache = sktext::gpu::StrikeCache; |
| |
| GrDirectContext::DirectContextID GrDirectContext::DirectContextID::Next() { |
| static std::atomic<uint32_t> nextID{1}; |
| uint32_t id; |
| do { |
| id = nextID.fetch_add(1, std::memory_order_relaxed); |
| } while (id == SK_InvalidUniqueID); |
| return DirectContextID(id); |
| } |
| |
| GrDirectContext::GrDirectContext(GrBackendApi backend, |
| const GrContextOptions& options, |
| sk_sp<GrContextThreadSafeProxy> proxy) |
| : GrRecordingContext(std::move(proxy), false) |
| , fDeleteCallbackHelper(new DeleteCallbackHelper(options.fContextDeleteContext, |
| options.fContextDeleteProc)) |
| , fDirectContextID(DirectContextID::Next()) {} |
| |
| GrDirectContext::~GrDirectContext() { |
| ASSERT_SINGLE_OWNER |
| // this if-test protects against the case where the context is being destroyed |
| // before having been fully created |
| if (fGpu) { |
| this->flushAndSubmit(); |
| } |
| |
| // We need to make sure all work is finished on the gpu before we start releasing resources. |
| this->syncAllOutstandingGpuWork(/*shouldExecuteWhileAbandoned=*/false); |
| |
| this->destroyDrawingManager(); |
| |
| // Ideally we could just let the ptr drop, but resource cache queries this ptr in releaseAll. |
| if (fResourceCache) { |
| fResourceCache->releaseAll(); |
| } |
| // This has to be after GrResourceCache::releaseAll so that other threads that are holding |
| // async pixel result don't try to destroy buffers off thread. |
| fMappedBufferManager.reset(); |
| } |
| |
| sk_sp<GrContextThreadSafeProxy> GrDirectContext::threadSafeProxy() { |
| return GrRecordingContext::threadSafeProxy(); |
| } |
| |
| void GrDirectContext::resetGLTextureBindings() { |
| if (this->abandoned() || this->backend() != GrBackendApi::kOpenGL) { |
| return; |
| } |
| fGpu->resetTextureBindings(); |
| } |
| |
| void GrDirectContext::resetContext(uint32_t state) { |
| ASSERT_SINGLE_OWNER |
| fGpu->markContextDirty(state); |
| } |
| |
| void GrDirectContext::abandonContext() { |
| if (GrRecordingContext::abandoned()) { |
| return; |
| } |
| |
| if (fInsideReleaseProcCnt) { |
| SkDEBUGFAIL("Calling GrDirectContext::abandonContext() while inside a ReleaseProc is not " |
| "allowed"); |
| return; |
| } |
| |
| GrRecordingContext::abandonContext(); |
| |
| // We need to make sure all work is finished on the gpu before we start releasing resources. |
| this->syncAllOutstandingGpuWork(this->caps()->mustSyncGpuDuringAbandon()); |
| |
| fStrikeCache->freeAll(); |
| |
| fMappedBufferManager->abandon(); |
| |
| fResourceProvider->abandon(); |
| |
| // abandon first so destructors don't try to free the resources in the API. |
| fResourceCache->abandonAll(); |
| |
| fGpu->disconnect(GrGpu::DisconnectType::kAbandon); |
| |
| #if !defined(SK_ENABLE_OPTIMIZE_SIZE) |
| if (fSmallPathAtlasMgr) { |
| fSmallPathAtlasMgr->reset(); |
| } |
| #endif |
| fAtlasManager->freeAll(); |
| } |
| |
| bool GrDirectContext::abandoned() { |
| if (GrRecordingContext::abandoned()) { |
| return true; |
| } |
| |
| if (fGpu && fGpu->isDeviceLost()) { |
| this->abandonContext(); |
| return true; |
| } |
| return false; |
| } |
| |
| bool GrDirectContext::isDeviceLost() { |
| if (fGpu && fGpu->isDeviceLost()) { |
| if (!GrRecordingContext::abandoned()) { |
| this->abandonContext(); |
| } |
| return true; |
| } |
| return false; |
| } |
| |
| bool GrDirectContext::oomed() { return fGpu ? fGpu->checkAndResetOOMed() : false; } |
| |
| void GrDirectContext::releaseResourcesAndAbandonContext() { |
| if (GrRecordingContext::abandoned()) { |
| return; |
| } |
| |
| GrRecordingContext::abandonContext(); |
| |
| // We need to make sure all work is finished on the gpu before we start releasing resources. |
| this->syncAllOutstandingGpuWork(/*shouldExecuteWhileAbandoned=*/true); |
| |
| fResourceProvider->abandon(); |
| |
| // Release all resources in the backend 3D API. |
| fResourceCache->releaseAll(); |
| |
| // Must be after GrResourceCache::releaseAll(). |
| fMappedBufferManager.reset(); |
| |
| fGpu->disconnect(GrGpu::DisconnectType::kCleanup); |
| #if !defined(SK_ENABLE_OPTIMIZE_SIZE) |
| if (fSmallPathAtlasMgr) { |
| fSmallPathAtlasMgr->reset(); |
| } |
| #endif |
| fAtlasManager->freeAll(); |
| } |
| |
| void GrDirectContext::freeGpuResources() { |
| ASSERT_SINGLE_OWNER |
| |
| if (this->abandoned()) { |
| return; |
| } |
| |
| this->flushAndSubmit(); |
| #if !defined(SK_ENABLE_OPTIMIZE_SIZE) |
| if (fSmallPathAtlasMgr) { |
| fSmallPathAtlasMgr->reset(); |
| } |
| #endif |
| fAtlasManager->freeAll(); |
| |
| // TODO: the glyph cache doesn't hold any GpuResources so this call should not be needed here. |
| // Some slack in the GrTextBlob's implementation requires it though. That could be fixed. |
| fStrikeCache->freeAll(); |
| |
| this->drawingManager()->freeGpuResources(); |
| |
| fResourceCache->purgeUnlockedResources(GrPurgeResourceOptions::kAllResources); |
| } |
| |
| bool GrDirectContext::init() { |
| ASSERT_SINGLE_OWNER |
| if (!fGpu) { |
| return false; |
| } |
| |
| fThreadSafeProxy->priv().init(fGpu->refCaps(), fGpu->refPipelineBuilder()); |
| if (!GrRecordingContext::init()) { |
| return false; |
| } |
| |
| SkASSERT(this->getTextBlobRedrawCoordinator()); |
| SkASSERT(this->threadSafeCache()); |
| |
| fStrikeCache = std::make_unique<StrikeCache>(); |
| fResourceCache = std::make_unique<GrResourceCache>(this->singleOwner(), |
| this->directContextID(), |
| this->contextID()); |
| fResourceCache->setProxyProvider(this->proxyProvider()); |
| fResourceCache->setThreadSafeCache(this->threadSafeCache()); |
| #if defined(GR_TEST_UTILS) |
| if (this->options().fResourceCacheLimitOverride != -1) { |
| this->setResourceCacheLimit(this->options().fResourceCacheLimitOverride); |
| } |
| #endif |
| fResourceProvider = std::make_unique<GrResourceProvider>(fGpu.get(), fResourceCache.get(), |
| this->singleOwner()); |
| fMappedBufferManager = std::make_unique<GrClientMappedBufferManager>(this->directContextID()); |
| |
| fDidTestPMConversions = false; |
| |
| // DDL TODO: we need to think through how the task group & persistent cache |
| // get passed on to/shared between all the DDLRecorders created with this context. |
| if (this->options().fExecutor) { |
| fTaskGroup = std::make_unique<SkTaskGroup>(*this->options().fExecutor); |
| } |
| |
| fPersistentCache = this->options().fPersistentCache; |
| |
| GrDrawOpAtlas::AllowMultitexturing allowMultitexturing; |
| if (GrContextOptions::Enable::kNo == this->options().fAllowMultipleGlyphCacheTextures || |
| // multitexturing supported only if range can represent the index + texcoords fully |
| !(this->caps()->shaderCaps()->fFloatIs32Bits || |
| this->caps()->shaderCaps()->fIntegerSupport)) { |
| allowMultitexturing = GrDrawOpAtlas::AllowMultitexturing::kNo; |
| } else { |
| allowMultitexturing = GrDrawOpAtlas::AllowMultitexturing::kYes; |
| } |
| |
| GrProxyProvider* proxyProvider = this->priv().proxyProvider(); |
| |
| fAtlasManager = std::make_unique<GrAtlasManager>(proxyProvider, |
| this->options().fGlyphCacheTextureMaximumBytes, |
| allowMultitexturing, |
| this->options().fSupportBilerpFromGlyphAtlas); |
| this->priv().addOnFlushCallbackObject(fAtlasManager.get()); |
| |
| return true; |
| } |
| |
| void GrDirectContext::getResourceCacheUsage(int* resourceCount, size_t* resourceBytes) const { |
| ASSERT_SINGLE_OWNER |
| |
| if (resourceCount) { |
| *resourceCount = fResourceCache->getBudgetedResourceCount(); |
| } |
| if (resourceBytes) { |
| *resourceBytes = fResourceCache->getBudgetedResourceBytes(); |
| } |
| } |
| |
| size_t GrDirectContext::getResourceCachePurgeableBytes() const { |
| ASSERT_SINGLE_OWNER |
| return fResourceCache->getPurgeableBytes(); |
| } |
| |
| void GrDirectContext::getResourceCacheLimits(int* maxResources, size_t* maxResourceBytes) const { |
| ASSERT_SINGLE_OWNER |
| if (maxResources) { |
| *maxResources = -1; |
| } |
| if (maxResourceBytes) { |
| *maxResourceBytes = this->getResourceCacheLimit(); |
| } |
| } |
| |
| size_t GrDirectContext::getResourceCacheLimit() const { |
| ASSERT_SINGLE_OWNER |
| return fResourceCache->getMaxResourceBytes(); |
| } |
| |
| void GrDirectContext::setResourceCacheLimits(int unused, size_t maxResourceBytes) { |
| ASSERT_SINGLE_OWNER |
| this->setResourceCacheLimit(maxResourceBytes); |
| } |
| |
| void GrDirectContext::setResourceCacheLimit(size_t maxResourceBytes) { |
| ASSERT_SINGLE_OWNER |
| fResourceCache->setLimit(maxResourceBytes); |
| } |
| |
| void GrDirectContext::purgeUnlockedResources(GrPurgeResourceOptions opts) { |
| ASSERT_SINGLE_OWNER |
| |
| if (this->abandoned()) { |
| return; |
| } |
| |
| fResourceCache->purgeUnlockedResources(opts); |
| fResourceCache->purgeAsNeeded(); |
| |
| // The textBlob Cache doesn't actually hold any GPU resource but this is a convenient |
| // place to purge stale blobs |
| this->getTextBlobRedrawCoordinator()->purgeStaleBlobs(); |
| |
| fGpu->releaseUnlockedBackendObjects(); |
| } |
| |
| void GrDirectContext::performDeferredCleanup(std::chrono::milliseconds msNotUsed, |
| GrPurgeResourceOptions opts) { |
| TRACE_EVENT0("skia.gpu", TRACE_FUNC); |
| |
| ASSERT_SINGLE_OWNER |
| |
| if (this->abandoned()) { |
| return; |
| } |
| |
| this->checkAsyncWorkCompletion(); |
| fMappedBufferManager->process(); |
| auto purgeTime = skgpu::StdSteadyClock::now() - msNotUsed; |
| |
| fResourceCache->purgeAsNeeded(); |
| fResourceCache->purgeResourcesNotUsedSince(purgeTime, opts); |
| |
| // The textBlob Cache doesn't actually hold any GPU resource but this is a convenient |
| // place to purge stale blobs |
| this->getTextBlobRedrawCoordinator()->purgeStaleBlobs(); |
| } |
| |
| void GrDirectContext::purgeUnlockedResources(size_t bytesToPurge, bool preferScratchResources) { |
| ASSERT_SINGLE_OWNER |
| |
| if (this->abandoned()) { |
| return; |
| } |
| |
| fResourceCache->purgeUnlockedResources(bytesToPurge, preferScratchResources); |
| } |
| |
| //////////////////////////////////////////////////////////////////////////////// |
| bool GrDirectContext::wait(int numSemaphores, const GrBackendSemaphore waitSemaphores[], |
| bool deleteSemaphoresAfterWait) { |
| if (!fGpu || !fGpu->caps()->backendSemaphoreSupport()) { |
| return false; |
| } |
| GrWrapOwnership ownership = |
| deleteSemaphoresAfterWait ? kAdopt_GrWrapOwnership : kBorrow_GrWrapOwnership; |
| for (int i = 0; i < numSemaphores; ++i) { |
| std::unique_ptr<GrSemaphore> sema = fResourceProvider->wrapBackendSemaphore( |
| waitSemaphores[i], GrSemaphoreWrapType::kWillWait, ownership); |
| // If we failed to wrap the semaphore it means the client didn't give us a valid semaphore |
| // to begin with. Therefore, it is fine to not wait on it. |
| if (sema) { |
| fGpu->waitSemaphore(sema.get()); |
| } |
| } |
| return true; |
| } |
| |
| #if !defined(SK_ENABLE_OPTIMIZE_SIZE) |
| skgpu::ganesh::SmallPathAtlasMgr* GrDirectContext::onGetSmallPathAtlasMgr() { |
| if (!fSmallPathAtlasMgr) { |
| fSmallPathAtlasMgr = std::make_unique<skgpu::ganesh::SmallPathAtlasMgr>(); |
| |
| this->priv().addOnFlushCallbackObject(fSmallPathAtlasMgr.get()); |
| } |
| |
| if (!fSmallPathAtlasMgr->initAtlas(this->proxyProvider(), this->caps())) { |
| return nullptr; |
| } |
| |
| return fSmallPathAtlasMgr.get(); |
| } |
| #endif |
| |
| //////////////////////////////////////////////////////////////////////////////// |
| |
| GrSemaphoresSubmitted GrDirectContext::flush(const GrFlushInfo& info) { |
| ASSERT_SINGLE_OWNER |
| if (this->abandoned()) { |
| if (info.fFinishedProc) { |
| info.fFinishedProc(info.fFinishedContext); |
| } |
| if (info.fSubmittedProc) { |
| info.fSubmittedProc(info.fSubmittedContext, false); |
| } |
| return GrSemaphoresSubmitted::kNo; |
| } |
| |
| return this->drawingManager()->flushSurfaces( |
| {}, SkSurfaces::BackendSurfaceAccess::kNoAccess, info, nullptr); |
| } |
| |
| bool GrDirectContext::submit(GrSyncCpu sync) { |
| ASSERT_SINGLE_OWNER |
| if (this->abandoned()) { |
| return false; |
| } |
| |
| if (!fGpu) { |
| return false; |
| } |
| |
| return fGpu->submitToGpu(sync); |
| } |
| |
| GrSemaphoresSubmitted GrDirectContext::flush(const sk_sp<const SkImage>& image, |
| const GrFlushInfo& flushInfo) { |
| if (!image) { |
| return GrSemaphoresSubmitted::kNo; |
| } |
| auto ib = as_IB(image); |
| if (!ib->isGaneshBacked()) { |
| return GrSemaphoresSubmitted::kNo; |
| } |
| auto igb = static_cast<const SkImage_GaneshBase*>(image.get()); |
| return igb->flush(this, flushInfo); |
| } |
| |
| void GrDirectContext::flush(const sk_sp<const SkImage>& image) { |
| this->flush(image, {}); |
| } |
| |
| void GrDirectContext::flushAndSubmit(const sk_sp<const SkImage>& image) { |
| this->flush(image, {}); |
| this->submit(); |
| } |
| |
| GrSemaphoresSubmitted GrDirectContext::flush(SkSurface* surface, |
| SkSurfaces::BackendSurfaceAccess access, |
| const GrFlushInfo& info) { |
| if (!surface) { |
| return GrSemaphoresSubmitted::kNo; |
| } |
| auto sb = asSB(surface); |
| if (!sb->isGaneshBacked()) { |
| return GrSemaphoresSubmitted::kNo; |
| } |
| |
| auto gs = static_cast<SkSurface_Ganesh*>(surface); |
| SkASSERT(this->priv().matches(gs->getDevice()->recordingContext()->asDirectContext())); |
| GrRenderTargetProxy* rtp = gs->getDevice()->targetProxy(); |
| |
| return this->priv().flushSurface(rtp, access, info, nullptr); |
| } |
| |
| GrSemaphoresSubmitted GrDirectContext::flush(SkSurface* surface, |
| const GrFlushInfo& info, |
| const skgpu::MutableTextureState* newState) { |
| if (!surface) { |
| return GrSemaphoresSubmitted::kNo; |
| } |
| auto sb = asSB(surface); |
| if (!sb->isGaneshBacked()) { |
| return GrSemaphoresSubmitted::kNo; |
| } |
| |
| auto gs = static_cast<SkSurface_Ganesh*>(surface); |
| SkASSERT(this->priv().matches(gs->getDevice()->recordingContext()->asDirectContext())); |
| GrRenderTargetProxy* rtp = gs->getDevice()->targetProxy(); |
| |
| return this->priv().flushSurface( |
| rtp, SkSurfaces::BackendSurfaceAccess::kNoAccess, info, newState); |
| } |
| |
| void GrDirectContext::flushAndSubmit(SkSurface* surface, GrSyncCpu sync) { |
| this->flush(surface, SkSurfaces::BackendSurfaceAccess::kNoAccess, GrFlushInfo()); |
| this->submit(sync); |
| } |
| |
| void GrDirectContext::flush(SkSurface* surface) { |
| this->flush(surface, GrFlushInfo(), nullptr); |
| } |
| |
| //////////////////////////////////////////////////////////////////////////////// |
| |
| void GrDirectContext::checkAsyncWorkCompletion() { |
| if (fGpu) { |
| fGpu->checkFinishProcs(); |
| } |
| } |
| |
| void GrDirectContext::syncAllOutstandingGpuWork(bool shouldExecuteWhileAbandoned) { |
| if (fGpu && (!this->abandoned() || shouldExecuteWhileAbandoned)) { |
| fGpu->finishOutstandingGpuWork(); |
| this->checkAsyncWorkCompletion(); |
| } |
| } |
| |
| //////////////////////////////////////////////////////////////////////////////// |
| |
| void GrDirectContext::storeVkPipelineCacheData() { |
| if (fGpu) { |
| fGpu->storeVkPipelineCacheData(); |
| } |
| } |
| |
| //////////////////////////////////////////////////////////////////////////////// |
| |
| bool GrDirectContext::supportsDistanceFieldText() const { |
| return this->caps()->shaderCaps()->supportsDistanceFieldText(); |
| } |
| |
| ////////////////////////////////////////////////////////////////////////////// |
| |
| void GrDirectContext::dumpMemoryStatistics(SkTraceMemoryDump* traceMemoryDump) const { |
| ASSERT_SINGLE_OWNER |
| fResourceCache->dumpMemoryStatistics(traceMemoryDump); |
| traceMemoryDump->dumpNumericValue("skia/gr_text_blob_cache", "size", "bytes", |
| this->getTextBlobRedrawCoordinator()->usedBytes()); |
| } |
| |
| GrBackendTexture GrDirectContext::createBackendTexture(int width, |
| int height, |
| const GrBackendFormat& backendFormat, |
| skgpu::Mipmapped mipmapped, |
| GrRenderable renderable, |
| GrProtected isProtected, |
| std::string_view label) { |
| TRACE_EVENT0("skia.gpu", TRACE_FUNC); |
| if (this->abandoned()) { |
| return GrBackendTexture(); |
| } |
| |
| return fGpu->createBackendTexture({width, height}, backendFormat, renderable, |
| mipmapped, isProtected, label); |
| } |
| |
| GrBackendTexture GrDirectContext::createBackendTexture(const SkPixmap& srcData, |
| GrSurfaceOrigin textureOrigin, |
| GrRenderable renderable, |
| GrProtected isProtected, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext, |
| std::string_view label) { |
| return this->createBackendTexture(&srcData, 1, textureOrigin, renderable, isProtected, |
| finishedProc, finishedContext, label); |
| } |
| |
| GrBackendTexture GrDirectContext::createBackendTexture(const SkPixmap& srcData, |
| GrRenderable renderable, |
| GrProtected isProtected, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext, |
| std::string_view label) { |
| return this->createBackendTexture(&srcData, |
| 1, |
| renderable, |
| isProtected, |
| finishedProc, |
| finishedContext, |
| label); |
| } |
| |
| GrBackendTexture GrDirectContext::createBackendTexture(const SkPixmap srcData[], |
| int numLevels, |
| GrRenderable renderable, |
| GrProtected isProtected, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext, |
| std::string_view label) { |
| return this->createBackendTexture(srcData, |
| numLevels, |
| kTopLeft_GrSurfaceOrigin, |
| renderable, |
| isProtected, |
| finishedProc, |
| finishedContext, |
| label); |
| } |
| |
| GrBackendTexture GrDirectContext::createBackendTexture(int width, |
| int height, |
| SkColorType skColorType, |
| skgpu::Mipmapped mipmapped, |
| GrRenderable renderable, |
| GrProtected isProtected, |
| std::string_view label) { |
| if (this->abandoned()) { |
| return GrBackendTexture(); |
| } |
| |
| const GrBackendFormat format = this->defaultBackendFormat(skColorType, renderable); |
| |
| return this->createBackendTexture( |
| width, height, format, mipmapped, renderable, isProtected, label); |
| } |
| |
| static GrBackendTexture create_and_clear_backend_texture( |
| GrDirectContext* dContext, |
| SkISize dimensions, |
| const GrBackendFormat& backendFormat, |
| skgpu::Mipmapped mipmapped, |
| GrRenderable renderable, |
| GrProtected isProtected, |
| sk_sp<skgpu::RefCntedCallback> finishedCallback, |
| std::array<float, 4> color, |
| std::string_view label) { |
| GrGpu* gpu = dContext->priv().getGpu(); |
| GrBackendTexture beTex = gpu->createBackendTexture(dimensions, backendFormat, renderable, |
| mipmapped, isProtected, label); |
| if (!beTex.isValid()) { |
| return {}; |
| } |
| |
| if (!dContext->priv().getGpu()->clearBackendTexture(beTex, |
| std::move(finishedCallback), |
| color)) { |
| dContext->deleteBackendTexture(beTex); |
| return {}; |
| } |
| return beTex; |
| } |
| |
| static bool update_texture_with_pixmaps(GrDirectContext* context, |
| const SkPixmap src[], |
| int numLevels, |
| const GrBackendTexture& backendTexture, |
| GrSurfaceOrigin textureOrigin, |
| sk_sp<skgpu::RefCntedCallback> finishedCallback) { |
| GrColorType ct = SkColorTypeToGrColorType(src[0].colorType()); |
| const GrBackendFormat& format = backendTexture.getBackendFormat(); |
| |
| if (!context->priv().caps()->areColorTypeAndFormatCompatible(ct, format)) { |
| return false; |
| } |
| |
| auto proxy = context->priv().proxyProvider()->wrapBackendTexture(backendTexture, |
| kBorrow_GrWrapOwnership, |
| GrWrapCacheable::kNo, |
| kRW_GrIOType, |
| std::move(finishedCallback)); |
| if (!proxy) { |
| return false; |
| } |
| |
| skgpu::Swizzle swizzle = context->priv().caps()->getReadSwizzle(format, ct); |
| GrSurfaceProxyView view(std::move(proxy), textureOrigin, swizzle); |
| skgpu::ganesh::SurfaceContext surfaceContext( |
| context, std::move(view), src[0].info().colorInfo()); |
| AutoSTArray<15, GrCPixmap> tmpSrc(numLevels); |
| for (int i = 0; i < numLevels; ++i) { |
| tmpSrc[i] = src[i]; |
| } |
| if (!surfaceContext.writePixels(context, tmpSrc.get(), numLevels)) { |
| return false; |
| } |
| |
| GrSurfaceProxy* p = surfaceContext.asSurfaceProxy(); |
| GrFlushInfo info; |
| context->priv().drawingManager()->flushSurfaces( |
| {&p, 1}, SkSurfaces::BackendSurfaceAccess::kNoAccess, info, nullptr); |
| return true; |
| } |
| |
| GrBackendTexture GrDirectContext::createBackendTexture(int width, |
| int height, |
| const GrBackendFormat& backendFormat, |
| const SkColor4f& color, |
| skgpu::Mipmapped mipmapped, |
| GrRenderable renderable, |
| GrProtected isProtected, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext, |
| std::string_view label) { |
| auto finishedCallback = skgpu::RefCntedCallback::Make(finishedProc, finishedContext); |
| |
| TRACE_EVENT0("skia.gpu", TRACE_FUNC); |
| if (this->abandoned()) { |
| return {}; |
| } |
| |
| return create_and_clear_backend_texture(this, |
| {width, height}, |
| backendFormat, |
| mipmapped, |
| renderable, |
| isProtected, |
| std::move(finishedCallback), |
| color.array(), |
| label); |
| } |
| |
| GrBackendTexture GrDirectContext::createBackendTexture(int width, |
| int height, |
| SkColorType skColorType, |
| const SkColor4f& color, |
| skgpu::Mipmapped mipmapped, |
| GrRenderable renderable, |
| GrProtected isProtected, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext, |
| std::string_view label) { |
| auto finishedCallback = skgpu::RefCntedCallback::Make(finishedProc, finishedContext); |
| |
| if (this->abandoned()) { |
| return {}; |
| } |
| |
| GrBackendFormat format = this->defaultBackendFormat(skColorType, renderable); |
| if (!format.isValid()) { |
| return {}; |
| } |
| |
| GrColorType grColorType = SkColorTypeToGrColorType(skColorType); |
| SkColor4f swizzledColor = this->caps()->getWriteSwizzle(format, grColorType).applyTo(color); |
| |
| return create_and_clear_backend_texture(this, |
| {width, height}, |
| format, |
| mipmapped, |
| renderable, |
| isProtected, |
| std::move(finishedCallback), |
| swizzledColor.array(), |
| label); |
| } |
| |
| GrBackendTexture GrDirectContext::createBackendTexture(const SkPixmap srcData[], |
| int numProvidedLevels, |
| GrSurfaceOrigin textureOrigin, |
| GrRenderable renderable, |
| GrProtected isProtected, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext, |
| std::string_view label) { |
| TRACE_EVENT0("skia.gpu", TRACE_FUNC); |
| |
| auto finishedCallback = skgpu::RefCntedCallback::Make(finishedProc, finishedContext); |
| |
| if (this->abandoned()) { |
| return {}; |
| } |
| |
| if (!srcData || numProvidedLevels <= 0) { |
| return {}; |
| } |
| |
| SkColorType colorType = srcData[0].colorType(); |
| |
| skgpu::Mipmapped mipmapped = skgpu::Mipmapped::kNo; |
| if (numProvidedLevels > 1) { |
| mipmapped = skgpu::Mipmapped::kYes; |
| } |
| |
| GrBackendFormat backendFormat = this->defaultBackendFormat(colorType, renderable); |
| GrBackendTexture beTex = this->createBackendTexture(srcData[0].width(), |
| srcData[0].height(), |
| backendFormat, |
| mipmapped, |
| renderable, |
| isProtected, |
| label); |
| if (!beTex.isValid()) { |
| return {}; |
| } |
| if (!update_texture_with_pixmaps(this, |
| srcData, |
| numProvidedLevels, |
| beTex, |
| textureOrigin, |
| std::move(finishedCallback))) { |
| this->deleteBackendTexture(beTex); |
| return {}; |
| } |
| return beTex; |
| } |
| |
| bool GrDirectContext::updateBackendTexture(const GrBackendTexture& texture, |
| const SkPixmap srcData[], |
| int numLevels, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext) { |
| return this->updateBackendTexture(texture, |
| srcData, |
| numLevels, |
| kTopLeft_GrSurfaceOrigin, |
| finishedProc, |
| finishedContext); |
| } |
| |
| bool GrDirectContext::updateBackendTexture(const GrBackendTexture& backendTexture, |
| const SkColor4f& color, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext) { |
| auto finishedCallback = skgpu::RefCntedCallback::Make(finishedProc, finishedContext); |
| |
| if (this->abandoned()) { |
| return false; |
| } |
| |
| return fGpu->clearBackendTexture(backendTexture, std::move(finishedCallback), color.array()); |
| } |
| |
| bool GrDirectContext::updateBackendTexture(const GrBackendTexture& backendTexture, |
| SkColorType skColorType, |
| const SkColor4f& color, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext) { |
| auto finishedCallback = skgpu::RefCntedCallback::Make(finishedProc, finishedContext); |
| |
| if (this->abandoned()) { |
| return false; |
| } |
| |
| GrBackendFormat format = backendTexture.getBackendFormat(); |
| GrColorType grColorType = SkColorTypeToGrColorType(skColorType); |
| |
| if (!this->caps()->areColorTypeAndFormatCompatible(grColorType, format)) { |
| return false; |
| } |
| |
| skgpu::Swizzle swizzle = this->caps()->getWriteSwizzle(format, grColorType); |
| SkColor4f swizzledColor = swizzle.applyTo(color); |
| |
| return fGpu->clearBackendTexture(backendTexture, |
| std::move(finishedCallback), |
| swizzledColor.array()); |
| } |
| |
| bool GrDirectContext::updateBackendTexture(const GrBackendTexture& backendTexture, |
| const SkPixmap srcData[], |
| int numLevels, |
| GrSurfaceOrigin textureOrigin, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext) { |
| auto finishedCallback = skgpu::RefCntedCallback::Make(finishedProc, finishedContext); |
| |
| if (this->abandoned()) { |
| return false; |
| } |
| |
| if (!srcData || numLevels <= 0) { |
| return false; |
| } |
| |
| // If the texture has MIP levels then we require that the full set is overwritten. |
| int numExpectedLevels = 1; |
| if (backendTexture.hasMipmaps()) { |
| numExpectedLevels = SkMipmap::ComputeLevelCount(backendTexture.width(), |
| backendTexture.height()) + 1; |
| } |
| if (numLevels != numExpectedLevels) { |
| return false; |
| } |
| return update_texture_with_pixmaps(this, |
| srcData, |
| numLevels, |
| backendTexture, |
| textureOrigin, |
| std::move(finishedCallback)); |
| } |
| |
| ////////////////////////////////////////////////////////////////////////////// |
| |
| static GrBackendTexture create_and_update_compressed_backend_texture( |
| GrDirectContext* dContext, |
| SkISize dimensions, |
| const GrBackendFormat& backendFormat, |
| skgpu::Mipmapped mipmapped, |
| GrProtected isProtected, |
| sk_sp<skgpu::RefCntedCallback> finishedCallback, |
| const void* data, |
| size_t size) { |
| GrGpu* gpu = dContext->priv().getGpu(); |
| |
| GrBackendTexture beTex = gpu->createCompressedBackendTexture(dimensions, backendFormat, |
| mipmapped, isProtected); |
| if (!beTex.isValid()) { |
| return {}; |
| } |
| |
| if (!dContext->priv().getGpu()->updateCompressedBackendTexture( |
| beTex, std::move(finishedCallback), data, size)) { |
| dContext->deleteBackendTexture(beTex); |
| return {}; |
| } |
| return beTex; |
| } |
| |
| GrBackendTexture GrDirectContext::createCompressedBackendTexture( |
| int width, |
| int height, |
| const GrBackendFormat& backendFormat, |
| const SkColor4f& color, |
| skgpu::Mipmapped mipmapped, |
| GrProtected isProtected, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext) { |
| TRACE_EVENT0("skia.gpu", TRACE_FUNC); |
| auto finishedCallback = skgpu::RefCntedCallback::Make(finishedProc, finishedContext); |
| |
| if (this->abandoned()) { |
| return {}; |
| } |
| |
| SkTextureCompressionType compression = GrBackendFormatToCompressionType(backendFormat); |
| if (compression == SkTextureCompressionType::kNone) { |
| return {}; |
| } |
| |
| size_t size = SkCompressedDataSize( |
| compression, {width, height}, nullptr, mipmapped == skgpu::Mipmapped::kYes); |
| auto storage = std::make_unique<char[]>(size); |
| skgpu::FillInCompressedData(compression, {width, height}, mipmapped, storage.get(), color); |
| return create_and_update_compressed_backend_texture(this, |
| {width, height}, |
| backendFormat, |
| mipmapped, |
| isProtected, |
| std::move(finishedCallback), |
| storage.get(), |
| size); |
| } |
| |
| GrBackendTexture GrDirectContext::createCompressedBackendTexture( |
| int width, |
| int height, |
| SkTextureCompressionType compression, |
| const SkColor4f& color, |
| skgpu::Mipmapped mipmapped, |
| GrProtected isProtected, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext) { |
| TRACE_EVENT0("skia.gpu", TRACE_FUNC); |
| GrBackendFormat format = this->compressedBackendFormat(compression); |
| return this->createCompressedBackendTexture(width, height, format, color, |
| mipmapped, isProtected, finishedProc, |
| finishedContext); |
| } |
| |
| GrBackendTexture GrDirectContext::createCompressedBackendTexture( |
| int width, |
| int height, |
| const GrBackendFormat& backendFormat, |
| const void* compressedData, |
| size_t dataSize, |
| skgpu::Mipmapped mipmapped, |
| GrProtected isProtected, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext) { |
| TRACE_EVENT0("skia.gpu", TRACE_FUNC); |
| auto finishedCallback = skgpu::RefCntedCallback::Make(finishedProc, finishedContext); |
| |
| if (this->abandoned()) { |
| return {}; |
| } |
| |
| return create_and_update_compressed_backend_texture(this, |
| {width, height}, |
| backendFormat, |
| mipmapped, |
| isProtected, |
| std::move(finishedCallback), |
| compressedData, |
| dataSize); |
| } |
| |
| GrBackendTexture GrDirectContext::createCompressedBackendTexture( |
| int width, |
| int height, |
| SkTextureCompressionType compression, |
| const void* data, |
| size_t dataSize, |
| skgpu::Mipmapped mipmapped, |
| GrProtected isProtected, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext) { |
| TRACE_EVENT0("skia.gpu", TRACE_FUNC); |
| GrBackendFormat format = this->compressedBackendFormat(compression); |
| return this->createCompressedBackendTexture(width, height, format, data, dataSize, mipmapped, |
| isProtected, finishedProc, finishedContext); |
| } |
| |
| bool GrDirectContext::updateCompressedBackendTexture(const GrBackendTexture& backendTexture, |
| const SkColor4f& color, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext) { |
| auto finishedCallback = skgpu::RefCntedCallback::Make(finishedProc, finishedContext); |
| |
| if (this->abandoned()) { |
| return false; |
| } |
| |
| SkTextureCompressionType compression = |
| GrBackendFormatToCompressionType(backendTexture.getBackendFormat()); |
| if (compression == SkTextureCompressionType::kNone) { |
| return {}; |
| } |
| size_t size = SkCompressedDataSize(compression, |
| backendTexture.dimensions(), |
| nullptr, |
| backendTexture.hasMipmaps()); |
| SkAutoMalloc storage(size); |
| skgpu::FillInCompressedData(compression, |
| backendTexture.dimensions(), |
| backendTexture.mipmapped(), |
| static_cast<char*>(storage.get()), |
| color); |
| return fGpu->updateCompressedBackendTexture(backendTexture, |
| std::move(finishedCallback), |
| storage.get(), |
| size); |
| } |
| |
| bool GrDirectContext::updateCompressedBackendTexture(const GrBackendTexture& backendTexture, |
| const void* compressedData, |
| size_t dataSize, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext) { |
| auto finishedCallback = skgpu::RefCntedCallback::Make(finishedProc, finishedContext); |
| |
| if (this->abandoned()) { |
| return false; |
| } |
| |
| if (!compressedData) { |
| return false; |
| } |
| |
| return fGpu->updateCompressedBackendTexture(backendTexture, |
| std::move(finishedCallback), |
| compressedData, |
| dataSize); |
| } |
| |
| ////////////////////////////////////////////////////////////////////////////// |
| |
| bool GrDirectContext::setBackendTextureState(const GrBackendTexture& backendTexture, |
| const skgpu::MutableTextureState& state, |
| skgpu::MutableTextureState* previousState, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext) { |
| auto callback = skgpu::RefCntedCallback::Make(finishedProc, finishedContext); |
| |
| if (this->abandoned()) { |
| return false; |
| } |
| |
| return fGpu->setBackendTextureState(backendTexture, state, previousState, std::move(callback)); |
| } |
| |
| |
| bool GrDirectContext::setBackendRenderTargetState(const GrBackendRenderTarget& backendRenderTarget, |
| const skgpu::MutableTextureState& state, |
| skgpu::MutableTextureState* previousState, |
| GrGpuFinishedProc finishedProc, |
| GrGpuFinishedContext finishedContext) { |
| auto callback = skgpu::RefCntedCallback::Make(finishedProc, finishedContext); |
| |
| if (this->abandoned()) { |
| return false; |
| } |
| |
| return fGpu->setBackendRenderTargetState(backendRenderTarget, state, previousState, |
| std::move(callback)); |
| } |
| |
| void GrDirectContext::deleteBackendTexture(const GrBackendTexture& backendTex) { |
| TRACE_EVENT0("skia.gpu", TRACE_FUNC); |
| // For the Vulkan backend we still must destroy the backend texture when the context is |
| // abandoned. |
| if ((this->abandoned() && this->backend() != GrBackendApi::kVulkan) || !backendTex.isValid()) { |
| return; |
| } |
| |
| fGpu->deleteBackendTexture(backendTex); |
| } |
| |
| ////////////////////////////////////////////////////////////////////////////// |
| |
| bool GrDirectContext::precompileShader(const SkData& key, const SkData& data) { |
| return fGpu->precompileShader(key, data); |
| } |
| |
| #ifdef SK_ENABLE_DUMP_GPU |
| #include "include/core/SkString.h" |
| #include "src/utils/SkJSONWriter.h" |
| SkString GrDirectContext::dump() const { |
| SkDynamicMemoryWStream stream; |
| SkJSONWriter writer(&stream, SkJSONWriter::Mode::kPretty); |
| writer.beginObject(); |
| |
| writer.appendCString("backend", GrBackendApiToStr(this->backend())); |
| |
| writer.appendName("caps"); |
| this->caps()->dumpJSON(&writer); |
| |
| writer.appendName("gpu"); |
| this->fGpu->dumpJSON(&writer); |
| |
| writer.appendName("context"); |
| this->dumpJSON(&writer); |
| |
| // Flush JSON to the memory stream |
| writer.endObject(); |
| writer.flush(); |
| |
| // Null terminate the JSON data in the memory stream |
| stream.write8(0); |
| |
| // Allocate a string big enough to hold all the data, then copy out of the stream |
| SkString result(stream.bytesWritten()); |
| stream.copyToAndReset(result.data()); |
| return result; |
| } |
| #endif |
| |
| /*************************************************************************************************/ |
| sk_sp<GrDirectContext> GrDirectContext::MakeMock(const GrMockOptions* mockOptions) { |
| GrContextOptions defaultOptions; |
| return MakeMock(mockOptions, defaultOptions); |
| } |
| |
| sk_sp<GrDirectContext> GrDirectContext::MakeMock(const GrMockOptions* mockOptions, |
| const GrContextOptions& options) { |
| sk_sp<GrDirectContext> direct( |
| new GrDirectContext(GrBackendApi::kMock, |
| options, |
| GrContextThreadSafeProxyPriv::Make(GrBackendApi::kMock, options))); |
| |
| direct->fGpu = GrMockGpu::Make(mockOptions, options, direct.get()); |
| if (!direct->init()) { |
| return nullptr; |
| } |
| |
| return direct; |
| } |
| |
| #ifdef SK_DIRECT3D |
| /*************************************************************************************************/ |
| sk_sp<GrDirectContext> GrDirectContext::MakeDirect3D(const GrD3DBackendContext& backendContext) { |
| GrContextOptions defaultOptions; |
| return MakeDirect3D(backendContext, defaultOptions); |
| } |
| |
| sk_sp<GrDirectContext> GrDirectContext::MakeDirect3D(const GrD3DBackendContext& backendContext, |
| const GrContextOptions& options) { |
| sk_sp<GrDirectContext> direct(new GrDirectContext( |
| GrBackendApi::kDirect3D, |
| options, |
| GrContextThreadSafeProxyPriv::Make(GrBackendApi::kDirect3D, options))); |
| |
| direct->fGpu = GrD3DGpu::Make(backendContext, options, direct.get()); |
| if (!direct->init()) { |
| return nullptr; |
| } |
| |
| return direct; |
| } |
| #endif |