blob: c6989d3bbd42ab257d94899773043d86aad79e22 [file] [log] [blame]
/*
* Copyright 2021 Google LLC
*
* Use of this source code is governed by a BSD-style license that can be
* found in the LICENSE file.
*/
#include "include/gpu/graphite/Context.h"
#include "include/core/SkPathTypes.h"
#include "include/effects/SkRuntimeEffect.h"
#include "include/gpu/graphite/BackendTexture.h"
#include "include/gpu/graphite/Recorder.h"
#include "include/gpu/graphite/Recording.h"
#include "include/gpu/graphite/TextureInfo.h"
#include "src/gpu/RefCntedCallback.h"
#include "src/gpu/graphite/BufferManager.h"
#include "src/gpu/graphite/Caps.h"
#include "src/gpu/graphite/ClientMappedBufferManager.h"
#include "src/gpu/graphite/CommandBuffer.h"
#include "src/gpu/graphite/ContextPriv.h"
#include "src/gpu/graphite/CopyTask.h"
#include "src/gpu/graphite/DrawAtlas.h"
#include "src/gpu/graphite/GlobalCache.h"
#include "src/gpu/graphite/GraphicsPipeline.h"
#include "src/gpu/graphite/GraphicsPipelineDesc.h"
#include "src/gpu/graphite/Image_Graphite.h"
#include "src/gpu/graphite/KeyContext.h"
#include "src/gpu/graphite/Log.h"
#include "src/gpu/graphite/QueueManager.h"
#include "src/gpu/graphite/RecorderPriv.h"
#include "src/gpu/graphite/RecordingPriv.h"
#include "src/gpu/graphite/Renderer.h"
#include "src/gpu/graphite/RendererProvider.h"
#include "src/gpu/graphite/ResourceProvider.h"
#include "src/gpu/graphite/RuntimeEffectDictionary.h"
#include "src/gpu/graphite/ShaderCodeDictionary.h"
#include "src/gpu/graphite/SharedContext.h"
#include "src/gpu/graphite/Surface_Graphite.h"
#include "src/gpu/graphite/SynchronizeToCpuTask.h"
#include "src/gpu/graphite/TextureProxyView.h"
#include "src/gpu/graphite/UploadTask.h"
namespace skgpu::graphite {
#define ASSERT_SINGLE_OWNER SKGPU_ASSERT_SINGLE_OWNER(this->singleOwner())
Context::ContextID Context::ContextID::Next() {
static std::atomic<uint32_t> nextID{1};
uint32_t id;
do {
id = nextID.fetch_add(1, std::memory_order_relaxed);
} while (id == SK_InvalidUniqueID);
return ContextID(id);
}
//--------------------------------------------------------------------------------------------------
Context::Context(sk_sp<SharedContext> sharedContext,
std::unique_ptr<QueueManager> queueManager,
const ContextOptions& options)
: fSharedContext(std::move(sharedContext))
, fQueueManager(std::move(queueManager))
#if GRAPHITE_TEST_UTILS
, fStoreContextRefInRecorder(options.fStoreContextRefInRecorder)
#endif
, fContextID(ContextID::Next()) {
// We have to create this outside the initializer list because we need to pass in the Context's
// SingleOwner object and it is declared last
fResourceProvider = fSharedContext->makeResourceProvider(&fSingleOwner);
fMappedBufferManager = std::make_unique<ClientMappedBufferManager>(this->contextID());
fPlotUploadTracker = std::make_unique<PlotUploadTracker>();
}
Context::~Context() {
#if GRAPHITE_TEST_UTILS
ASSERT_SINGLE_OWNER
for (auto& recorder : fTrackedRecorders) {
recorder->priv().setContext(nullptr);
}
#endif
}
bool Context::finishInitialization() {
SkASSERT(!fSharedContext->rendererProvider()); // Can only initialize once
StaticBufferManager bufferManager{fResourceProvider.get(), fSharedContext->caps()};
std::unique_ptr<RendererProvider> renderers{
new RendererProvider(fSharedContext->caps(), &bufferManager)};
auto result = bufferManager.finalize(this, fQueueManager.get(), fSharedContext->globalCache());
if (result == StaticBufferManager::FinishResult::kFailure) {
// If something went wrong filling out the static vertex buffers, any Renderer that would
// use it will draw incorrectly, so it's better to fail the Context creation.
return false;
}
if (result == StaticBufferManager::FinishResult::kSuccess &&
!fQueueManager->submitToGpu()) {
SKGPU_LOG_W("Failed to submit initial command buffer for Context creation.\n");
return false;
} // else result was kNoWork so skip submitting to the GPU
fSharedContext->setRendererProvider(std::move(renderers));
return true;
}
BackendApi Context::backend() const { return fSharedContext->backend(); }
std::unique_ptr<Recorder> Context::makeRecorder(const RecorderOptions& options) {
ASSERT_SINGLE_OWNER
auto recorder = std::unique_ptr<Recorder>(new Recorder(fSharedContext, options));
#if GRAPHITE_TEST_UTILS
if (fStoreContextRefInRecorder) {
recorder->priv().setContext(this);
}
#endif
return recorder;
}
bool Context::insertRecording(const InsertRecordingInfo& info) {
ASSERT_SINGLE_OWNER
return fQueueManager->addRecording(info, this);
}
bool Context::submit(SyncToCpu syncToCpu) {
ASSERT_SINGLE_OWNER
bool success = fQueueManager->submitToGpu();
fQueueManager->checkForFinishedWork(syncToCpu);
return success;
}
void Context::asyncReadPixels(const SkImage* image,
const SkColorInfo& dstColorInfo,
const SkIRect& srcRect,
SkImage::ReadPixelsCallback callback,
SkImage::ReadPixelsContext callbackContext) {
if (!as_IB(image)->isGraphiteBacked()) {
callback(callbackContext, nullptr);
return;
}
auto graphiteImage = reinterpret_cast<const skgpu::graphite::Image*>(image);
TextureProxyView proxyView = graphiteImage->textureProxyView();
this->asyncReadPixels(proxyView.proxy(),
image->imageInfo(),
dstColorInfo,
srcRect,
callback,
callbackContext);
}
void Context::asyncReadPixels(const SkSurface* surface,
const SkColorInfo& dstColorInfo,
const SkIRect& srcRect,
SkImage::ReadPixelsCallback callback,
SkImage::ReadPixelsContext callbackContext) {
if (!static_cast<const SkSurface_Base*>(surface)->isGraphiteBacked()) {
callback(callbackContext, nullptr);
return;
}
auto graphiteSurface = reinterpret_cast<const skgpu::graphite::Surface*>(surface);
TextureProxyView proxyView = graphiteSurface->readSurfaceView();
this->asyncReadPixels(proxyView.proxy(),
surface->imageInfo(),
dstColorInfo,
srcRect,
callback,
callbackContext);
}
void Context::asyncReadPixels(const TextureProxy* proxy,
const SkImageInfo& srcImageInfo,
const SkColorInfo& dstColorInfo,
const SkIRect& srcRect,
SkImage::ReadPixelsCallback callback,
SkImage::ReadPixelsContext callbackContext) {
if (!proxy) {
callback(callbackContext, nullptr);
return;
}
if (!SkImageInfoIsValid(srcImageInfo) || !SkColorInfoIsValid(dstColorInfo)) {
callback(callbackContext, nullptr);
return;
}
if (!SkIRect::MakeSize(srcImageInfo.dimensions()).contains(srcRect)) {
callback(callbackContext, nullptr);
return;
}
const Caps* caps = fSharedContext->caps();
if (!caps->supportsReadPixels(proxy->textureInfo())) {
// TODO: try to copy to a readable texture instead
callback(callbackContext, nullptr);
return;
}
PixelTransferResult transferResult = this->transferPixels(proxy, srcImageInfo,
dstColorInfo, srcRect);
if (!transferResult.fTransferBuffer) {
// TODO: try to do a synchronous readPixels instead
callback(callbackContext, nullptr);
return;
}
using AsyncReadResult = skgpu::TAsyncReadResult<Buffer, ContextID, PixelTransferResult>;
struct FinishContext {
SkImage::ReadPixelsCallback* fClientCallback;
SkImage::ReadPixelsContext fClientContext;
SkISize fSize;
size_t fRowBytes;
ClientMappedBufferManager* fMappedBufferManager;
PixelTransferResult fTransferResult;
};
size_t rowBytes = fSharedContext->caps()->getAlignedTextureDataRowBytes(
srcRect.width() * SkColorTypeBytesPerPixel(dstColorInfo.colorType()));
auto* finishContext = new FinishContext{callback,
callbackContext,
srcRect.size(),
rowBytes,
fMappedBufferManager.get(),
std::move(transferResult)};
GpuFinishedProc finishCallback = [](GpuFinishedContext c, CallbackResult status) {
const auto* context = reinterpret_cast<const FinishContext*>(c);
if (status == CallbackResult::kSuccess) {
ClientMappedBufferManager* manager = context->fMappedBufferManager;
auto result = std::make_unique<AsyncReadResult>(manager->ownerID());
if (!result->addTransferResult(context->fTransferResult, context->fSize,
context->fRowBytes, manager)) {
result.reset();
}
(*context->fClientCallback)(context->fClientContext, std::move(result));
} else {
(*context->fClientCallback)(context->fClientContext, nullptr);
}
delete context;
};
InsertFinishInfo info;
info.fFinishedContext = finishContext;
info.fFinishedProc = finishCallback;
// If addFinishInfo() fails, it invokes the finish callback automatically, which handles all the
// required clean up for us, just log an error message.
if (!fQueueManager->addFinishInfo(info, fResourceProvider.get())) {
SKGPU_LOG_E("Failed to register finish callbacks for asyncReadPixels.");
}
}
Context::PixelTransferResult Context::transferPixels(const TextureProxy* proxy,
const SkImageInfo& srcImageInfo,
const SkColorInfo& dstColorInfo,
const SkIRect& srcRect) {
SkASSERT(srcImageInfo.bounds().contains(srcRect));
const Caps* caps = fSharedContext->caps();
SkColorType supportedColorType =
caps->supportedReadPixelsColorType(srcImageInfo.colorType(),
proxy->textureInfo(),
dstColorInfo.colorType());
if (supportedColorType == kUnknown_SkColorType) {
return {};
}
// Fail if read color type does not have all of dstCT's color channels and those missing color
// channels are in the src.
uint32_t dstChannels = SkColorTypeChannelFlags(dstColorInfo.colorType());
uint32_t legalReadChannels = SkColorTypeChannelFlags(supportedColorType);
uint32_t srcChannels = SkColorTypeChannelFlags(srcImageInfo.colorType());
if ((~legalReadChannels & dstChannels) & srcChannels) {
return {};
}
size_t rowBytes = caps->getAlignedTextureDataRowBytes(
SkColorTypeBytesPerPixel(supportedColorType) * srcRect.width());
size_t size = SkAlignTo(rowBytes * srcRect.height(), caps->requiredTransferBufferAlignment());
sk_sp<Buffer> buffer = fResourceProvider->findOrCreateBuffer(
size,
BufferType::kXferCpuToGpu,
PrioritizeGpuReads::kNo);
if (!buffer) {
return {};
}
// Set up copy task. Since we always use a new buffer the offset can be 0 and we don't need to
// worry about aligning it to the required transfer buffer alignment.
sk_sp<CopyTextureToBufferTask> copyTask = CopyTextureToBufferTask::Make(sk_ref_sp(proxy),
srcRect,
buffer,
/*bufferOffset=*/0,
rowBytes);
if (!copyTask || !fQueueManager->addTask(copyTask.get(), this)) {
return {};
}
sk_sp<SynchronizeToCpuTask> syncTask = SynchronizeToCpuTask::Make(buffer);
if (!syncTask || !fQueueManager->addTask(syncTask.get(), this)) {
return {};
}
PixelTransferResult result;
result.fTransferBuffer = std::move(buffer);
if (srcImageInfo.colorInfo() != dstColorInfo) {
result.fPixelConverter = [dims = srcRect.size(), dstColorInfo, srcImageInfo, rowBytes](
void* dst, const void* src) {
SkImageInfo srcInfo = SkImageInfo::Make(dims, srcImageInfo.colorInfo());
SkImageInfo dstInfo = SkImageInfo::Make(dims, dstColorInfo);
SkAssertResult(SkConvertPixels(dstInfo, dst, dstInfo.minRowBytes(),
srcInfo, src, rowBytes));
};
}
return result;
}
void Context::checkAsyncWorkCompletion() {
ASSERT_SINGLE_OWNER
fQueueManager->checkForFinishedWork(SyncToCpu::kNo);
}
void Context::deleteBackendTexture(BackendTexture& texture) {
ASSERT_SINGLE_OWNER
if (!texture.isValid() || texture.backend() != this->backend()) {
return;
}
fResourceProvider->deleteBackendTexture(texture);
}
///////////////////////////////////////////////////////////////////////////////////
#if GRAPHITE_TEST_UTILS
bool ContextPriv::readPixels(const SkPixmap& pm,
const TextureProxy* textureProxy,
const SkImageInfo& srcImageInfo,
int srcX, int srcY) {
auto rect = SkIRect::MakeXYWH(srcX, srcY, pm.width(), pm.height());
struct AsyncContext {
bool fCalled = false;
std::unique_ptr<const SkImage::AsyncReadResult> fResult;
} asyncContext;
fContext->asyncReadPixels(textureProxy, srcImageInfo, pm.info().colorInfo(), rect,
[](void* c, std::unique_ptr<const SkImage::AsyncReadResult> result) {
auto context = static_cast<AsyncContext*>(c);
context->fResult = std::move(result);
context->fCalled = true;
},
&asyncContext);
if (!asyncContext.fCalled) {
fContext->submit(SyncToCpu::kYes);
}
SkASSERT(asyncContext.fCalled);
if (!asyncContext.fResult) {
return false;
}
SkRectMemcpy(pm.writable_addr(), pm.rowBytes(), asyncContext.fResult->data(0),
asyncContext.fResult->rowBytes(0), pm.info().minRowBytes(),
pm.height());
return true;
}
void ContextPriv::deregisterRecorder(const Recorder* recorder) {
SKGPU_ASSERT_SINGLE_OWNER(fContext->singleOwner())
for (auto it = fContext->fTrackedRecorders.begin();
it != fContext->fTrackedRecorders.end();
it++) {
if (*it == recorder) {
fContext->fTrackedRecorders.erase(it);
return;
}
}
}
#endif
///////////////////////////////////////////////////////////////////////////////////
std::unique_ptr<Context> ContextCtorAccessor::MakeContext(
sk_sp<SharedContext> sharedContext,
std::unique_ptr<QueueManager> queueManager,
const ContextOptions& options) {
auto context = std::unique_ptr<Context>(new Context(std::move(sharedContext),
std::move(queueManager),
options));
if (context && context->finishInitialization()) {
return context;
} else {
return nullptr;
}
}
} // namespace skgpu::graphite