blob: 09e243852f1a82891cd98489315c6c0ab9ff93db [file] [log] [blame]
/*
* Copyright 2020 Google Inc.
*
* Use of this source code is governed by a BSD-style license that can be
* found in the LICENSE file.
*/
#ifndef GrThreadSafeCache_DEFINED
#define GrThreadSafeCache_DEFINED
#include "include/core/SkRefCnt.h"
#include "include/private/SkSpinlock.h"
#include "src/base/SkArenaAlloc.h"
#include "src/core/SkTDynamicHash.h"
#include "src/core/SkTInternalLList.h"
#include "src/gpu/ganesh/GrGpuBuffer.h"
#include "src/gpu/ganesh/GrSurfaceProxy.h"
#include "src/gpu/ganesh/GrSurfaceProxyView.h"
// Ganesh creates a lot of utility textures (e.g., blurred-rrect masks) that need to be shared
// between the direct context and all the DDL recording contexts. This thread-safe cache
// allows this sharing.
//
// In operation, each thread will first check if the threaded cache possesses the required texture.
//
// If a DDL thread doesn't find a needed texture it will go off and create it on the cpu and then
// attempt to add it to the cache. If another thread had added it in the interim, the losing thread
// will discard its work and use the texture the winning thread had created.
//
// If the thread in possession of the direct context doesn't find the needed texture it should
// add a place holder view and then queue up the draw calls to complete it. In this way the
// gpu-thread has precedence over the recording threads.
//
// The invariants for this cache differ a bit from those of the proxy and resource caches.
// For this cache:
//
// only this cache knows the unique key - neither the proxy nor backing resource should
// be discoverable in any other cache by the unique key
// if a backing resource resides in the resource cache then there should be an entry in this
// cache
// an entry in this cache, however, doesn't guarantee that there is a corresponding entry in
// the resource cache - although the entry here should be able to generate that entry
// (i.e., be a lazy proxy)
//
// Wrt interactions w/ GrContext/GrResourceCache purging, we have:
//
// Both GrContext::abandonContext and GrContext::releaseResourcesAndAbandonContext will cause
// all the refs held in this cache to be dropped prior to clearing out the resource cache.
//
// For the size_t-variant of GrContext::purgeUnlockedResources, after an initial attempt
// to purge the requested amount of resources fails, uniquely held resources in this cache
// will be dropped in LRU to MRU order until the cache is under budget. Note that this
// prioritizes the survival of resources in this cache over those just in the resource cache.
//
// For the 'scratchResourcesOnly' variant of GrContext::purgeUnlockedResources, this cache
// won't be modified in the scratch-only case unless the resource cache is over budget (in
// which case it will purge uniquely-held resources in LRU to MRU order to get
// back under budget). In the non-scratch-only case, all uniquely held resources in this cache
// will be released prior to the resource cache being cleared out.
//
// For GrContext::setResourceCacheLimit, if an initial pass through the resource cache doesn't
// reach the budget, uniquely held resources in this cache will be released in LRU to MRU order.
//
// For GrContext::performDeferredCleanup, any uniquely held resources that haven't been accessed
// w/in 'msNotUsed' will be released from this cache prior to the resource cache being cleaned.
class GrThreadSafeCache {
public:
GrThreadSafeCache();
~GrThreadSafeCache();
#if GR_TEST_UTILS
int numEntries() const SK_EXCLUDES(fSpinLock);
size_t approxBytesUsedForHash() const SK_EXCLUDES(fSpinLock);
#endif
void dropAllRefs() SK_EXCLUDES(fSpinLock);
// Drop uniquely held refs until under the resource cache's budget.
// A null parameter means drop all uniquely held refs.
void dropUniqueRefs(GrResourceCache* resourceCache) SK_EXCLUDES(fSpinLock);
// Drop uniquely held refs that were last accessed before 'purgeTime'
void dropUniqueRefsOlderThan(GrStdSteadyClock::time_point purgeTime) SK_EXCLUDES(fSpinLock);
SkDEBUGCODE(bool has(const skgpu::UniqueKey&) SK_EXCLUDES(fSpinLock);)
GrSurfaceProxyView find(const skgpu::UniqueKey&) SK_EXCLUDES(fSpinLock);
std::tuple<GrSurfaceProxyView, sk_sp<SkData>> findWithData(
const skgpu::UniqueKey&) SK_EXCLUDES(fSpinLock);
GrSurfaceProxyView add(
const skgpu::UniqueKey&, const GrSurfaceProxyView&) SK_EXCLUDES(fSpinLock);
std::tuple<GrSurfaceProxyView, sk_sp<SkData>> addWithData(
const skgpu::UniqueKey&, const GrSurfaceProxyView&) SK_EXCLUDES(fSpinLock);
GrSurfaceProxyView findOrAdd(const skgpu::UniqueKey&,
const GrSurfaceProxyView&) SK_EXCLUDES(fSpinLock);
std::tuple<GrSurfaceProxyView, sk_sp<SkData>> findOrAddWithData(
const skgpu::UniqueKey&, const GrSurfaceProxyView&) SK_EXCLUDES(fSpinLock);
// To hold vertex data in the cache and have it transparently transition from cpu-side to
// gpu-side while being shared between all the threads we need a ref counted object that
// keeps hold of the cpu-side data but allows deferred filling in of the mirroring gpu buffer.
class VertexData : public SkNVRefCnt<VertexData> {
public:
~VertexData();
const void* vertices() const { return fVertices; }
size_t size() const { return fNumVertices * fVertexSize; }
int numVertices() const { return fNumVertices; }
size_t vertexSize() const { return fVertexSize; }
// TODO: make these return const GrGpuBuffers?
GrGpuBuffer* gpuBuffer() { return fGpuBuffer.get(); }
sk_sp<GrGpuBuffer> refGpuBuffer() { return fGpuBuffer; }
void setGpuBuffer(sk_sp<GrGpuBuffer> gpuBuffer) {
// TODO: once we add the gpuBuffer we could free 'fVertices'. Deinstantiable
// DDLs could throw a monkey wrench into that plan though.
SkASSERT(!fGpuBuffer);
fGpuBuffer = gpuBuffer;
}
void reset() {
sk_free(const_cast<void*>(fVertices));
fVertices = nullptr;
fNumVertices = 0;
fVertexSize = 0;
fGpuBuffer.reset();
}
private:
friend class GrThreadSafeCache; // for access to ctor
VertexData(const void* vertices, int numVertices, size_t vertexSize)
: fVertices(vertices)
, fNumVertices(numVertices)
, fVertexSize(vertexSize) {
}
VertexData(sk_sp<GrGpuBuffer> gpuBuffer, int numVertices, size_t vertexSize)
: fVertices(nullptr)
, fNumVertices(numVertices)
, fVertexSize(vertexSize)
, fGpuBuffer(std::move(gpuBuffer)) {
}
const void* fVertices;
int fNumVertices;
size_t fVertexSize;
sk_sp<GrGpuBuffer> fGpuBuffer;
};
// The returned VertexData object takes ownership of 'vertices' which had better have been
// allocated with malloc!
static sk_sp<VertexData> MakeVertexData(const void* vertices,
int vertexCount,
size_t vertexSize);
static sk_sp<VertexData> MakeVertexData(sk_sp<GrGpuBuffer> buffer,
int vertexCount,
size_t vertexSize);
std::tuple<sk_sp<VertexData>, sk_sp<SkData>> findVertsWithData(
const skgpu::UniqueKey&) SK_EXCLUDES(fSpinLock);
typedef bool (*IsNewerBetter)(SkData* incumbent, SkData* challenger);
std::tuple<sk_sp<VertexData>, sk_sp<SkData>> addVertsWithData(
const skgpu::UniqueKey&,
sk_sp<VertexData>,
IsNewerBetter) SK_EXCLUDES(fSpinLock);
void remove(const skgpu::UniqueKey&) SK_EXCLUDES(fSpinLock);
// To allow gpu-created resources to have priority, we pre-emptively place a lazy proxy
// in the thread-safe cache (with findOrAdd). The Trampoline object allows that lazy proxy to
// be instantiated with some later generated rendering result.
class Trampoline : public SkRefCnt {
public:
sk_sp<GrTextureProxy> fProxy;
};
static std::tuple<GrSurfaceProxyView, sk_sp<Trampoline>> CreateLazyView(GrDirectContext*,
GrColorType,
SkISize dimensions,
GrSurfaceOrigin,
SkBackingFit);
private:
struct Entry {
Entry(const skgpu::UniqueKey& key, const GrSurfaceProxyView& view)
: fKey(key)
, fView(view)
, fTag(Entry::kView) {
}
Entry(const skgpu::UniqueKey& key, sk_sp<VertexData> vertData)
: fKey(key)
, fVertData(std::move(vertData))
, fTag(Entry::kVertData) {
}
~Entry() {
this->makeEmpty();
}
bool uniquelyHeld() const {
SkASSERT(fTag != kEmpty);
if (fTag == kView && fView.proxy()->unique()) {
return true;
} else if (fTag == kVertData && fVertData->unique()) {
return true;
}
return false;
}
const skgpu::UniqueKey& key() const {
SkASSERT(fTag != kEmpty);
return fKey;
}
SkData* getCustomData() const {
SkASSERT(fTag != kEmpty);
return fKey.getCustomData();
}
sk_sp<SkData> refCustomData() const {
SkASSERT(fTag != kEmpty);
return fKey.refCustomData();
}
GrSurfaceProxyView view() {
SkASSERT(fTag == kView);
return fView;
}
sk_sp<VertexData> vertexData() {
SkASSERT(fTag == kVertData);
return fVertData;
}
void set(const skgpu::UniqueKey& key, const GrSurfaceProxyView& view) {
SkASSERT(fTag == kEmpty);
fKey = key;
fView = view;
fTag = kView;
}
void makeEmpty() {
fKey.reset();
if (fTag == kView) {
fView.reset();
} else if (fTag == kVertData) {
fVertData.reset();
}
fTag = kEmpty;
}
void set(const skgpu::UniqueKey& key, sk_sp<VertexData> vertData) {
SkASSERT(fTag == kEmpty || fTag == kVertData);
fKey = key;
fVertData = vertData;
fTag = kVertData;
}
// The thread-safe cache gets to directly manipulate the llist and last-access members
GrStdSteadyClock::time_point fLastAccess;
SK_DECLARE_INTERNAL_LLIST_INTERFACE(Entry);
// for SkTDynamicHash
static const skgpu::UniqueKey& GetKey(const Entry& e) {
SkASSERT(e.fTag != kEmpty);
return e.fKey;
}
static uint32_t Hash(const skgpu::UniqueKey& key) { return key.hash(); }
private:
// Note: the unique key is stored here bc it is never attached to a proxy or a GrTexture
skgpu::UniqueKey fKey;
union {
GrSurfaceProxyView fView;
sk_sp<VertexData> fVertData;
};
enum {
kEmpty,
kView,
kVertData,
} fTag { kEmpty };
};
void makeExistingEntryMRU(Entry*) SK_REQUIRES(fSpinLock);
Entry* makeNewEntryMRU(Entry*) SK_REQUIRES(fSpinLock);
Entry* getEntry(const skgpu::UniqueKey&, const GrSurfaceProxyView&) SK_REQUIRES(fSpinLock);
Entry* getEntry(const skgpu::UniqueKey&, sk_sp<VertexData>) SK_REQUIRES(fSpinLock);
void recycleEntry(Entry*) SK_REQUIRES(fSpinLock);
std::tuple<GrSurfaceProxyView, sk_sp<SkData>> internalFind(
const skgpu::UniqueKey&) SK_REQUIRES(fSpinLock);
std::tuple<GrSurfaceProxyView, sk_sp<SkData>> internalAdd(
const skgpu::UniqueKey&, const GrSurfaceProxyView&) SK_REQUIRES(fSpinLock);
std::tuple<sk_sp<VertexData>, sk_sp<SkData>> internalFindVerts(
const skgpu::UniqueKey&) SK_REQUIRES(fSpinLock);
std::tuple<sk_sp<VertexData>, sk_sp<SkData>> internalAddVerts(
const skgpu::UniqueKey&, sk_sp<VertexData>, IsNewerBetter) SK_REQUIRES(fSpinLock);
mutable SkSpinlock fSpinLock;
SkTDynamicHash<Entry, skgpu::UniqueKey> fUniquelyKeyedEntryMap SK_GUARDED_BY(fSpinLock);
// The head of this list is the MRU
SkTInternalLList<Entry> fUniquelyKeyedEntryList SK_GUARDED_BY(fSpinLock);
// TODO: empirically determine this from the skps
static const int kInitialArenaSize = 64 * sizeof(Entry);
char fStorage[kInitialArenaSize];
SkArenaAlloc fEntryAllocator{fStorage, kInitialArenaSize, kInitialArenaSize};
Entry* fFreeEntryList SK_GUARDED_BY(fSpinLock);
};
#endif // GrThreadSafeCache_DEFINED