xref: /aosp_15_r20/external/skia/src/gpu/ganesh/GrThreadSafeCache.h (revision c8dee2aa9b3f27cf6c858bd81872bdeb2c07ed17)
1 /*
2  * Copyright 2020 Google Inc.
3  *
4  * Use of this source code is governed by a BSD-style license that can be
5  * found in the LICENSE file.
6  */
7 
8 #ifndef GrThreadSafeCache_DEFINED
9 #define GrThreadSafeCache_DEFINED
10 
11 #include "include/core/SkRefCnt.h"
12 #include "include/private/base/SkAssert.h"
13 #include "include/private/base/SkDebug.h"
14 #include "include/private/base/SkMalloc.h"
15 #include "include/private/base/SkThreadAnnotations.h"
16 #include "src/base/SkArenaAlloc.h"
17 #include "src/base/SkSpinlock.h"
18 #include "src/base/SkTInternalLList.h"
19 #include "src/core/SkTDynamicHash.h"
20 #include "src/gpu/GpuTypesPriv.h"
21 #include "src/gpu/ResourceKey.h"
22 #include "src/gpu/ganesh/GrGpuBuffer.h"
23 #include "src/gpu/ganesh/GrSurfaceProxy.h"
24 #include "src/gpu/ganesh/GrSurfaceProxyView.h"
25 #include "src/gpu/ganesh/GrTextureProxy.h"
26 
27 #include <cstddef>
28 #include <cstdint>
29 #include <tuple>
30 #include <utility>
31 
32 class GrDirectContext;
33 class GrResourceCache;
34 class SkData;
35 enum GrSurfaceOrigin : int;
36 enum class GrColorType;
37 enum class SkBackingFit;
38 struct SkISize;
39 
40 // Ganesh creates a lot of utility textures (e.g., blurred-rrect masks) that need to be shared
41 // between the direct context and all the DDL recording contexts. This thread-safe cache
42 // allows this sharing.
43 //
44 // In operation, each thread will first check if the threaded cache possesses the required texture.
45 //
46 // If a DDL thread doesn't find a needed texture it will go off and create it on the cpu and then
47 // attempt to add it to the cache. If another thread had added it in the interim, the losing thread
48 // will discard its work and use the texture the winning thread had created.
49 //
50 // If the thread in possession of the direct context doesn't find the needed texture it should
51 // add a place holder view and then queue up the draw calls to complete it. In this way the
52 // gpu-thread has precedence over the recording threads.
53 //
54 // The invariants for this cache differ a bit from those of the proxy and resource caches.
55 // For this cache:
56 //
57 //   only this cache knows the unique key - neither the proxy nor backing resource should
58 //              be discoverable in any other cache by the unique key
59 //   if a backing resource resides in the resource cache then there should be an entry in this
60 //              cache
61 //   an entry in this cache, however, doesn't guarantee that there is a corresponding entry in
62 //              the resource cache - although the entry here should be able to generate that entry
63 //              (i.e., be a lazy proxy)
64 //
65 // Wrt interactions w/ GrContext/GrResourceCache purging, we have:
66 //
67 //    Both GrContext::abandonContext and GrContext::releaseResourcesAndAbandonContext will cause
68 //    all the refs held in this cache to be dropped prior to clearing out the resource cache.
69 //
70 //    For the size_t-variant of GrContext::purgeUnlockedResources, after an initial attempt
71 //    to purge the requested amount of resources fails, uniquely held resources in this cache
72 //    will be dropped in LRU to MRU order until the cache is under budget. Note that this
73 //    prioritizes the survival of resources in this cache over those just in the resource cache.
74 //
75 //    For the 'scratchResourcesOnly' variant of GrContext::purgeUnlockedResources, this cache
76 //    won't be modified in the scratch-only case unless the resource cache is over budget (in
77 //    which case it will purge uniquely-held resources in LRU to MRU order to get
78 //    back under budget). In the non-scratch-only case, all uniquely held resources in this cache
79 //    will be released prior to the resource cache being cleared out.
80 //
81 //    For GrContext::setResourceCacheLimit, if an initial pass through the resource cache doesn't
82 //    reach the budget, uniquely held resources in this cache will be released in LRU to MRU order.
83 //
84 //    For GrContext::performDeferredCleanup, any uniquely held resources that haven't been accessed
85 //    w/in 'msNotUsed' will be released from this cache prior to the resource cache being cleaned.
86 class GrThreadSafeCache {
87 public:
88     GrThreadSafeCache();
89     ~GrThreadSafeCache();
90 
91 #if defined(GPU_TEST_UTILS)
92     int numEntries() const  SK_EXCLUDES(fSpinLock);
93 
94     size_t approxBytesUsedForHash() const  SK_EXCLUDES(fSpinLock);
95 #endif
96 
97     void dropAllRefs()  SK_EXCLUDES(fSpinLock);
98 
99     // Drop uniquely held refs until under the resource cache's budget.
100     // A null parameter means drop all uniquely held refs.
101     void dropUniqueRefs(GrResourceCache* resourceCache)  SK_EXCLUDES(fSpinLock);
102 
103     // Drop uniquely held refs that were last accessed before 'purgeTime'
104     void dropUniqueRefsOlderThan(
105             skgpu::StdSteadyClock::time_point purgeTime)  SK_EXCLUDES(fSpinLock);
106 
107     SkDEBUGCODE(bool has(const skgpu::UniqueKey&)  SK_EXCLUDES(fSpinLock);)
108 
109     GrSurfaceProxyView find(const skgpu::UniqueKey&)  SK_EXCLUDES(fSpinLock);
110     std::tuple<GrSurfaceProxyView, sk_sp<SkData>> findWithData(
111             const skgpu::UniqueKey&)  SK_EXCLUDES(fSpinLock);
112 
113     GrSurfaceProxyView add(
114             const skgpu::UniqueKey&, const GrSurfaceProxyView&)  SK_EXCLUDES(fSpinLock);
115     std::tuple<GrSurfaceProxyView, sk_sp<SkData>> addWithData(
116             const skgpu::UniqueKey&, const GrSurfaceProxyView&)  SK_EXCLUDES(fSpinLock);
117 
118     GrSurfaceProxyView findOrAdd(const skgpu::UniqueKey&,
119                                  const GrSurfaceProxyView&)  SK_EXCLUDES(fSpinLock);
120     std::tuple<GrSurfaceProxyView, sk_sp<SkData>> findOrAddWithData(
121             const skgpu::UniqueKey&, const GrSurfaceProxyView&)  SK_EXCLUDES(fSpinLock);
122 
123     // To hold vertex data in the cache and have it transparently transition from cpu-side to
124     // gpu-side while being shared between all the threads we need a ref counted object that
125     // keeps hold of the cpu-side data but allows deferred filling in of the mirroring gpu buffer.
126     class VertexData : public SkNVRefCnt<VertexData> {
127     public:
128         ~VertexData();
129 
vertices()130         const void* vertices() const { return fVertices; }
size()131         size_t size() const { return fNumVertices * fVertexSize; }
132 
numVertices()133         int numVertices() const { return fNumVertices; }
vertexSize()134         size_t vertexSize() const { return fVertexSize; }
135 
136         // TODO: make these return const GrGpuBuffers?
gpuBuffer()137         GrGpuBuffer* gpuBuffer() { return fGpuBuffer.get(); }
refGpuBuffer()138         sk_sp<GrGpuBuffer> refGpuBuffer() { return fGpuBuffer; }
139 
setGpuBuffer(sk_sp<GrGpuBuffer> gpuBuffer)140         void setGpuBuffer(sk_sp<GrGpuBuffer> gpuBuffer) {
141             // TODO: once we add the gpuBuffer we could free 'fVertices'. Deinstantiable
142             // DDLs could throw a monkey wrench into that plan though.
143             SkASSERT(!fGpuBuffer);
144             fGpuBuffer = std::move(gpuBuffer);
145         }
146 
reset()147         void reset() {
148             sk_free(const_cast<void*>(fVertices));
149             fVertices = nullptr;
150             fNumVertices = 0;
151             fVertexSize = 0;
152             fGpuBuffer.reset();
153         }
154 
155     private:
156         friend class GrThreadSafeCache;  // for access to ctor
157 
VertexData(const void * vertices,int numVertices,size_t vertexSize)158         VertexData(const void* vertices, int numVertices, size_t vertexSize)
159                 : fVertices(vertices)
160                 , fNumVertices(numVertices)
161                 , fVertexSize(vertexSize) {
162         }
163 
VertexData(sk_sp<GrGpuBuffer> gpuBuffer,int numVertices,size_t vertexSize)164         VertexData(sk_sp<GrGpuBuffer> gpuBuffer, int numVertices, size_t vertexSize)
165                 : fVertices(nullptr)
166                 , fNumVertices(numVertices)
167                 , fVertexSize(vertexSize)
168                 , fGpuBuffer(std::move(gpuBuffer)) {
169         }
170 
171         const void*        fVertices;
172         int                fNumVertices;
173         size_t             fVertexSize;
174 
175         sk_sp<GrGpuBuffer> fGpuBuffer;
176     };
177 
178     // The returned VertexData object takes ownership of 'vertices' which had better have been
179     // allocated with malloc!
180     static sk_sp<VertexData> MakeVertexData(const void* vertices,
181                                             int vertexCount,
182                                             size_t vertexSize);
183     static sk_sp<VertexData> MakeVertexData(sk_sp<GrGpuBuffer> buffer,
184                                             int vertexCount,
185                                             size_t vertexSize);
186 
187     std::tuple<sk_sp<VertexData>, sk_sp<SkData>> findVertsWithData(
188             const skgpu::UniqueKey&)  SK_EXCLUDES(fSpinLock);
189 
190     typedef bool (*IsNewerBetter)(SkData* incumbent, SkData* challenger);
191 
192     std::tuple<sk_sp<VertexData>, sk_sp<SkData>> addVertsWithData(
193                                                         const skgpu::UniqueKey&,
194                                                         sk_sp<VertexData>,
195                                                         IsNewerBetter)  SK_EXCLUDES(fSpinLock);
196 
197     void remove(const skgpu::UniqueKey&)  SK_EXCLUDES(fSpinLock);
198 
199     // To allow gpu-created resources to have priority, we pre-emptively place a lazy proxy
200     // in the thread-safe cache (with findOrAdd). The Trampoline object allows that lazy proxy to
201     // be instantiated with some later generated rendering result.
202     class Trampoline : public SkRefCnt {
203     public:
204         sk_sp<GrTextureProxy> fProxy;
205     };
206 
207     static std::tuple<GrSurfaceProxyView, sk_sp<Trampoline>> CreateLazyView(GrDirectContext*,
208                                                                             GrColorType,
209                                                                             SkISize dimensions,
210                                                                             GrSurfaceOrigin,
211                                                                             SkBackingFit);
212 private:
213     struct Entry {
EntryEntry214         Entry(const skgpu::UniqueKey& key, const GrSurfaceProxyView& view)
215                 : fKey(key), fView(view), fTag(Entry::Tag::kView) {}
216 
EntryEntry217         Entry(const skgpu::UniqueKey& key, sk_sp<VertexData> vertData)
218                 : fKey(key), fVertData(std::move(vertData)), fTag(Entry::Tag::kVertData) {}
219 
~EntryEntry220         ~Entry() {
221             this->makeEmpty();
222         }
223 
uniquelyHeldEntry224         bool uniquelyHeld() const {
225             SkASSERT(fTag != Tag::kEmpty);
226 
227             if (fTag == Tag::kView && fView.proxy()->unique()) {
228                 return true;
229             } else if (fTag == Tag::kVertData && fVertData->unique()) {
230                 return true;
231             }
232 
233             return false;
234         }
235 
keyEntry236         const skgpu::UniqueKey& key() const {
237             SkASSERT(fTag != Tag::kEmpty);
238             return fKey;
239         }
240 
getCustomDataEntry241         SkData* getCustomData() const {
242             SkASSERT(fTag != Tag::kEmpty);
243             return fKey.getCustomData();
244         }
245 
refCustomDataEntry246         sk_sp<SkData> refCustomData() const {
247             SkASSERT(fTag != Tag::kEmpty);
248             return fKey.refCustomData();
249         }
250 
viewEntry251         GrSurfaceProxyView view() {
252             SkASSERT(fTag == Tag::kView);
253             return fView;
254         }
255 
vertexDataEntry256         sk_sp<VertexData> vertexData() {
257             SkASSERT(fTag == Tag::kVertData);
258             return fVertData;
259         }
260 
setEntry261         void set(const skgpu::UniqueKey& key, const GrSurfaceProxyView& view) {
262             SkASSERT(fTag == Tag::kEmpty);
263             fKey = key;
264             fView = view;
265             fTag = Tag::kView;
266         }
267 
makeEmptyEntry268         void makeEmpty() {
269             fKey.reset();
270             if (fTag == Tag::kView) {
271                 fView.reset();
272             } else if (fTag == Tag::kVertData) {
273                 fVertData.reset();
274             }
275             fTag = Tag::kEmpty;
276         }
277 
setEntry278         void set(const skgpu::UniqueKey& key, sk_sp<VertexData> vertData) {
279             SkASSERT(fTag == Tag::kEmpty || fTag == Tag::kVertData);
280             fKey = key;
281             fVertData = std::move(vertData);
282             fTag = Tag::kVertData;
283         }
284 
285         // The thread-safe cache gets to directly manipulate the llist and last-access members
286         skgpu::StdSteadyClock::time_point fLastAccess;
287         SK_DECLARE_INTERNAL_LLIST_INTERFACE(Entry);
288 
289         // for SkTDynamicHash
GetKeyEntry290         static const skgpu::UniqueKey& GetKey(const Entry& e) {
291             SkASSERT(e.fTag != Tag::kEmpty);
292             return e.fKey;
293         }
HashEntry294         static uint32_t Hash(const skgpu::UniqueKey& key) { return key.hash(); }
295 
296     private:
297         // Note: the unique key is stored here bc it is never attached to a proxy or a GrTexture
298         skgpu::UniqueKey             fKey;
299         union {
300             GrSurfaceProxyView  fView;
301             sk_sp<VertexData>   fVertData;
302         };
303 
304         enum class Tag {
305             kEmpty,
306             kView,
307             kVertData,
308         };
309         Tag fTag{Tag::kEmpty};
310     };
311 
312     void makeExistingEntryMRU(Entry*)  SK_REQUIRES(fSpinLock);
313     Entry* makeNewEntryMRU(Entry*)  SK_REQUIRES(fSpinLock);
314 
315     Entry* getEntry(const skgpu::UniqueKey&, const GrSurfaceProxyView&)  SK_REQUIRES(fSpinLock);
316     Entry* getEntry(const skgpu::UniqueKey&, sk_sp<VertexData>)  SK_REQUIRES(fSpinLock);
317 
318     void recycleEntry(Entry*)  SK_REQUIRES(fSpinLock);
319 
320     std::tuple<GrSurfaceProxyView, sk_sp<SkData>> internalFind(
321             const skgpu::UniqueKey&)  SK_REQUIRES(fSpinLock);
322     std::tuple<GrSurfaceProxyView, sk_sp<SkData>> internalAdd(
323             const skgpu::UniqueKey&, const GrSurfaceProxyView&)  SK_REQUIRES(fSpinLock);
324 
325     std::tuple<sk_sp<VertexData>, sk_sp<SkData>> internalFindVerts(
326             const skgpu::UniqueKey&)  SK_REQUIRES(fSpinLock);
327     std::tuple<sk_sp<VertexData>, sk_sp<SkData>> internalAddVerts(
328             const skgpu::UniqueKey&, sk_sp<VertexData>, IsNewerBetter)  SK_REQUIRES(fSpinLock);
329 
330     mutable SkSpinlock fSpinLock;
331 
332     SkTDynamicHash<Entry, skgpu::UniqueKey> fUniquelyKeyedEntryMap  SK_GUARDED_BY(fSpinLock);
333     // The head of this list is the MRU
334     SkTInternalLList<Entry>            fUniquelyKeyedEntryList  SK_GUARDED_BY(fSpinLock);
335 
336     // TODO: empirically determine this from the skps
337     static const int kInitialArenaSize = 64 * sizeof(Entry);
338 
339     char                         fStorage[kInitialArenaSize];
340     SkArenaAlloc                 fEntryAllocator{fStorage, kInitialArenaSize, kInitialArenaSize};
341     Entry*                       fFreeEntryList  SK_GUARDED_BY(fSpinLock);
342 };
343 
344 #endif // GrThreadSafeCache_DEFINED
345