1 /* 2 * Copyright 2020 Google Inc. 3 * 4 * Use of this source code is governed by a BSD-style license that can be 5 * found in the LICENSE file. 6 */ 7 8 #ifndef GrThreadSafeCache_DEFINED 9 #define GrThreadSafeCache_DEFINED 10 11 #include "include/core/SkRefCnt.h" 12 #include "src/base/SkArenaAlloc.h" 13 #include "src/base/SkSpinlock.h" 14 #include "src/base/SkTInternalLList.h" 15 #include "src/core/SkTDynamicHash.h" 16 #include "src/gpu/ganesh/GrGpuBuffer.h" 17 #include "src/gpu/ganesh/GrSurfaceProxy.h" 18 #include "src/gpu/ganesh/GrSurfaceProxyView.h" 19 20 // Ganesh creates a lot of utility textures (e.g., blurred-rrect masks) that need to be shared 21 // between the direct context and all the DDL recording contexts. This thread-safe cache 22 // allows this sharing. 23 // 24 // In operation, each thread will first check if the threaded cache possesses the required texture. 25 // 26 // If a DDL thread doesn't find a needed texture it will go off and create it on the cpu and then 27 // attempt to add it to the cache. If another thread had added it in the interim, the losing thread 28 // will discard its work and use the texture the winning thread had created. 29 // 30 // If the thread in possession of the direct context doesn't find the needed texture it should 31 // add a place holder view and then queue up the draw calls to complete it. In this way the 32 // gpu-thread has precedence over the recording threads. 33 // 34 // The invariants for this cache differ a bit from those of the proxy and resource caches. 35 // For this cache: 36 // 37 // only this cache knows the unique key - neither the proxy nor backing resource should 38 // be discoverable in any other cache by the unique key 39 // if a backing resource resides in the resource cache then there should be an entry in this 40 // cache 41 // an entry in this cache, however, doesn't guarantee that there is a corresponding entry in 42 // the resource cache - although the entry here should be able to generate that entry 43 // (i.e., be a lazy proxy) 44 // 45 // Wrt interactions w/ GrContext/GrResourceCache purging, we have: 46 // 47 // Both GrContext::abandonContext and GrContext::releaseResourcesAndAbandonContext will cause 48 // all the refs held in this cache to be dropped prior to clearing out the resource cache. 49 // 50 // For the size_t-variant of GrContext::purgeUnlockedResources, after an initial attempt 51 // to purge the requested amount of resources fails, uniquely held resources in this cache 52 // will be dropped in LRU to MRU order until the cache is under budget. Note that this 53 // prioritizes the survival of resources in this cache over those just in the resource cache. 54 // 55 // For the 'scratchResourcesOnly' variant of GrContext::purgeUnlockedResources, this cache 56 // won't be modified in the scratch-only case unless the resource cache is over budget (in 57 // which case it will purge uniquely-held resources in LRU to MRU order to get 58 // back under budget). In the non-scratch-only case, all uniquely held resources in this cache 59 // will be released prior to the resource cache being cleared out. 60 // 61 // For GrContext::setResourceCacheLimit, if an initial pass through the resource cache doesn't 62 // reach the budget, uniquely held resources in this cache will be released in LRU to MRU order. 63 // 64 // For GrContext::performDeferredCleanup, any uniquely held resources that haven't been accessed 65 // w/in 'msNotUsed' will be released from this cache prior to the resource cache being cleaned. 66 class GrThreadSafeCache { 67 public: 68 GrThreadSafeCache(); 69 ~GrThreadSafeCache(); 70 71 #if defined(GR_TEST_UTILS) 72 int numEntries() const SK_EXCLUDES(fSpinLock); 73 74 size_t approxBytesUsedForHash() const SK_EXCLUDES(fSpinLock); 75 #endif 76 77 void dropAllRefs() SK_EXCLUDES(fSpinLock); 78 79 // Drop uniquely held refs until under the resource cache's budget. 80 // A null parameter means drop all uniquely held refs. 81 void dropUniqueRefs(GrResourceCache* resourceCache) SK_EXCLUDES(fSpinLock); 82 83 // Drop uniquely held refs that were last accessed before 'purgeTime' 84 void dropUniqueRefsOlderThan( 85 skgpu::StdSteadyClock::time_point purgeTime) SK_EXCLUDES(fSpinLock); 86 87 SkDEBUGCODE(bool has(const skgpu::UniqueKey&) SK_EXCLUDES(fSpinLock);) 88 89 GrSurfaceProxyView find(const skgpu::UniqueKey&) SK_EXCLUDES(fSpinLock); 90 std::tuple<GrSurfaceProxyView, sk_sp<SkData>> findWithData( 91 const skgpu::UniqueKey&) SK_EXCLUDES(fSpinLock); 92 93 GrSurfaceProxyView add( 94 const skgpu::UniqueKey&, const GrSurfaceProxyView&) SK_EXCLUDES(fSpinLock); 95 std::tuple<GrSurfaceProxyView, sk_sp<SkData>> addWithData( 96 const skgpu::UniqueKey&, const GrSurfaceProxyView&) SK_EXCLUDES(fSpinLock); 97 98 GrSurfaceProxyView findOrAdd(const skgpu::UniqueKey&, 99 const GrSurfaceProxyView&) SK_EXCLUDES(fSpinLock); 100 std::tuple<GrSurfaceProxyView, sk_sp<SkData>> findOrAddWithData( 101 const skgpu::UniqueKey&, const GrSurfaceProxyView&) SK_EXCLUDES(fSpinLock); 102 103 // To hold vertex data in the cache and have it transparently transition from cpu-side to 104 // gpu-side while being shared between all the threads we need a ref counted object that 105 // keeps hold of the cpu-side data but allows deferred filling in of the mirroring gpu buffer. 106 class VertexData : public SkNVRefCnt<VertexData> { 107 public: 108 ~VertexData(); 109 vertices()110 const void* vertices() const { return fVertices; } size()111 size_t size() const { return fNumVertices * fVertexSize; } 112 numVertices()113 int numVertices() const { return fNumVertices; } vertexSize()114 size_t vertexSize() const { return fVertexSize; } 115 116 // TODO: make these return const GrGpuBuffers? gpuBuffer()117 GrGpuBuffer* gpuBuffer() { return fGpuBuffer.get(); } refGpuBuffer()118 sk_sp<GrGpuBuffer> refGpuBuffer() { return fGpuBuffer; } 119 setGpuBuffer(sk_sp<GrGpuBuffer> gpuBuffer)120 void setGpuBuffer(sk_sp<GrGpuBuffer> gpuBuffer) { 121 // TODO: once we add the gpuBuffer we could free 'fVertices'. Deinstantiable 122 // DDLs could throw a monkey wrench into that plan though. 123 SkASSERT(!fGpuBuffer); 124 fGpuBuffer = std::move(gpuBuffer); 125 } 126 reset()127 void reset() { 128 sk_free(const_cast<void*>(fVertices)); 129 fVertices = nullptr; 130 fNumVertices = 0; 131 fVertexSize = 0; 132 fGpuBuffer.reset(); 133 } 134 135 private: 136 friend class GrThreadSafeCache; // for access to ctor 137 VertexData(const void * vertices,int numVertices,size_t vertexSize)138 VertexData(const void* vertices, int numVertices, size_t vertexSize) 139 : fVertices(vertices) 140 , fNumVertices(numVertices) 141 , fVertexSize(vertexSize) { 142 } 143 VertexData(sk_sp<GrGpuBuffer> gpuBuffer,int numVertices,size_t vertexSize)144 VertexData(sk_sp<GrGpuBuffer> gpuBuffer, int numVertices, size_t vertexSize) 145 : fVertices(nullptr) 146 , fNumVertices(numVertices) 147 , fVertexSize(vertexSize) 148 , fGpuBuffer(std::move(gpuBuffer)) { 149 } 150 151 const void* fVertices; 152 int fNumVertices; 153 size_t fVertexSize; 154 155 sk_sp<GrGpuBuffer> fGpuBuffer; 156 }; 157 158 // The returned VertexData object takes ownership of 'vertices' which had better have been 159 // allocated with malloc! 160 static sk_sp<VertexData> MakeVertexData(const void* vertices, 161 int vertexCount, 162 size_t vertexSize); 163 static sk_sp<VertexData> MakeVertexData(sk_sp<GrGpuBuffer> buffer, 164 int vertexCount, 165 size_t vertexSize); 166 167 std::tuple<sk_sp<VertexData>, sk_sp<SkData>> findVertsWithData( 168 const skgpu::UniqueKey&) SK_EXCLUDES(fSpinLock); 169 170 typedef bool (*IsNewerBetter)(SkData* incumbent, SkData* challenger); 171 172 std::tuple<sk_sp<VertexData>, sk_sp<SkData>> addVertsWithData( 173 const skgpu::UniqueKey&, 174 sk_sp<VertexData>, 175 IsNewerBetter) SK_EXCLUDES(fSpinLock); 176 177 void remove(const skgpu::UniqueKey&) SK_EXCLUDES(fSpinLock); 178 179 // To allow gpu-created resources to have priority, we pre-emptively place a lazy proxy 180 // in the thread-safe cache (with findOrAdd). The Trampoline object allows that lazy proxy to 181 // be instantiated with some later generated rendering result. 182 class Trampoline : public SkRefCnt { 183 public: 184 sk_sp<GrTextureProxy> fProxy; 185 }; 186 187 static std::tuple<GrSurfaceProxyView, sk_sp<Trampoline>> CreateLazyView(GrDirectContext*, 188 GrColorType, 189 SkISize dimensions, 190 GrSurfaceOrigin, 191 SkBackingFit); 192 private: 193 struct Entry { EntryEntry194 Entry(const skgpu::UniqueKey& key, const GrSurfaceProxyView& view) 195 : fKey(key) 196 , fView(view) 197 , fTag(Entry::kView) { 198 } 199 EntryEntry200 Entry(const skgpu::UniqueKey& key, sk_sp<VertexData> vertData) 201 : fKey(key) 202 , fVertData(std::move(vertData)) 203 , fTag(Entry::kVertData) { 204 } 205 ~EntryEntry206 ~Entry() { 207 this->makeEmpty(); 208 } 209 uniquelyHeldEntry210 bool uniquelyHeld() const { 211 SkASSERT(fTag != kEmpty); 212 213 if (fTag == kView && fView.proxy()->unique()) { 214 return true; 215 } else if (fTag == kVertData && fVertData->unique()) { 216 return true; 217 } 218 219 return false; 220 } 221 keyEntry222 const skgpu::UniqueKey& key() const { 223 SkASSERT(fTag != kEmpty); 224 return fKey; 225 } 226 getCustomDataEntry227 SkData* getCustomData() const { 228 SkASSERT(fTag != kEmpty); 229 return fKey.getCustomData(); 230 } 231 refCustomDataEntry232 sk_sp<SkData> refCustomData() const { 233 SkASSERT(fTag != kEmpty); 234 return fKey.refCustomData(); 235 } 236 viewEntry237 GrSurfaceProxyView view() { 238 SkASSERT(fTag == kView); 239 return fView; 240 } 241 vertexDataEntry242 sk_sp<VertexData> vertexData() { 243 SkASSERT(fTag == kVertData); 244 return fVertData; 245 } 246 setEntry247 void set(const skgpu::UniqueKey& key, const GrSurfaceProxyView& view) { 248 SkASSERT(fTag == kEmpty); 249 fKey = key; 250 fView = view; 251 fTag = kView; 252 } 253 makeEmptyEntry254 void makeEmpty() { 255 fKey.reset(); 256 if (fTag == kView) { 257 fView.reset(); 258 } else if (fTag == kVertData) { 259 fVertData.reset(); 260 } 261 fTag = kEmpty; 262 } 263 setEntry264 void set(const skgpu::UniqueKey& key, sk_sp<VertexData> vertData) { 265 SkASSERT(fTag == kEmpty || fTag == kVertData); 266 fKey = key; 267 fVertData = std::move(vertData); 268 fTag = kVertData; 269 } 270 271 // The thread-safe cache gets to directly manipulate the llist and last-access members 272 skgpu::StdSteadyClock::time_point fLastAccess; 273 SK_DECLARE_INTERNAL_LLIST_INTERFACE(Entry); 274 275 // for SkTDynamicHash GetKeyEntry276 static const skgpu::UniqueKey& GetKey(const Entry& e) { 277 SkASSERT(e.fTag != kEmpty); 278 return e.fKey; 279 } HashEntry280 static uint32_t Hash(const skgpu::UniqueKey& key) { return key.hash(); } 281 282 private: 283 // Note: the unique key is stored here bc it is never attached to a proxy or a GrTexture 284 skgpu::UniqueKey fKey; 285 union { 286 GrSurfaceProxyView fView; 287 sk_sp<VertexData> fVertData; 288 }; 289 290 enum { 291 kEmpty, 292 kView, 293 kVertData, 294 } fTag { kEmpty }; 295 }; 296 297 void makeExistingEntryMRU(Entry*) SK_REQUIRES(fSpinLock); 298 Entry* makeNewEntryMRU(Entry*) SK_REQUIRES(fSpinLock); 299 300 Entry* getEntry(const skgpu::UniqueKey&, const GrSurfaceProxyView&) SK_REQUIRES(fSpinLock); 301 Entry* getEntry(const skgpu::UniqueKey&, sk_sp<VertexData>) SK_REQUIRES(fSpinLock); 302 303 void recycleEntry(Entry*) SK_REQUIRES(fSpinLock); 304 305 std::tuple<GrSurfaceProxyView, sk_sp<SkData>> internalFind( 306 const skgpu::UniqueKey&) SK_REQUIRES(fSpinLock); 307 std::tuple<GrSurfaceProxyView, sk_sp<SkData>> internalAdd( 308 const skgpu::UniqueKey&, const GrSurfaceProxyView&) SK_REQUIRES(fSpinLock); 309 310 std::tuple<sk_sp<VertexData>, sk_sp<SkData>> internalFindVerts( 311 const skgpu::UniqueKey&) SK_REQUIRES(fSpinLock); 312 std::tuple<sk_sp<VertexData>, sk_sp<SkData>> internalAddVerts( 313 const skgpu::UniqueKey&, sk_sp<VertexData>, IsNewerBetter) SK_REQUIRES(fSpinLock); 314 315 mutable SkSpinlock fSpinLock; 316 317 SkTDynamicHash<Entry, skgpu::UniqueKey> fUniquelyKeyedEntryMap SK_GUARDED_BY(fSpinLock); 318 // The head of this list is the MRU 319 SkTInternalLList<Entry> fUniquelyKeyedEntryList SK_GUARDED_BY(fSpinLock); 320 321 // TODO: empirically determine this from the skps 322 static const int kInitialArenaSize = 64 * sizeof(Entry); 323 324 char fStorage[kInitialArenaSize]; 325 SkArenaAlloc fEntryAllocator{fStorage, kInitialArenaSize, kInitialArenaSize}; 326 Entry* fFreeEntryList SK_GUARDED_BY(fSpinLock); 327 }; 328 329 #endif // GrThreadSafeCache_DEFINED 330