/* * Copyright 2021 Google Inc. * * Use of this source code is governed by a BSD-style license that can be * found in the LICENSE file. */ #include "src/gpu/graphite/BufferManager.h" #include "include/gpu/graphite/Recording.h" #include "src/gpu/graphite/Caps.h" #include "src/gpu/graphite/ContextPriv.h" #include "src/gpu/graphite/Log.h" #include "src/gpu/graphite/QueueManager.h" #include "src/gpu/graphite/RecordingPriv.h" #include "src/gpu/graphite/ResourceProvider.h" #include "src/gpu/graphite/SharedContext.h" #include "src/gpu/graphite/UploadBufferManager.h" #include "src/gpu/graphite/task/ClearBuffersTask.h" #include "src/gpu/graphite/task/CopyTask.h" namespace skgpu::graphite { namespace { // TODO: Tune these values on real world data static constexpr size_t kVertexBufferSize = 16 << 10; // 16 KB static constexpr size_t kIndexBufferSize = 2 << 10; // 2 KB static constexpr size_t kUniformBufferSize = 2 << 10; // 2 KB static constexpr size_t kStorageBufferSize = 2 << 10; // 2 KB // The limit for all data created by the StaticBufferManager. This data remains alive for // the entire SharedContext so we want to keep it small and give a concrete upper bound to // clients for our steady-state memory usage. // FIXME The current usage is 4732 bytes across static vertex and index buffers, but that includes // multiple copies of tessellation data, and an unoptimized AnalyticRRect mesh. Once those issues // are addressed, we can tighten this and decide on the transfer buffer sizing as well. [[maybe_unused]] static constexpr size_t kMaxStaticDataSize = 6 << 10; size_t sufficient_block_size(size_t requiredBytes, size_t blockSize) { // Always request a buffer at least 'requiredBytes', but keep them in multiples of // 'blockSize' for improved reuse. static constexpr size_t kMaxSize = std::numeric_limits::max(); size_t maxBlocks = kMaxSize / blockSize; size_t blocks = (requiredBytes / blockSize) + 1; size_t bufferSize = blocks > maxBlocks ? kMaxSize : (blocks * blockSize); SkASSERT(requiredBytes < bufferSize); return bufferSize; } bool can_fit(size_t requestedSize, size_t allocatedSize, size_t currentOffset, size_t alignment) { size_t startOffset = SkAlignTo(currentOffset, alignment); return requestedSize <= (allocatedSize - startOffset); } size_t starting_alignment(BufferType type, bool useTransferBuffers, const Caps* caps) { // Both vertex and index data is aligned to 4 bytes by default size_t alignment = 4; if (type == BufferType::kUniform) { alignment = caps->requiredUniformBufferAlignment(); } else if (type == BufferType::kStorage || type == BufferType::kVertexStorage || type == BufferType::kIndexStorage || type == BufferType::kIndirect) { alignment = caps->requiredStorageBufferAlignment(); } if (useTransferBuffers) { alignment = std::max(alignment, caps->requiredTransferBufferAlignment()); } return alignment; } } // anonymous namespace // ------------------------------------------------------------------------------------------------ // ScratchBuffer ScratchBuffer::ScratchBuffer(size_t size, size_t alignment, sk_sp buffer, DrawBufferManager* owner) : fSize(size) , fAlignment(alignment) , fBuffer(std::move(buffer)) , fOwner(owner) { SkASSERT(fSize > 0); SkASSERT(fBuffer); SkASSERT(fOwner); } ScratchBuffer::~ScratchBuffer() { this->returnToPool(); } BindBufferInfo ScratchBuffer::suballocate(size_t requiredBytes) { if (!this->isValid()) { return {}; } if (!can_fit(requiredBytes, fBuffer->size(), fOffset, fAlignment)) { return {}; } const size_t offset = SkAlignTo(fOffset, fAlignment); fOffset = offset + requiredBytes; return {fBuffer.get(), offset}; } void ScratchBuffer::returnToPool() { if (fOwner && fBuffer) { // TODO: Generalize the pool to other buffer types. fOwner->fReusableScratchStorageBuffers.push_back(std::move(fBuffer)); SkASSERT(!fBuffer); } } // ------------------------------------------------------------------------------------------------ // DrawBufferManager DrawBufferManager::DrawBufferManager(ResourceProvider* resourceProvider, const Caps* caps, UploadBufferManager* uploadManager) : fResourceProvider(resourceProvider) , fCaps(caps) , fUploadManager(uploadManager) , fCurrentBuffers{{ { BufferType::kVertex, kVertexBufferSize, caps }, { BufferType::kIndex, kIndexBufferSize, caps }, { BufferType::kUniform, kUniformBufferSize, caps }, { BufferType::kStorage, kStorageBufferSize, caps }, // mapped storage { BufferType::kStorage, kStorageBufferSize, caps }, // GPU-only storage { BufferType::kVertexStorage, kVertexBufferSize, caps }, { BufferType::kIndexStorage, kIndexBufferSize, caps }, { BufferType::kIndirect, kStorageBufferSize, caps } }} {} DrawBufferManager::~DrawBufferManager() {} // For simplicity, if transfer buffers are being used, we align the data to the max alignment of // either the final buffer type or cpu->gpu transfer alignment so that the buffers are laid out // the same in memory. DrawBufferManager::BufferInfo::BufferInfo(BufferType type, size_t blockSize, const Caps* caps) : fType(type) , fStartAlignment(starting_alignment(type, !caps->drawBufferCanBeMapped(), caps)) , fBlockSize(SkAlignTo(blockSize, fStartAlignment)) {} std::pair DrawBufferManager::getVertexWriter(size_t requiredBytes) { if (!requiredBytes) { return {}; } auto& info = fCurrentBuffers[kVertexBufferIndex]; auto [ptr, bindInfo] = this->prepareMappedBindBuffer(&info, requiredBytes, "VertexBuffer"); return {VertexWriter(ptr, requiredBytes), bindInfo}; } void DrawBufferManager::returnVertexBytes(size_t unusedBytes) { if (fMappingFailed) { // The caller can be unaware that the written data went to no-where and will still call // this function. return; } SkASSERT(fCurrentBuffers[kVertexBufferIndex].fOffset >= unusedBytes); fCurrentBuffers[kVertexBufferIndex].fOffset -= unusedBytes; } std::pair DrawBufferManager::getIndexWriter(size_t requiredBytes) { if (!requiredBytes) { return {}; } auto& info = fCurrentBuffers[kIndexBufferIndex]; auto [ptr, bindInfo] = this->prepareMappedBindBuffer(&info, requiredBytes, "IndexBuffer"); return {IndexWriter(ptr, requiredBytes), bindInfo}; } std::pair DrawBufferManager::getUniformWriter(size_t requiredBytes) { if (!requiredBytes) { return {}; } auto& info = fCurrentBuffers[kUniformBufferIndex]; auto [ptr, bindInfo] = this->prepareMappedBindBuffer(&info, requiredBytes, "UniformBuffer"); return {UniformWriter(ptr, requiredBytes), bindInfo}; } std::pair DrawBufferManager::getSsboWriter(size_t requiredBytes) { if (!requiredBytes) { return {}; } auto& info = fCurrentBuffers[kStorageBufferIndex]; auto [ptr, bindInfo] = this->prepareMappedBindBuffer(&info, requiredBytes, "StorageBuffer"); return {UniformWriter(ptr, requiredBytes), bindInfo}; } std::pair DrawBufferManager::getUniformPointer( size_t requiredBytes) { if (!requiredBytes) { return {}; } auto& info = fCurrentBuffers[kUniformBufferIndex]; return this->prepareMappedBindBuffer(&info, requiredBytes, "UniformBuffer"); } std::pair DrawBufferManager::getStoragePointer( size_t requiredBytes) { if (!requiredBytes) { return {}; } auto& info = fCurrentBuffers[kStorageBufferIndex]; return this->prepareMappedBindBuffer(&info, requiredBytes, "StorageBuffer"); } BindBufferInfo DrawBufferManager::getStorage(size_t requiredBytes, ClearBuffer cleared) { if (!requiredBytes) { return {}; } auto& info = fCurrentBuffers[kGpuOnlyStorageBufferIndex]; return this->prepareBindBuffer(&info, requiredBytes, "StorageBuffer", /*supportCpuUpload=*/false, cleared); } BindBufferInfo DrawBufferManager::getVertexStorage(size_t requiredBytes) { if (!requiredBytes) { return {}; } auto& info = fCurrentBuffers[kVertexStorageBufferIndex]; return this->prepareBindBuffer(&info, requiredBytes, "VertexStorageBuffer"); } BindBufferInfo DrawBufferManager::getIndexStorage(size_t requiredBytes) { if (!requiredBytes) { return {}; } auto& info = fCurrentBuffers[kIndexStorageBufferIndex]; return this->prepareBindBuffer(&info, requiredBytes, "IndexStorageBuffer"); } BindBufferInfo DrawBufferManager::getIndirectStorage(size_t requiredBytes, ClearBuffer cleared) { if (!requiredBytes) { return {}; } auto& info = fCurrentBuffers[kIndirectStorageBufferIndex]; return this->prepareBindBuffer(&info, requiredBytes, "IndirectStorageBuffer", /*supportCpuUpload=*/false, cleared); } ScratchBuffer DrawBufferManager::getScratchStorage(size_t requiredBytes) { if (!requiredBytes || fMappingFailed) { return {}; } // TODO: Generalize the pool to other buffer types. auto& info = fCurrentBuffers[kStorageBufferIndex]; size_t bufferSize = sufficient_block_size(requiredBytes, info.fBlockSize); sk_sp buffer = this->findReusableSbo(bufferSize); if (!buffer) { buffer = fResourceProvider->findOrCreateBuffer( bufferSize, BufferType::kStorage, AccessPattern::kGpuOnly, "ScratchStorageBuffer"); if (!buffer) { this->onFailedBuffer(); return {}; } } return {requiredBytes, info.fStartAlignment, std::move(buffer), this}; } void DrawBufferManager::onFailedBuffer() { fMappingFailed = true; // Clean up and unmap everything now fReusableScratchStorageBuffers.clear(); for (auto& [buffer, _] : fUsedBuffers) { if (buffer->isMapped()) { buffer->unmap(); } } fUsedBuffers.clear(); for (auto& info : fCurrentBuffers) { if (info.fBuffer && info.fBuffer->isMapped()) { info.fBuffer->unmap(); } info.fBuffer = nullptr; info.fTransferBuffer = {}; info.fOffset = 0; } } void DrawBufferManager::transferToRecording(Recording* recording) { // We could allow this to be called when the mapping has failed, since the transfer will be a no // op, but in practice, the caller will want to check the error state as soon as possible to // limit any unnecessary resource preparation from other tasks. SkASSERT(!fMappingFailed); if (!fClearList.empty()) { recording->priv().addTask(ClearBuffersTask::Make(std::move(fClearList))); } // Transfer the buffers in the reuse pool to the recording. // TODO: Allow reuse across different Recordings? for (auto& buffer : fReusableScratchStorageBuffers) { recording->priv().addResourceRef(std::move(buffer)); } fReusableScratchStorageBuffers.clear(); for (auto& [buffer, transferBuffer] : fUsedBuffers) { if (transferBuffer) { SkASSERT(buffer); SkASSERT(!fCaps->drawBufferCanBeMapped()); // Since the transfer buffer is managed by the UploadManager, we don't manually unmap // it here or need to pass a ref into CopyBufferToBufferTask. size_t copySize = buffer->size(); recording->priv().addTask( CopyBufferToBufferTask::Make(transferBuffer.fBuffer, transferBuffer.fOffset, std::move(buffer), /*dstOffset=*/0, copySize)); } else { if (buffer->isMapped()) { buffer->unmap(); } recording->priv().addResourceRef(std::move(buffer)); } } fUsedBuffers.clear(); // The current draw buffers have not been added to fUsedBuffers, // so we need to handle them as well. for (auto& info : fCurrentBuffers) { if (!info.fBuffer) { continue; } if (info.fTransferBuffer) { // A transfer buffer should always be mapped at this stage SkASSERT(info.fBuffer); SkASSERT(!fCaps->drawBufferCanBeMapped()); // Since the transfer buffer is managed by the UploadManager, we don't manually unmap // it here or need to pass a ref into CopyBufferToBufferTask. recording->priv().addTask( CopyBufferToBufferTask::Make(info.fTransferBuffer.fBuffer, info.fTransferBuffer.fOffset, info.fBuffer, /*dstOffset=*/0, info.fBuffer->size())); } else { if (info.fBuffer->isMapped()) { info.fBuffer->unmap(); } recording->priv().addResourceRef(std::move(info.fBuffer)); } info.fTransferBuffer = {}; info.fOffset = 0; } } std::pair DrawBufferManager::prepareMappedBindBuffer( BufferInfo* info, size_t requiredBytes, std::string_view label) { BindBufferInfo bindInfo = this->prepareBindBuffer(info, requiredBytes, std::move(label), /*supportCpuUpload=*/true); if (!bindInfo) { // prepareBindBuffer() already called onFailedBuffer() SkASSERT(fMappingFailed); return {nullptr, {}}; } // If there's a transfer buffer, its mapped pointer should already have been validated SkASSERT(!info->fTransferBuffer || info->fTransferMapPtr); void* mapPtr = info->fTransferBuffer ? info->fTransferMapPtr : info->fBuffer->map(); if (!mapPtr) { // Mapping a direct draw buffer failed this->onFailedBuffer(); return {nullptr, {}}; } mapPtr = SkTAddOffset(mapPtr, static_cast(bindInfo.fOffset)); return {mapPtr, bindInfo}; } BindBufferInfo DrawBufferManager::prepareBindBuffer(BufferInfo* info, size_t requiredBytes, std::string_view label, bool supportCpuUpload, ClearBuffer cleared) { SkASSERT(info); SkASSERT(requiredBytes); if (fMappingFailed) { return {}; } // A transfer buffer is not necessary if the caller does not intend to upload CPU data to it. bool useTransferBuffer = supportCpuUpload && !fCaps->drawBufferCanBeMapped(); if (info->fBuffer && !can_fit(requiredBytes, info->fBuffer->size(), info->fOffset, info->fStartAlignment)) { fUsedBuffers.emplace_back(std::move(info->fBuffer), info->fTransferBuffer); info->fTransferBuffer = {}; } if (!info->fBuffer) { // This buffer can be GPU-only if // a) the caller does not intend to ever upload CPU data to the buffer; or // b) CPU data will get uploaded to fBuffer only via a transfer buffer AccessPattern accessPattern = (useTransferBuffer || !supportCpuUpload) ? AccessPattern::kGpuOnly : AccessPattern::kHostVisible; size_t bufferSize = sufficient_block_size(requiredBytes, info->fBlockSize); info->fBuffer = fResourceProvider->findOrCreateBuffer(bufferSize, info->fType, accessPattern, std::move(label)); info->fOffset = 0; if (!info->fBuffer) { this->onFailedBuffer(); return {}; } } if (useTransferBuffer && !info->fTransferBuffer) { std::tie(info->fTransferMapPtr, info->fTransferBuffer) = fUploadManager->makeBindInfo(info->fBuffer->size(), fCaps->requiredTransferBufferAlignment(), "TransferForDataBuffer"); if (!info->fTransferBuffer) { this->onFailedBuffer(); return {}; } SkASSERT(info->fTransferMapPtr); } info->fOffset = SkAlignTo(info->fOffset, info->fStartAlignment); BindBufferInfo bindInfo{info->fBuffer.get(), info->fOffset}; info->fOffset += requiredBytes; if (cleared == ClearBuffer::kYes) { fClearList.push_back({bindInfo.fBuffer, bindInfo.fOffset, requiredBytes}); } return bindInfo; } sk_sp DrawBufferManager::findReusableSbo(size_t bufferSize) { SkASSERT(bufferSize); SkASSERT(!fMappingFailed); for (int i = 0; i < fReusableScratchStorageBuffers.size(); ++i) { sk_sp* buffer = &fReusableScratchStorageBuffers[i]; if ((*buffer)->size() >= bufferSize) { auto found = std::move(*buffer); // Fill the hole left by the move (if necessary) and shrink the pool. if (i < fReusableScratchStorageBuffers.size() - 1) { *buffer = std::move(fReusableScratchStorageBuffers.back()); } fReusableScratchStorageBuffers.pop_back(); return found; } } return nullptr; } // ------------------------------------------------------------------------------------------------ // StaticBufferManager StaticBufferManager::StaticBufferManager(ResourceProvider* resourceProvider, const Caps* caps) : fResourceProvider(resourceProvider) , fUploadManager(resourceProvider, caps) , fRequiredTransferAlignment(caps->requiredTransferBufferAlignment()) , fVertexBufferInfo(BufferType::kVertex, caps) , fIndexBufferInfo(BufferType::kIndex, caps) {} StaticBufferManager::~StaticBufferManager() = default; StaticBufferManager::BufferInfo::BufferInfo(BufferType type, const Caps* caps) : fBufferType(type) , fAlignment(starting_alignment(type, /*useTransferBuffers=*/true, caps)) , fTotalRequiredBytes(0) {} VertexWriter StaticBufferManager::getVertexWriter(size_t size, BindBufferInfo* binding) { void* data = this->prepareStaticData(&fVertexBufferInfo, size, binding); return VertexWriter{data, size}; } VertexWriter StaticBufferManager::getIndexWriter(size_t size, BindBufferInfo* binding) { void* data = this->prepareStaticData(&fIndexBufferInfo, size, binding); return VertexWriter{data, size}; } void* StaticBufferManager::prepareStaticData(BufferInfo* info, size_t size, BindBufferInfo* target) { // Zero-out the target binding in the event of any failure in actually transfering data later. SkASSERT(target); *target = {nullptr, 0}; if (!size || fMappingFailed) { return nullptr; } // Both the transfer buffer and static buffers are aligned to the max required alignment for // the pair of buffer types involved (transfer cpu->gpu and either index or vertex). Copies // must also copy an aligned amount of bytes. size = SkAlignTo(size, info->fAlignment); auto [transferMapPtr, transferBindInfo] = fUploadManager.makeBindInfo(size, fRequiredTransferAlignment, "TransferForStaticBuffer"); if (!transferMapPtr) { SKGPU_LOG_E("Failed to create or map transfer buffer that initializes static GPU data."); fMappingFailed = true; return nullptr; } info->fData.push_back({transferBindInfo, target, size}); info->fTotalRequiredBytes += size; return transferMapPtr; } bool StaticBufferManager::BufferInfo::createAndUpdateBindings( ResourceProvider* resourceProvider, Context* context, QueueManager* queueManager, GlobalCache* globalCache, std::string_view label) const { if (!fTotalRequiredBytes) { return true; // No buffer needed } sk_sp staticBuffer = resourceProvider->findOrCreateBuffer( fTotalRequiredBytes, fBufferType, AccessPattern::kGpuOnly, std::move(label)); if (!staticBuffer) { SKGPU_LOG_E("Failed to create static buffer for type %d of size %zu bytes.\n", (int) fBufferType, fTotalRequiredBytes); return false; } size_t offset = 0; for (const CopyRange& data : fData) { // Each copy range's size should be aligned to the max of the required buffer alignment and // the transfer alignment, so we can just increment the offset into the static buffer. SkASSERT(offset % fAlignment == 0); data.fTarget->fBuffer = staticBuffer.get(); data.fTarget->fOffset = offset; auto copyTask = CopyBufferToBufferTask::Make( data.fSource.fBuffer, data.fSource.fOffset, sk_ref_sp(data.fTarget->fBuffer), data.fTarget->fOffset, data.fSize); if (!queueManager->addTask(copyTask.get(), context)) { SKGPU_LOG_E("Failed to copy data to static buffer.\n"); return false; } offset += data.fSize; } SkASSERT(offset == fTotalRequiredBytes); globalCache->addStaticResource(std::move(staticBuffer)); return true; } StaticBufferManager::FinishResult StaticBufferManager::finalize(Context* context, QueueManager* queueManager, GlobalCache* globalCache) { if (fMappingFailed) { return FinishResult::kFailure; } const size_t totalRequiredBytes = fVertexBufferInfo.fTotalRequiredBytes + fIndexBufferInfo.fTotalRequiredBytes; SkASSERT(totalRequiredBytes <= kMaxStaticDataSize); if (!totalRequiredBytes) { return FinishResult::kNoWork; } if (!fVertexBufferInfo.createAndUpdateBindings(fResourceProvider, context, queueManager, globalCache, "StaticVertexBuffer")) { return FinishResult::kFailure; } if (!fIndexBufferInfo.createAndUpdateBindings(fResourceProvider, context, queueManager, globalCache, "StaticIndexBuffer")) { return FinishResult::kFailure; } queueManager->addUploadBufferManagerRefs(&fUploadManager); // Reset the static buffer manager since the Recording's copy tasks now manage ownership of // the transfer buffers and the GlobalCache owns the final static buffers. fVertexBufferInfo.reset(); fIndexBufferInfo.reset(); return FinishResult::kSuccess; } } // namespace skgpu::graphite