Remove device dependencies from ringbuffer.

Allows ringbuffer sub-allocator to be used for non-staging memory.

BUG=dawn:155

Change-Id: Id0021907f520909aaebaf79e992124a47797d38d
Reviewed-on: https://dawn-review.googlesource.com/c/dawn/+/9760
Reviewed-by: Corentin Wallez <cwallez@chromium.org>
Reviewed-by: Austin Eng <enga@chromium.org>
Commit-Queue: Bryan Bernhart <bryan.bernhart@intel.com>
This commit is contained in:
Bryan Bernhart 2019-09-18 22:06:41 +00:00 committed by Commit Bot service account
parent b11bd2dfe5
commit 450e212cf5
17 changed files with 290 additions and 378 deletions

View File

@ -216,8 +216,8 @@ source_set("libdawn_native_sources") {
"src/dawn_native/ResourceHeap.h", "src/dawn_native/ResourceHeap.h",
"src/dawn_native/ResourceMemoryAllocation.cpp", "src/dawn_native/ResourceMemoryAllocation.cpp",
"src/dawn_native/ResourceMemoryAllocation.h", "src/dawn_native/ResourceMemoryAllocation.h",
"src/dawn_native/RingBuffer.cpp", "src/dawn_native/RingBufferAllocator.cpp",
"src/dawn_native/RingBuffer.h", "src/dawn_native/RingBufferAllocator.h",
"src/dawn_native/Sampler.cpp", "src/dawn_native/Sampler.cpp",
"src/dawn_native/Sampler.h", "src/dawn_native/Sampler.h",
"src/dawn_native/ShaderModule.cpp", "src/dawn_native/ShaderModule.cpp",
@ -769,7 +769,7 @@ test("dawn_unittests") {
"src/tests/unittests/PerStageTests.cpp", "src/tests/unittests/PerStageTests.cpp",
"src/tests/unittests/RefCountedTests.cpp", "src/tests/unittests/RefCountedTests.cpp",
"src/tests/unittests/ResultTests.cpp", "src/tests/unittests/ResultTests.cpp",
"src/tests/unittests/RingBufferTests.cpp", "src/tests/unittests/RingBufferAllocatorTests.cpp",
"src/tests/unittests/SerialMapTests.cpp", "src/tests/unittests/SerialMapTests.cpp",
"src/tests/unittests/SerialQueueTests.cpp", "src/tests/unittests/SerialQueueTests.cpp",
"src/tests/unittests/ToBackendTests.cpp", "src/tests/unittests/ToBackendTests.cpp",

View File

@ -167,9 +167,7 @@ namespace dawn_native {
// error buffer. // error buffer.
// TODO(enga): Suballocate and reuse memory from a larger staging buffer so we don't create // TODO(enga): Suballocate and reuse memory from a larger staging buffer so we don't create
// many small buffers. // many small buffers.
DynamicUploader* uploader = nullptr; DAWN_TRY_ASSIGN(mStagingBuffer, GetDevice()->CreateStagingBuffer(GetSize()));
DAWN_TRY_ASSIGN(uploader, GetDevice()->GetDynamicUploader());
DAWN_TRY_ASSIGN(mStagingBuffer, uploader->CreateStagingBuffer(GetSize()));
ASSERT(mStagingBuffer->GetMappedPointer() != nullptr); ASSERT(mStagingBuffer->GetMappedPointer() != nullptr);
*mappedPointer = reinterpret_cast<uint8_t*>(mStagingBuffer->GetMappedPointer()); *mappedPointer = reinterpret_cast<uint8_t*>(mStagingBuffer->GetMappedPointer());
@ -252,11 +250,11 @@ namespace dawn_native {
} }
MaybeError BufferBase::SetSubDataImpl(uint32_t start, uint32_t count, const void* data) { MaybeError BufferBase::SetSubDataImpl(uint32_t start, uint32_t count, const void* data) {
DynamicUploader* uploader = nullptr; DynamicUploader* uploader = GetDevice()->GetDynamicUploader();
DAWN_TRY_ASSIGN(uploader, GetDevice()->GetDynamicUploader());
UploadHandle uploadHandle; UploadHandle uploadHandle;
DAWN_TRY_ASSIGN(uploadHandle, uploader->Allocate(count)); DAWN_TRY_ASSIGN(uploadHandle,
uploader->Allocate(count, GetDevice()->GetPendingCommandSerial()));
ASSERT(uploadHandle.mappedBuffer != nullptr); ASSERT(uploadHandle.mappedBuffer != nullptr);
memcpy(uploadHandle.mappedBuffer, data, count); memcpy(uploadHandle.mappedBuffer, data, count);
@ -311,8 +309,7 @@ namespace dawn_native {
ASSERT(mStagingBuffer); ASSERT(mStagingBuffer);
DAWN_TRY(GetDevice()->CopyFromStagingToBuffer(mStagingBuffer.get(), 0, this, 0, GetSize())); DAWN_TRY(GetDevice()->CopyFromStagingToBuffer(mStagingBuffer.get(), 0, this, 0, GetSize()));
DynamicUploader* uploader = nullptr; DynamicUploader* uploader = GetDevice()->GetDynamicUploader();
DAWN_TRY_ASSIGN(uploader, GetDevice()->GetDynamicUploader());
uploader->ReleaseStagingBuffer(std::move(mStagingBuffer)); uploader->ReleaseStagingBuffer(std::move(mStagingBuffer));
return {}; return {};

View File

@ -704,10 +704,7 @@ namespace dawn_native {
// Other implementation details // Other implementation details
ResultOrError<DynamicUploader*> DeviceBase::GetDynamicUploader() const { DynamicUploader* DeviceBase::GetDynamicUploader() const {
if (mDynamicUploader->IsEmpty()) {
DAWN_TRY(mDynamicUploader->CreateAndAppendBuffer());
}
return mDynamicUploader.get(); return mDynamicUploader.get();
} }

View File

@ -167,7 +167,7 @@ namespace dawn_native {
uint64_t destinationOffset, uint64_t destinationOffset,
uint64_t size) = 0; uint64_t size) = 0;
ResultOrError<DynamicUploader*> GetDynamicUploader() const; DynamicUploader* GetDynamicUploader() const;
std::vector<const char*> GetEnabledExtensions() const; std::vector<const char*> GetEnabledExtensions() const;
std::vector<const char*> GetTogglesUsed() const; std::vector<const char*> GetTogglesUsed() const;

View File

@ -18,15 +18,9 @@
namespace dawn_native { namespace dawn_native {
DynamicUploader::DynamicUploader(DeviceBase* device) : mDevice(device) { DynamicUploader::DynamicUploader(DeviceBase* device, size_t size) : mDevice(device) {
} mRingBuffers.emplace_back(
std::unique_ptr<RingBuffer>(new RingBuffer{nullptr, RingBufferAllocator(size)}));
ResultOrError<std::unique_ptr<StagingBufferBase>> DynamicUploader::CreateStagingBuffer(
size_t size) {
std::unique_ptr<StagingBufferBase> stagingBuffer;
DAWN_TRY_ASSIGN(stagingBuffer, mDevice->CreateStagingBuffer(size));
DAWN_TRY(stagingBuffer->Initialize());
return stagingBuffer;
} }
void DynamicUploader::ReleaseStagingBuffer(std::unique_ptr<StagingBufferBase> stagingBuffer) { void DynamicUploader::ReleaseStagingBuffer(std::unique_ptr<StagingBufferBase> stagingBuffer) {
@ -34,73 +28,78 @@ namespace dawn_native {
mDevice->GetPendingCommandSerial()); mDevice->GetPendingCommandSerial());
} }
MaybeError DynamicUploader::CreateAndAppendBuffer(size_t size) { ResultOrError<UploadHandle> DynamicUploader::Allocate(size_t allocationSize, Serial serial) {
std::unique_ptr<RingBuffer> ringBuffer = std::make_unique<RingBuffer>(mDevice, size);
DAWN_TRY(ringBuffer->Initialize());
mRingBuffers.emplace_back(std::move(ringBuffer));
return {};
}
ResultOrError<UploadHandle> DynamicUploader::Allocate(uint32_t size) {
// Note: Validation ensures size is already aligned. // Note: Validation ensures size is already aligned.
// First-fit: find next smallest buffer large enough to satisfy the allocation request. // First-fit: find next smallest buffer large enough to satisfy the allocation request.
RingBuffer* targetRingBuffer = GetLargestBuffer(); RingBuffer* targetRingBuffer = mRingBuffers.back().get();
for (auto& ringBuffer : mRingBuffers) { for (auto& ringBuffer : mRingBuffers) {
const RingBufferAllocator& ringBufferAllocator = ringBuffer->mAllocator;
// Prevent overflow. // Prevent overflow.
ASSERT(ringBuffer->GetSize() >= ringBuffer->GetUsedSize()); ASSERT(ringBufferAllocator.GetSize() >= ringBufferAllocator.GetUsedSize());
const size_t remainingSize = ringBuffer->GetSize() - ringBuffer->GetUsedSize(); const size_t remainingSize =
if (size <= remainingSize) { ringBufferAllocator.GetSize() - ringBufferAllocator.GetUsedSize();
if (allocationSize <= remainingSize) {
targetRingBuffer = ringBuffer.get(); targetRingBuffer = ringBuffer.get();
break; break;
} }
} }
UploadHandle uploadHandle = UploadHandle{}; size_t startOffset = kInvalidOffset;
if (targetRingBuffer != nullptr) { if (targetRingBuffer != nullptr) {
uploadHandle = targetRingBuffer->SubAllocate(size); startOffset = targetRingBuffer->mAllocator.Allocate(allocationSize, serial);
} }
// Upon failure, append a newly created (and much larger) ring buffer to fulfill the // Upon failure, append a newly created (and much larger) ring buffer to fulfill the
// request. // request.
if (uploadHandle.mappedBuffer == nullptr) { if (startOffset == kInvalidOffset) {
// Compute the new max size (in powers of two to preserve alignment). // Compute the new max size (in powers of two to preserve alignment).
size_t newMaxSize = targetRingBuffer->GetSize() * 2; size_t newMaxSize = targetRingBuffer->mAllocator.GetSize() * 2;
while (newMaxSize < size) { while (newMaxSize < allocationSize) {
newMaxSize *= 2; newMaxSize *= 2;
} }
// TODO(bryan.bernhart@intel.com): Fall-back to no sub-allocations should this fail. // TODO(bryan.bernhart@intel.com): Fall-back to no sub-allocations should this fail.
DAWN_TRY(CreateAndAppendBuffer(newMaxSize)); mRingBuffers.emplace_back(std::unique_ptr<RingBuffer>(
targetRingBuffer = GetLargestBuffer(); new RingBuffer{nullptr, RingBufferAllocator(newMaxSize)}));
uploadHandle = targetRingBuffer->SubAllocate(size);
targetRingBuffer = mRingBuffers.back().get();
startOffset = targetRingBuffer->mAllocator.Allocate(allocationSize, serial);
} }
uploadHandle.stagingBuffer = targetRingBuffer->GetStagingBuffer(); ASSERT(startOffset != kInvalidOffset);
// Allocate the staging buffer backing the ringbuffer.
// Note: the first ringbuffer will be lazily created.
if (targetRingBuffer->mStagingBuffer == nullptr) {
std::unique_ptr<StagingBufferBase> stagingBuffer;
DAWN_TRY_ASSIGN(stagingBuffer,
mDevice->CreateStagingBuffer(targetRingBuffer->mAllocator.GetSize()));
targetRingBuffer->mStagingBuffer = std::move(stagingBuffer);
}
ASSERT(targetRingBuffer->mStagingBuffer != nullptr);
UploadHandle uploadHandle;
uploadHandle.stagingBuffer = targetRingBuffer->mStagingBuffer.get();
uploadHandle.mappedBuffer =
static_cast<uint8_t*>(uploadHandle.stagingBuffer->GetMappedPointer()) + startOffset;
uploadHandle.startOffset = startOffset;
return uploadHandle; return uploadHandle;
} }
void DynamicUploader::Tick(Serial lastCompletedSerial) { void DynamicUploader::Deallocate(Serial lastCompletedSerial) {
// Reclaim memory within the ring buffers by ticking (or removing requests no longer // Reclaim memory within the ring buffers by ticking (or removing requests no longer
// in-flight). // in-flight).
for (size_t i = 0; i < mRingBuffers.size(); ++i) { for (size_t i = 0; i < mRingBuffers.size(); ++i) {
mRingBuffers[i]->Tick(lastCompletedSerial); mRingBuffers[i]->mAllocator.Deallocate(lastCompletedSerial);
// Never erase the last buffer as to prevent re-creating smaller buffers // Never erase the last buffer as to prevent re-creating smaller buffers
// again. The last buffer is the largest. // again. The last buffer is the largest.
if (mRingBuffers[i]->Empty() && i < mRingBuffers.size() - 1) { if (mRingBuffers[i]->mAllocator.Empty() && i < mRingBuffers.size() - 1) {
mRingBuffers.erase(mRingBuffers.begin() + i); mRingBuffers.erase(mRingBuffers.begin() + i);
} }
} }
mReleasedStagingBuffers.ClearUpTo(lastCompletedSerial); mReleasedStagingBuffers.ClearUpTo(lastCompletedSerial);
} }
RingBuffer* DynamicUploader::GetLargestBuffer() {
ASSERT(!mRingBuffers.empty());
return mRingBuffers.back().get();
}
bool DynamicUploader::IsEmpty() const {
return mRingBuffers.empty();
}
} // namespace dawn_native } // namespace dawn_native

View File

@ -16,37 +16,42 @@
#define DAWNNATIVE_DYNAMICUPLOADER_H_ #define DAWNNATIVE_DYNAMICUPLOADER_H_
#include "dawn_native/Forward.h" #include "dawn_native/Forward.h"
#include "dawn_native/RingBuffer.h" #include "dawn_native/RingBufferAllocator.h"
#include "dawn_native/StagingBuffer.h"
// DynamicUploader is the front-end implementation used to manage multiple ring buffers for upload // DynamicUploader is the front-end implementation used to manage multiple ring buffers for upload
// usage. // usage.
namespace dawn_native { namespace dawn_native {
struct UploadHandle {
uint8_t* mappedBuffer = nullptr;
size_t startOffset = 0;
StagingBufferBase* stagingBuffer = nullptr;
};
class DynamicUploader { class DynamicUploader {
public: public:
DynamicUploader(DeviceBase* device); DynamicUploader(DeviceBase* device, size_t size = kBaseUploadBufferSize);
~DynamicUploader() = default; ~DynamicUploader() = default;
// We add functions to Create/Release StagingBuffers to the DynamicUploader as there's // We add functions to Release StagingBuffers to the DynamicUploader as there's
// currently no place to track the allocated staging buffers such that they're freed after // currently no place to track the allocated staging buffers such that they're freed after
// pending commands are finished. This should be changed when better resource allocation is // pending commands are finished. This should be changed when better resource allocation is
// implemented. // implemented.
ResultOrError<std::unique_ptr<StagingBufferBase>> CreateStagingBuffer(size_t size);
void ReleaseStagingBuffer(std::unique_ptr<StagingBufferBase> stagingBuffer); void ReleaseStagingBuffer(std::unique_ptr<StagingBufferBase> stagingBuffer);
ResultOrError<UploadHandle> Allocate(uint32_t size); ResultOrError<UploadHandle> Allocate(size_t allocationSize, Serial serial);
void Tick(Serial lastCompletedSerial); void Deallocate(Serial lastCompletedSerial);
RingBuffer* GetLargestBuffer();
MaybeError CreateAndAppendBuffer(size_t size = kBaseUploadBufferSize);
bool IsEmpty() const;
private: private:
// TODO(bryan.bernhart@intel.com): Figure out this value. // TODO(bryan.bernhart@intel.com): Figure out this value.
static constexpr size_t kBaseUploadBufferSize = 64000; static constexpr size_t kBaseUploadBufferSize = 64000;
struct RingBuffer {
std::unique_ptr<StagingBufferBase> mStagingBuffer;
RingBufferAllocator mAllocator;
};
std::vector<std::unique_ptr<RingBuffer>> mRingBuffers; std::vector<std::unique_ptr<RingBuffer>> mRingBuffers;
SerialQueue<std::unique_ptr<StagingBufferBase>> mReleasedStagingBuffers; SerialQueue<std::unique_ptr<StagingBufferBase>> mReleasedStagingBuffers;
DeviceBase* mDevice; DeviceBase* mDevice;

View File

@ -12,13 +12,10 @@
// See the License for the specific language governing permissions and // See the License for the specific language governing permissions and
// limitations under the License. // limitations under the License.
#include "dawn_native/RingBuffer.h" #include "dawn_native/RingBufferAllocator.h"
#include "dawn_native/Device.h"
#include <limits> // Note: Current RingBufferAllocator implementation uses two indices (start and end) to implement a
// circular queue. However, this approach defines a full queue when one element is still unused.
// Note: Current RingBuffer implementation uses two indices (start and end) to implement a circular
// queue. However, this approach defines a full queue when one element is still unused.
// //
// For example, [E,E,E,E] would be equivelent to [U,U,U,U]. // For example, [E,E,E,E] would be equivelent to [U,U,U,U].
// ^ ^ // ^ ^
@ -32,36 +29,10 @@
// TODO(bryan.bernhart@intel.com): Follow-up with ringbuffer optimization. // TODO(bryan.bernhart@intel.com): Follow-up with ringbuffer optimization.
namespace dawn_native { namespace dawn_native {
static constexpr size_t INVALID_OFFSET = std::numeric_limits<size_t>::max(); RingBufferAllocator::RingBufferAllocator(size_t maxSize) : mMaxBlockSize(maxSize) {
RingBuffer::RingBuffer(DeviceBase* device, size_t size) : mBufferSize(size), mDevice(device) {
} }
MaybeError RingBuffer::Initialize() { void RingBufferAllocator::Deallocate(Serial lastCompletedSerial) {
DAWN_TRY_ASSIGN(mStagingBuffer, mDevice->CreateStagingBuffer(mBufferSize));
DAWN_TRY(mStagingBuffer->Initialize());
return {};
}
// Record allocations in a request when serial advances.
// This method has been split from Tick() for testing.
void RingBuffer::Track() {
if (mCurrentRequestSize == 0)
return;
const Serial currentSerial = mDevice->GetPendingCommandSerial();
if (mInflightRequests.Empty() || currentSerial > mInflightRequests.LastSerial()) {
Request request;
request.endOffset = mUsedEndOffset;
request.size = mCurrentRequestSize;
mInflightRequests.Enqueue(std::move(request), currentSerial);
mCurrentRequestSize = 0; // reset
}
}
void RingBuffer::Tick(Serial lastCompletedSerial) {
Track();
// Reclaim memory from previously recorded blocks. // Reclaim memory from previously recorded blocks.
for (Request& request : mInflightRequests.IterateUpTo(lastCompletedSerial)) { for (Request& request : mInflightRequests.IterateUpTo(lastCompletedSerial)) {
mUsedStartOffset = request.endOffset; mUsedStartOffset = request.endOffset;
@ -72,23 +43,18 @@ namespace dawn_native {
mInflightRequests.ClearUpTo(lastCompletedSerial); mInflightRequests.ClearUpTo(lastCompletedSerial);
} }
size_t RingBuffer::GetSize() const { size_t RingBufferAllocator::GetSize() const {
return mBufferSize; return mMaxBlockSize;
} }
size_t RingBuffer::GetUsedSize() const { size_t RingBufferAllocator::GetUsedSize() const {
return mUsedSize; return mUsedSize;
} }
bool RingBuffer::Empty() const { bool RingBufferAllocator::Empty() const {
return mInflightRequests.Empty(); return mInflightRequests.Empty();
} }
StagingBufferBase* RingBuffer::GetStagingBuffer() const {
ASSERT(mStagingBuffer != nullptr);
return mStagingBuffer.get();
}
// Sub-allocate the ring-buffer by requesting a chunk of the specified size. // Sub-allocate the ring-buffer by requesting a chunk of the specified size.
// This is a serial-based resource scheme, the life-span of resources (and the allocations) get // This is a serial-based resource scheme, the life-span of resources (and the allocations) get
// tracked by GPU progress via serials. Memory can be reused by determining if the GPU has // tracked by GPU progress via serials. Memory can be reused by determining if the GPU has
@ -96,55 +62,55 @@ namespace dawn_native {
// queue, which identifies an existing (or new) frames-worth of resources. Internally, the // queue, which identifies an existing (or new) frames-worth of resources. Internally, the
// ring-buffer maintains offsets of 3 "memory" states: Free, Reclaimed, and Used. This is done // ring-buffer maintains offsets of 3 "memory" states: Free, Reclaimed, and Used. This is done
// in FIFO order as older frames would free resources before newer ones. // in FIFO order as older frames would free resources before newer ones.
UploadHandle RingBuffer::SubAllocate(size_t allocSize) { size_t RingBufferAllocator::Allocate(size_t allocationSize, Serial serial) {
ASSERT(mStagingBuffer != nullptr);
// Check if the buffer is full by comparing the used size. // Check if the buffer is full by comparing the used size.
// If the buffer is not split where waste occurs (e.g. cannot fit new sub-alloc in front), a // If the buffer is not split where waste occurs (e.g. cannot fit new sub-alloc in front), a
// subsequent sub-alloc could fail where the used size was previously adjusted to include // subsequent sub-alloc could fail where the used size was previously adjusted to include
// the wasted. // the wasted.
if (mUsedSize >= mBufferSize) if (allocationSize == 0 || mUsedSize >= mMaxBlockSize) {
return UploadHandle{}; return kInvalidOffset;
}
size_t startOffset = INVALID_OFFSET; size_t startOffset = kInvalidOffset;
// Check if the buffer is NOT split (i.e sub-alloc on ends) // Check if the buffer is NOT split (i.e sub-alloc on ends)
if (mUsedStartOffset <= mUsedEndOffset) { if (mUsedStartOffset <= mUsedEndOffset) {
// Order is important (try to sub-alloc at end first). // Order is important (try to sub-alloc at end first).
// This is due to FIFO order where sub-allocs are inserted from left-to-right (when not // This is due to FIFO order where sub-allocs are inserted from left-to-right (when not
// wrapped). // wrapped).
if (mUsedEndOffset + allocSize <= mBufferSize) { if (mUsedEndOffset + allocationSize <= mMaxBlockSize) {
startOffset = mUsedEndOffset; startOffset = mUsedEndOffset;
mUsedEndOffset += allocSize; mUsedEndOffset += allocationSize;
mUsedSize += allocSize; mUsedSize += allocationSize;
mCurrentRequestSize += allocSize; mCurrentRequestSize += allocationSize;
} else if (allocSize <= mUsedStartOffset) { // Try to sub-alloc at front. } else if (allocationSize <= mUsedStartOffset) { // Try to sub-alloc at front.
// Count the space at front in the request size so that a subsequent // Count the space at the end so that a subsequent
// sub-alloc cannot not succeed when the buffer is full. // sub-alloc cannot not succeed when the buffer is full.
const size_t requestSize = (mBufferSize - mUsedEndOffset) + allocSize; const size_t requestSize = (mMaxBlockSize - mUsedEndOffset) + allocationSize;
startOffset = 0; startOffset = 0;
mUsedEndOffset = allocSize; mUsedEndOffset = allocationSize;
mUsedSize += requestSize; mUsedSize += requestSize;
mCurrentRequestSize += requestSize; mCurrentRequestSize += requestSize;
} }
} else if (mUsedEndOffset + allocSize <= } else if (mUsedEndOffset + allocationSize <=
mUsedStartOffset) { // Otherwise, buffer is split where sub-alloc must be mUsedStartOffset) { // Otherwise, buffer is split where sub-alloc must be
// in-between. // in-between.
startOffset = mUsedEndOffset; startOffset = mUsedEndOffset;
mUsedEndOffset += allocSize; mUsedEndOffset += allocationSize;
mUsedSize += allocSize; mUsedSize += allocationSize;
mCurrentRequestSize += allocSize; mCurrentRequestSize += allocationSize;
} }
if (startOffset == INVALID_OFFSET) if (startOffset != kInvalidOffset) {
return UploadHandle{}; Request request;
request.endOffset = mUsedEndOffset;
request.size = mCurrentRequestSize;
UploadHandle uploadHandle; mInflightRequests.Enqueue(std::move(request), serial);
uploadHandle.mappedBuffer = mCurrentRequestSize = 0; // reset
static_cast<uint8_t*>(mStagingBuffer->GetMappedPointer()) + startOffset; }
uploadHandle.startOffset = startOffset;
return uploadHandle; return startOffset;
} }
} // namespace dawn_native } // namespace dawn_native

View File

@ -12,46 +12,32 @@
// See the License for the specific language governing permissions and // See the License for the specific language governing permissions and
// limitations under the License. // limitations under the License.
#ifndef DAWNNATIVE_RINGBUFFER_H_ #ifndef DAWNNATIVE_RINGBUFFERALLOCATOR_H_
#define DAWNNATIVE_RINGBUFFER_H_ #define DAWNNATIVE_RINGBUFFERALLOCATOR_H_
#include "common/SerialQueue.h" #include "common/SerialQueue.h"
#include "dawn_native/StagingBuffer.h"
#include <limits>
#include <memory> #include <memory>
// RingBuffer is the front-end implementation used to manage a ring buffer in GPU memory. // RingBufferAllocator is the front-end implementation used to manage a ring buffer in GPU memory.
namespace dawn_native { namespace dawn_native {
struct UploadHandle { static constexpr size_t kInvalidOffset = std::numeric_limits<size_t>::max();
uint8_t* mappedBuffer = nullptr;
size_t startOffset = 0;
StagingBufferBase* stagingBuffer = nullptr;
};
class DeviceBase; class RingBufferAllocator {
class RingBuffer {
public: public:
RingBuffer(DeviceBase* device, size_t size); RingBufferAllocator(size_t maxSize);
~RingBuffer() = default; ~RingBufferAllocator() = default;
MaybeError Initialize(); size_t Allocate(size_t allocationSize, Serial serial);
void Deallocate(Serial lastCompletedSerial);
UploadHandle SubAllocate(size_t requestedSize);
void Tick(Serial lastCompletedSerial);
size_t GetSize() const; size_t GetSize() const;
bool Empty() const; bool Empty() const;
size_t GetUsedSize() const; size_t GetUsedSize() const;
StagingBufferBase* GetStagingBuffer() const;
// Seperated for testing.
void Track();
private: private:
std::unique_ptr<StagingBufferBase> mStagingBuffer;
struct Request { struct Request {
size_t endOffset; size_t endOffset;
size_t size; size_t size;
@ -62,13 +48,11 @@ namespace dawn_native {
size_t mUsedEndOffset = 0; // Tail of used sub-alloc requests (in bytes). size_t mUsedEndOffset = 0; // Tail of used sub-alloc requests (in bytes).
size_t mUsedStartOffset = 0; // Head of used sub-alloc requests (in bytes). size_t mUsedStartOffset = 0; // Head of used sub-alloc requests (in bytes).
size_t mBufferSize = 0; // Max size of the ring buffer (in bytes). size_t mMaxBlockSize = 0; // Max size of the ring buffer (in bytes).
size_t mUsedSize = 0; // Size of the sub-alloc requests (in bytes) of the ring buffer. size_t mUsedSize = 0; // Size of the sub-alloc requests (in bytes) of the ring buffer.
size_t mCurrentRequestSize = size_t mCurrentRequestSize =
0; // Size of the sub-alloc requests (in bytes) of the current serial. 0; // Size of the sub-alloc requests (in bytes) of the current serial.
DeviceBase* mDevice;
}; };
} // namespace dawn_native } // namespace dawn_native
#endif // DAWNNATIVE_RINGBUFFER_H_ #endif // DAWNNATIVE_RINGBUFFERALLOCATOR_H_

View File

@ -216,8 +216,8 @@ namespace dawn_native { namespace d3d12 {
mCompletedSerial = mFence->GetCompletedValue(); mCompletedSerial = mFence->GetCompletedValue();
// Uploader should tick before the resource allocator // Uploader should tick before the resource allocator
// as it enqueues resources to be released. // as it enqueued resources to be released.
mDynamicUploader->Tick(mCompletedSerial); mDynamicUploader->Deallocate(mCompletedSerial);
mResourceAllocator->Tick(mCompletedSerial); mResourceAllocator->Tick(mCompletedSerial);
mCommandAllocatorManager->Tick(mCompletedSerial); mCommandAllocatorManager->Tick(mCompletedSerial);
@ -319,6 +319,7 @@ namespace dawn_native { namespace d3d12 {
ResultOrError<std::unique_ptr<StagingBufferBase>> Device::CreateStagingBuffer(size_t size) { ResultOrError<std::unique_ptr<StagingBufferBase>> Device::CreateStagingBuffer(size_t size) {
std::unique_ptr<StagingBufferBase> stagingBuffer = std::unique_ptr<StagingBufferBase> stagingBuffer =
std::make_unique<StagingBuffer>(size, this); std::make_unique<StagingBuffer>(size, this);
DAWN_TRY(stagingBuffer->Initialize());
return std::move(stagingBuffer); return std::move(stagingBuffer);
} }

View File

@ -538,10 +538,10 @@ namespace dawn_native { namespace d3d12 {
return DAWN_OUT_OF_MEMORY_ERROR("Unable to allocate buffer."); return DAWN_OUT_OF_MEMORY_ERROR("Unable to allocate buffer.");
} }
uint32_t bufferSize = static_cast<uint32_t>(bufferSize64); uint32_t bufferSize = static_cast<uint32_t>(bufferSize64);
DynamicUploader* uploader = nullptr; DynamicUploader* uploader = device->GetDynamicUploader();
DAWN_TRY_ASSIGN(uploader, device->GetDynamicUploader());
UploadHandle uploadHandle; UploadHandle uploadHandle;
DAWN_TRY_ASSIGN(uploadHandle, uploader->Allocate(bufferSize)); DAWN_TRY_ASSIGN(uploadHandle,
uploader->Allocate(bufferSize, device->GetPendingCommandSerial()));
std::fill(reinterpret_cast<uint32_t*>(uploadHandle.mappedBuffer), std::fill(reinterpret_cast<uint32_t*>(uploadHandle.mappedBuffer),
reinterpret_cast<uint32_t*>(uploadHandle.mappedBuffer + bufferSize), reinterpret_cast<uint32_t*>(uploadHandle.mappedBuffer + bufferSize),
clearColor); clearColor);

View File

@ -154,7 +154,7 @@ namespace dawn_native { namespace metal {
void Device::TickImpl() { void Device::TickImpl() {
Serial completedSerial = GetCompletedCommandSerial(); Serial completedSerial = GetCompletedCommandSerial();
mDynamicUploader->Tick(completedSerial); mDynamicUploader->Deallocate(completedSerial);
mMapTracker->Tick(completedSerial); mMapTracker->Tick(completedSerial);
if (mPendingCommands != nil) { if (mPendingCommands != nil) {
@ -239,6 +239,7 @@ namespace dawn_native { namespace metal {
ResultOrError<std::unique_ptr<StagingBufferBase>> Device::CreateStagingBuffer(size_t size) { ResultOrError<std::unique_ptr<StagingBufferBase>> Device::CreateStagingBuffer(size_t size) {
std::unique_ptr<StagingBufferBase> stagingBuffer = std::unique_ptr<StagingBufferBase> stagingBuffer =
std::make_unique<StagingBuffer>(size, this); std::make_unique<StagingBuffer>(size, this);
DAWN_TRY(stagingBuffer->Initialize());
return std::move(stagingBuffer); return std::move(stagingBuffer);
} }

View File

@ -150,6 +150,7 @@ namespace dawn_native { namespace null {
ResultOrError<std::unique_ptr<StagingBufferBase>> Device::CreateStagingBuffer(size_t size) { ResultOrError<std::unique_ptr<StagingBufferBase>> Device::CreateStagingBuffer(size_t size) {
std::unique_ptr<StagingBufferBase> stagingBuffer = std::unique_ptr<StagingBufferBase> stagingBuffer =
std::make_unique<StagingBuffer>(size, this); std::make_unique<StagingBuffer>(size, this);
DAWN_TRY(stagingBuffer->Initialize());
return std::move(stagingBuffer); return std::move(stagingBuffer);
} }

View File

@ -26,7 +26,7 @@
#include "dawn_native/PipelineLayout.h" #include "dawn_native/PipelineLayout.h"
#include "dawn_native/Queue.h" #include "dawn_native/Queue.h"
#include "dawn_native/RenderPipeline.h" #include "dawn_native/RenderPipeline.h"
#include "dawn_native/RingBuffer.h" #include "dawn_native/RingBufferAllocator.h"
#include "dawn_native/Sampler.h" #include "dawn_native/Sampler.h"
#include "dawn_native/ShaderModule.h" #include "dawn_native/ShaderModule.h"
#include "dawn_native/StagingBuffer.h" #include "dawn_native/StagingBuffer.h"

View File

@ -217,7 +217,7 @@ namespace dawn_native { namespace vulkan {
// Uploader should tick before the resource allocator // Uploader should tick before the resource allocator
// as it enqueues resources to be released. // as it enqueues resources to be released.
mDynamicUploader->Tick(mCompletedSerial); mDynamicUploader->Deallocate(mCompletedSerial);
mMemoryAllocator->Tick(mCompletedSerial); mMemoryAllocator->Tick(mCompletedSerial);
@ -561,6 +561,7 @@ namespace dawn_native { namespace vulkan {
ResultOrError<std::unique_ptr<StagingBufferBase>> Device::CreateStagingBuffer(size_t size) { ResultOrError<std::unique_ptr<StagingBufferBase>> Device::CreateStagingBuffer(size_t size) {
std::unique_ptr<StagingBufferBase> stagingBuffer = std::unique_ptr<StagingBufferBase> stagingBuffer =
std::make_unique<StagingBuffer>(size, this); std::make_unique<StagingBuffer>(size, this);
DAWN_TRY(stagingBuffer->Initialize());
return std::move(stagingBuffer); return std::move(stagingBuffer);
} }

View File

@ -687,10 +687,10 @@ namespace dawn_native { namespace vulkan {
return DAWN_OUT_OF_MEMORY_ERROR("Unable to allocate buffer."); return DAWN_OUT_OF_MEMORY_ERROR("Unable to allocate buffer.");
} }
uint32_t bufferSize = static_cast<uint32_t>(bufferSize64); uint32_t bufferSize = static_cast<uint32_t>(bufferSize64);
DynamicUploader* uploader = nullptr; DynamicUploader* uploader = device->GetDynamicUploader();
DAWN_TRY_ASSIGN(uploader, device->GetDynamicUploader());
UploadHandle uploadHandle; UploadHandle uploadHandle;
DAWN_TRY_ASSIGN(uploadHandle, uploader->Allocate(bufferSize)); DAWN_TRY_ASSIGN(uploadHandle,
uploader->Allocate(bufferSize, device->GetPendingCommandSerial()));
std::fill(reinterpret_cast<uint32_t*>(uploadHandle.mappedBuffer), std::fill(reinterpret_cast<uint32_t*>(uploadHandle.mappedBuffer),
reinterpret_cast<uint32_t*>(uploadHandle.mappedBuffer + bufferSize), reinterpret_cast<uint32_t*>(uploadHandle.mappedBuffer + bufferSize),
clearColor); clearColor);

View File

@ -0,0 +1,161 @@
// Copyright 2018 The Dawn Authors
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include <gtest/gtest.h>
#include "dawn_native/RingBufferAllocator.h"
using namespace dawn_native;
// Number of basic tests for Ringbuffer
TEST(RingBufferAllocatorTests, BasicTest) {
constexpr size_t sizeInBytes = 64000;
RingBufferAllocator allocator(sizeInBytes);
// Ensure no requests exist on empty buffer.
EXPECT_TRUE(allocator.Empty());
ASSERT_EQ(allocator.GetSize(), sizeInBytes);
// Ensure failure upon sub-allocating an oversized request.
ASSERT_EQ(allocator.Allocate(sizeInBytes + 1, 0), kInvalidOffset);
// Fill the entire buffer with two requests of equal size.
ASSERT_EQ(allocator.Allocate(sizeInBytes / 2, 1), 0u);
ASSERT_EQ(allocator.Allocate(sizeInBytes / 2, 2), 32000u);
// Ensure the buffer is full.
ASSERT_EQ(allocator.Allocate(1, 3), kInvalidOffset);
}
// Tests that several ringbuffer allocations do not fail.
TEST(RingBufferAllocatorTests, RingBufferManyAlloc) {
constexpr size_t maxNumOfFrames = 64000;
constexpr size_t frameSizeInBytes = 4;
RingBufferAllocator allocator(maxNumOfFrames * frameSizeInBytes);
size_t offset = 0;
for (size_t i = 0; i < maxNumOfFrames; ++i) {
offset = allocator.Allocate(frameSizeInBytes, i);
ASSERT_EQ(offset, i * frameSizeInBytes);
}
}
// Tests ringbuffer sub-allocations of the same serial are correctly tracked.
TEST(RingBufferAllocatorTests, AllocInSameFrame) {
constexpr size_t maxNumOfFrames = 3;
constexpr size_t frameSizeInBytes = 4;
RingBufferAllocator allocator(maxNumOfFrames * frameSizeInBytes);
// F1
// [xxxx|--------]
size_t offset = allocator.Allocate(frameSizeInBytes, 1);
// F1 F2
// [xxxx|xxxx|----]
offset = allocator.Allocate(frameSizeInBytes, 2);
// F1 F2
// [xxxx|xxxxxxxx]
offset = allocator.Allocate(frameSizeInBytes, 2);
ASSERT_EQ(offset, 8u);
ASSERT_EQ(allocator.GetUsedSize(), frameSizeInBytes * 3);
allocator.Deallocate(2);
ASSERT_EQ(allocator.GetUsedSize(), 0u);
EXPECT_TRUE(allocator.Empty());
}
// Tests ringbuffer sub-allocation at various offsets.
TEST(RingBufferAllocatorTests, RingBufferSubAlloc) {
constexpr size_t maxNumOfFrames = 10;
constexpr size_t frameSizeInBytes = 4;
RingBufferAllocator allocator(maxNumOfFrames * frameSizeInBytes);
// Sub-alloc the first eight frames.
Serial serial = 1;
for (size_t i = 0; i < 8; ++i) {
allocator.Allocate(frameSizeInBytes, serial);
serial += 1;
}
// Each frame corrresponds to the serial number (for simplicity).
//
// F1 F2 F3 F4 F5 F6 F7 F8
// [xxxx|xxxx|xxxx|xxxx|xxxx|xxxx|xxxx|xxxx|--------]
//
// Ensure an oversized allocation fails (only 8 bytes left)
ASSERT_EQ(allocator.Allocate(frameSizeInBytes * 3, serial + 1), kInvalidOffset);
ASSERT_EQ(allocator.GetUsedSize(), frameSizeInBytes * 8);
// Reclaim the first 3 frames.
allocator.Deallocate(3);
// F4 F5 F6 F7 F8
// [------------|xxxx|xxxx|xxxx|xxxx|xxxx|--------]
//
ASSERT_EQ(allocator.GetUsedSize(), frameSizeInBytes * 5);
// Re-try the over-sized allocation.
size_t offset = allocator.Allocate(frameSizeInBytes * 3, serial);
// F9 F4 F5 F6 F7 F8
// [xxxxxxxxxxxx|xxxx|xxxx|xxxx|xxxx|xxxx|xxxxxxxx]
// ^^^^^^^^ wasted
// In this example, Deallocate(8) could not reclaim the wasted bytes. The wasted bytes
// were added to F9's sub-allocation.
// TODO(bryan.bernhart@intel.com): Decide if Deallocate(8) should free these wasted bytes.
ASSERT_EQ(offset, 0u);
ASSERT_EQ(allocator.GetUsedSize(), frameSizeInBytes * maxNumOfFrames);
// Ensure we are full.
ASSERT_EQ(allocator.Allocate(frameSizeInBytes, serial + 1), kInvalidOffset);
// Reclaim the next two frames.
allocator.Deallocate(5);
// F9 F4 F5 F6 F7 F8
// [xxxxxxxxxxxx|----|----|xxxx|xxxx|xxxx|xxxxxxxx]
//
ASSERT_EQ(allocator.GetUsedSize(), frameSizeInBytes * 8);
// Sub-alloc the chunk in the middle.
serial += 1;
offset = allocator.Allocate(frameSizeInBytes * 2, serial);
ASSERT_EQ(offset, frameSizeInBytes * 3);
ASSERT_EQ(allocator.GetUsedSize(), frameSizeInBytes * maxNumOfFrames);
// F9 F10 F6 F7 F8
// [xxxxxxxxxxxx|xxxxxxxxx|xxxx|xxxx|xxxx|xxxxxxxx]
//
// Ensure we are full.
ASSERT_EQ(allocator.Allocate(frameSizeInBytes, serial + 1), kInvalidOffset);
// Reclaim all.
allocator.Deallocate(maxNumOfFrames);
EXPECT_TRUE(allocator.Empty());
}

View File

@ -1,201 +0,0 @@
// Copyright 2018 The Dawn Authors
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include <gtest/gtest.h>
#include "dawn_native/null/DeviceNull.h"
using namespace dawn_native;
namespace {
size_t ValidateValidUploadHandle(const UploadHandle& uploadHandle) {
ASSERT(uploadHandle.mappedBuffer != nullptr);
return uploadHandle.startOffset;
}
void ValidateInvalidUploadHandle(const UploadHandle& uploadHandle) {
ASSERT_EQ(uploadHandle.mappedBuffer, nullptr);
}
} // namespace
class RingBufferTests : public testing::Test {
protected:
void SetUp() override {
// TODO(bryan.bernhart@intel.com): Create this device through the adapter.
mDevice = std::make_unique<null::Device>(/*adapter*/ nullptr, /*deviceDescriptor*/ nullptr);
}
null::Device* GetDevice() const {
return mDevice.get();
}
std::unique_ptr<RingBuffer> CreateRingBuffer(size_t size) {
std::unique_ptr<RingBuffer> ringBuffer = std::make_unique<RingBuffer>(mDevice.get(), size);
DAWN_UNUSED(ringBuffer->Initialize());
return ringBuffer;
}
private:
std::unique_ptr<null::Device> mDevice;
};
// Number of basic tests for Ringbuffer
TEST_F(RingBufferTests, BasicTest) {
constexpr size_t sizeInBytes = 64000;
std::unique_ptr<RingBuffer> buffer = CreateRingBuffer(sizeInBytes);
// Ensure no requests exist on empty buffer.
EXPECT_TRUE(buffer->Empty());
ASSERT_EQ(buffer->GetSize(), sizeInBytes);
// Ensure failure upon sub-allocating an oversized request.
ValidateInvalidUploadHandle(buffer->SubAllocate(sizeInBytes + 1));
// Fill the entire buffer with two requests of equal size.
ValidateValidUploadHandle(buffer->SubAllocate(sizeInBytes / 2));
ValidateValidUploadHandle(buffer->SubAllocate(sizeInBytes / 2));
// Ensure the buffer is full.
ValidateInvalidUploadHandle(buffer->SubAllocate(1));
}
// Tests that several ringbuffer allocations do not fail.
TEST_F(RingBufferTests, RingBufferManyAlloc) {
constexpr size_t maxNumOfFrames = 64000;
constexpr size_t frameSizeInBytes = 4;
std::unique_ptr<RingBuffer> buffer = CreateRingBuffer(maxNumOfFrames * frameSizeInBytes);
size_t offset = 0;
for (size_t i = 0; i < maxNumOfFrames; ++i) {
offset = ValidateValidUploadHandle(buffer->SubAllocate(frameSizeInBytes));
GetDevice()->Tick();
ASSERT_EQ(offset, i * frameSizeInBytes);
}
}
// Tests ringbuffer sub-allocations of the same serial are correctly tracked.
TEST_F(RingBufferTests, AllocInSameFrame) {
constexpr size_t maxNumOfFrames = 3;
constexpr size_t frameSizeInBytes = 4;
std::unique_ptr<RingBuffer> buffer = CreateRingBuffer(maxNumOfFrames * frameSizeInBytes);
// F1
// [xxxx|--------]
ValidateValidUploadHandle(buffer->SubAllocate(frameSizeInBytes));
GetDevice()->Tick();
// F1 F2
// [xxxx|xxxx|----]
ValidateValidUploadHandle(buffer->SubAllocate(frameSizeInBytes));
// F1 F2
// [xxxx|xxxxxxxx]
size_t offset = ValidateValidUploadHandle(buffer->SubAllocate(frameSizeInBytes));
ASSERT_EQ(offset, 8u);
ASSERT_EQ(buffer->GetUsedSize(), frameSizeInBytes * 3);
buffer->Tick(1);
// Used size does not change as previous sub-allocations were not tracked.
ASSERT_EQ(buffer->GetUsedSize(), frameSizeInBytes * 3);
buffer->Tick(2);
ASSERT_EQ(buffer->GetUsedSize(), 0u);
EXPECT_TRUE(buffer->Empty());
}
// Tests ringbuffer sub-allocation at various offsets.
TEST_F(RingBufferTests, RingBufferSubAlloc) {
constexpr size_t maxNumOfFrames = 10;
constexpr size_t frameSizeInBytes = 4;
std::unique_ptr<RingBuffer> buffer = CreateRingBuffer(maxNumOfFrames * frameSizeInBytes);
// Sub-alloc the first eight frames.
for (size_t i = 0; i < 8; ++i) {
ValidateValidUploadHandle(buffer->SubAllocate(frameSizeInBytes));
buffer->Track();
GetDevice()->Tick();
}
// Each frame corrresponds to the serial number (for simplicity).
//
// F1 F2 F3 F4 F5 F6 F7 F8
// [xxxx|xxxx|xxxx|xxxx|xxxx|xxxx|xxxx|xxxx|--------]
//
// Ensure an oversized allocation fails (only 8 bytes left)
ValidateInvalidUploadHandle(buffer->SubAllocate(frameSizeInBytes * 3));
ASSERT_EQ(buffer->GetUsedSize(), frameSizeInBytes * 8);
// Reclaim the first 3 frames.
buffer->Tick(3);
// F4 F5 F6 F7 F8
// [------------|xxxx|xxxx|xxxx|xxxx|xxxx|--------]
//
ASSERT_EQ(buffer->GetUsedSize(), frameSizeInBytes * 5);
// Re-try the over-sized allocation.
size_t offset = ValidateValidUploadHandle(buffer->SubAllocate(frameSizeInBytes * 3));
// F9 F4 F5 F6 F7 F8
// [xxxxxxxxxxxx|xxxx|xxxx|xxxx|xxxx|xxxx|xxxxxxxx]
// ^^^^^^^^ wasted
// In this example, Tick(8) could not reclaim the wasted bytes. The wasted bytes
// were add to F9's sub-allocation.
// TODO(bryan.bernhart@intel.com): Decide if Tick(8) should free these wasted bytes.
ASSERT_EQ(offset, 0u);
ASSERT_EQ(buffer->GetUsedSize(), frameSizeInBytes * maxNumOfFrames);
// Ensure we are full.
ValidateInvalidUploadHandle(buffer->SubAllocate(frameSizeInBytes));
// Reclaim the next two frames.
buffer->Tick(5);
// F9 F4 F5 F6 F7 F8
// [xxxxxxxxxxxx|----|----|xxxx|xxxx|xxxx|xxxxxxxx]
//
ASSERT_EQ(buffer->GetUsedSize(), frameSizeInBytes * 8);
// Sub-alloc the chunk in the middle.
offset = ValidateValidUploadHandle(buffer->SubAllocate(frameSizeInBytes * 2));
ASSERT_EQ(offset, frameSizeInBytes * 3);
ASSERT_EQ(buffer->GetUsedSize(), frameSizeInBytes * maxNumOfFrames);
// F9 F6 F7 F8
// [xxxxxxxxxxxx|xxxx|xxxx|xxxx|xxxx|xxxx|xxxxxxxx]
// ^^^^^^^^^ untracked
// Ensure we are full.
ValidateInvalidUploadHandle(buffer->SubAllocate(frameSizeInBytes));
// Reclaim all.
buffer->Tick(maxNumOfFrames);
EXPECT_TRUE(buffer->Empty());
}