Implement buffer lazy initialization before mapping
This patch implements buffer lazy initialization beffor MapAsync() and buffer creation with BufferDescriptor.mappedAtCreation == true. Note that this patch doesn't initialize buffers in MapReadAsyc() and MapWriteAsync() because they are deprecated and will be removed soon. BUG=dawn:414 TEST=dawn_end2end_tests Change-Id: Ifea99833897081f599c45797e0829c57de1ac926 Reviewed-on: https://dawn-review.googlesource.com/c/dawn/+/24687 Reviewed-by: Corentin Wallez <cwallez@chromium.org> Commit-Queue: Jiawei Shao <jiawei.shao@intel.com>
This commit is contained in:
parent
2733af5c75
commit
77eb64eb8c
|
@ -275,6 +275,10 @@ namespace dawn_native { namespace d3d12 {
|
||||||
}
|
}
|
||||||
|
|
||||||
MaybeError Buffer::MapAtCreationImpl() {
|
MaybeError Buffer::MapAtCreationImpl() {
|
||||||
|
CommandRecordingContext* commandContext;
|
||||||
|
DAWN_TRY_ASSIGN(commandContext, ToBackend(GetDevice())->GetPendingCommandContext());
|
||||||
|
DAWN_TRY(EnsureDataInitialized(commandContext));
|
||||||
|
|
||||||
// Setting isMapWrite to false on MapRead buffers to silence D3D12 debug layer warning.
|
// Setting isMapWrite to false on MapRead buffers to silence D3D12 debug layer warning.
|
||||||
bool isMapWrite = (GetUsage() & wgpu::BufferUsage::MapWrite) != 0;
|
bool isMapWrite = (GetUsage() & wgpu::BufferUsage::MapWrite) != 0;
|
||||||
DAWN_TRY(MapInternal(isMapWrite, 0, size_t(GetSize()), "D3D12 map at creation"));
|
DAWN_TRY(MapInternal(isMapWrite, 0, size_t(GetSize()), "D3D12 map at creation"));
|
||||||
|
@ -290,6 +294,10 @@ namespace dawn_native { namespace d3d12 {
|
||||||
}
|
}
|
||||||
|
|
||||||
MaybeError Buffer::MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) {
|
MaybeError Buffer::MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) {
|
||||||
|
CommandRecordingContext* commandContext;
|
||||||
|
DAWN_TRY_ASSIGN(commandContext, ToBackend(GetDevice())->GetPendingCommandContext());
|
||||||
|
DAWN_TRY(EnsureDataInitialized(commandContext));
|
||||||
|
|
||||||
return MapInternal(mode & wgpu::MapMode::Write, offset, size, "D3D12 map async");
|
return MapInternal(mode & wgpu::MapMode::Write, offset, size, "D3D12 map async");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -110,6 +110,10 @@ namespace dawn_native { namespace metal {
|
||||||
}
|
}
|
||||||
|
|
||||||
MaybeError Buffer::MapAtCreationImpl() {
|
MaybeError Buffer::MapAtCreationImpl() {
|
||||||
|
CommandRecordingContext* commandContext =
|
||||||
|
ToBackend(GetDevice())->GetPendingCommandContext();
|
||||||
|
EnsureDataInitialized(commandContext);
|
||||||
|
|
||||||
return {};
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -122,6 +126,10 @@ namespace dawn_native { namespace metal {
|
||||||
}
|
}
|
||||||
|
|
||||||
MaybeError Buffer::MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) {
|
MaybeError Buffer::MapAsyncImpl(wgpu::MapMode mode, size_t offset, size_t size) {
|
||||||
|
CommandRecordingContext* commandContext =
|
||||||
|
ToBackend(GetDevice())->GetPendingCommandContext();
|
||||||
|
EnsureDataInitialized(commandContext);
|
||||||
|
|
||||||
return {};
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -99,6 +99,8 @@ namespace dawn_native { namespace opengl {
|
||||||
}
|
}
|
||||||
|
|
||||||
MaybeError Buffer::MapAtCreationImpl() {
|
MaybeError Buffer::MapAtCreationImpl() {
|
||||||
|
EnsureDataInitialized();
|
||||||
|
|
||||||
const OpenGLFunctions& gl = ToBackend(GetDevice())->gl;
|
const OpenGLFunctions& gl = ToBackend(GetDevice())->gl;
|
||||||
gl.BindBuffer(GL_ARRAY_BUFFER, mBuffer);
|
gl.BindBuffer(GL_ARRAY_BUFFER, mBuffer);
|
||||||
mMappedData = gl.MapBuffer(GL_ARRAY_BUFFER, GL_WRITE_ONLY);
|
mMappedData = gl.MapBuffer(GL_ARRAY_BUFFER, GL_WRITE_ONLY);
|
||||||
|
@ -137,6 +139,8 @@ namespace dawn_native { namespace opengl {
|
||||||
size = 4;
|
size = 4;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
EnsureDataInitialized();
|
||||||
|
|
||||||
// TODO(cwallez@chromium.org): this does GPU->CPU synchronization, we could require a high
|
// TODO(cwallez@chromium.org): this does GPU->CPU synchronization, we could require a high
|
||||||
// version of OpenGL that would let us map the buffer unsynchronized.
|
// version of OpenGL that would let us map the buffer unsynchronized.
|
||||||
gl.BindBuffer(GL_ARRAY_BUFFER, mBuffer);
|
gl.BindBuffer(GL_ARRAY_BUFFER, mBuffer);
|
||||||
|
|
|
@ -240,6 +240,12 @@ namespace dawn_native { namespace vulkan {
|
||||||
}
|
}
|
||||||
|
|
||||||
MaybeError Buffer::MapAtCreationImpl() {
|
MaybeError Buffer::MapAtCreationImpl() {
|
||||||
|
CommandRecordingContext* recordingContext =
|
||||||
|
ToBackend(GetDevice())->GetPendingRecordingContext();
|
||||||
|
|
||||||
|
// TODO(jiawei.shao@intel.com): initialize mapped buffer in CPU side.
|
||||||
|
EnsureDataInitialized(recordingContext);
|
||||||
|
|
||||||
return {};
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -263,6 +269,10 @@ namespace dawn_native { namespace vulkan {
|
||||||
Device* device = ToBackend(GetDevice());
|
Device* device = ToBackend(GetDevice());
|
||||||
|
|
||||||
CommandRecordingContext* recordingContext = device->GetPendingRecordingContext();
|
CommandRecordingContext* recordingContext = device->GetPendingRecordingContext();
|
||||||
|
|
||||||
|
// TODO(jiawei.shao@intel.com): initialize mapped buffer in CPU side.
|
||||||
|
EnsureDataInitialized(recordingContext);
|
||||||
|
|
||||||
if (mode & wgpu::MapMode::Read) {
|
if (mode & wgpu::MapMode::Read) {
|
||||||
TransitionUsageNow(recordingContext, wgpu::BufferUsage::MapRead);
|
TransitionUsageNow(recordingContext, wgpu::BufferUsage::MapRead);
|
||||||
} else {
|
} else {
|
||||||
|
|
|
@ -30,12 +30,35 @@
|
||||||
|
|
||||||
class BufferZeroInitTest : public DawnTest {
|
class BufferZeroInitTest : public DawnTest {
|
||||||
public:
|
public:
|
||||||
wgpu::Buffer CreateBuffer(uint64_t size, wgpu::BufferUsage usage) {
|
wgpu::Buffer CreateBuffer(uint64_t size,
|
||||||
|
wgpu::BufferUsage usage,
|
||||||
|
bool mappedAtCreation = false) {
|
||||||
wgpu::BufferDescriptor descriptor;
|
wgpu::BufferDescriptor descriptor;
|
||||||
descriptor.size = size;
|
descriptor.size = size;
|
||||||
descriptor.usage = usage;
|
descriptor.usage = usage;
|
||||||
|
descriptor.mappedAtCreation = mappedAtCreation;
|
||||||
return device.CreateBuffer(&descriptor);
|
return device.CreateBuffer(&descriptor);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
void MapAsyncAndWait(wgpu::Buffer buffer,
|
||||||
|
wgpu::MapMode mapMode,
|
||||||
|
uint64_t offset,
|
||||||
|
uint64_t size) {
|
||||||
|
ASSERT(mapMode == wgpu::MapMode::Read || mapMode == wgpu::MapMode::Write);
|
||||||
|
|
||||||
|
bool done = false;
|
||||||
|
buffer.MapAsync(
|
||||||
|
mapMode, offset, size,
|
||||||
|
[](WGPUBufferMapAsyncStatus status, void* userdata) {
|
||||||
|
ASSERT_EQ(WGPUBufferMapAsyncStatus_Success, status);
|
||||||
|
*static_cast<bool*>(userdata) = true;
|
||||||
|
},
|
||||||
|
&done);
|
||||||
|
|
||||||
|
while (!done) {
|
||||||
|
WaitABit();
|
||||||
|
}
|
||||||
|
}
|
||||||
};
|
};
|
||||||
|
|
||||||
// Test that calling writeBuffer to overwrite the entire buffer doesn't need to lazily initialize
|
// Test that calling writeBuffer to overwrite the entire buffer doesn't need to lazily initialize
|
||||||
|
@ -260,6 +283,101 @@ TEST_P(BufferZeroInitTest, CopyBufferToBufferDestination) {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Test that the code path of readable buffer mapping clears the buffer correctly when it is the
|
||||||
|
// first use of the buffer.
|
||||||
|
TEST_P(BufferZeroInitTest, MapReadAsync) {
|
||||||
|
constexpr uint32_t kBufferSize = 16u;
|
||||||
|
constexpr wgpu::BufferUsage kBufferUsage =
|
||||||
|
wgpu::BufferUsage::MapRead | wgpu::BufferUsage::CopyDst;
|
||||||
|
|
||||||
|
constexpr wgpu::MapMode kMapMode = wgpu::MapMode::Read;
|
||||||
|
|
||||||
|
// Map the whole buffer
|
||||||
|
{
|
||||||
|
wgpu::Buffer buffer = CreateBuffer(kBufferSize, kBufferUsage);
|
||||||
|
EXPECT_LAZY_CLEAR(1u, MapAsyncAndWait(buffer, kMapMode, 0, kBufferSize));
|
||||||
|
|
||||||
|
const uint32_t* mappedDataUint = static_cast<const uint32_t*>(buffer.GetConstMappedRange());
|
||||||
|
for (uint32_t i = 0; i < kBufferSize / sizeof(uint32_t); ++i) {
|
||||||
|
EXPECT_EQ(0u, mappedDataUint[i]);
|
||||||
|
}
|
||||||
|
buffer.Unmap();
|
||||||
|
}
|
||||||
|
|
||||||
|
// Map a range of a buffer
|
||||||
|
{
|
||||||
|
wgpu::Buffer buffer = CreateBuffer(kBufferSize, kBufferUsage);
|
||||||
|
|
||||||
|
constexpr uint64_t kOffset = 4u;
|
||||||
|
constexpr uint64_t kSize = 8u;
|
||||||
|
EXPECT_LAZY_CLEAR(1u, MapAsyncAndWait(buffer, kMapMode, kOffset, kSize));
|
||||||
|
|
||||||
|
const uint32_t* mappedDataUint = static_cast<const uint32_t*>(buffer.GetConstMappedRange());
|
||||||
|
for (uint32_t i = 0; i < kSize / sizeof(uint32_t); ++i) {
|
||||||
|
EXPECT_EQ(0u, mappedDataUint[i]);
|
||||||
|
}
|
||||||
|
buffer.Unmap();
|
||||||
|
|
||||||
|
EXPECT_LAZY_CLEAR(0u, MapAsyncAndWait(buffer, kMapMode, 0, kBufferSize));
|
||||||
|
mappedDataUint = static_cast<const uint32_t*>(buffer.GetConstMappedRange());
|
||||||
|
for (uint32_t i = 0; i < kBufferSize / sizeof(uint32_t); ++i) {
|
||||||
|
EXPECT_EQ(0u, mappedDataUint[i]);
|
||||||
|
}
|
||||||
|
buffer.Unmap();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Test that the code path of writable buffer mapping clears the buffer correctly when it is the
|
||||||
|
// first use of the buffer.
|
||||||
|
TEST_P(BufferZeroInitTest, MapWriteAsync) {
|
||||||
|
constexpr uint32_t kBufferSize = 16u;
|
||||||
|
constexpr wgpu::BufferUsage kBufferUsage =
|
||||||
|
wgpu::BufferUsage::MapWrite | wgpu::BufferUsage::CopySrc;
|
||||||
|
|
||||||
|
constexpr wgpu::MapMode kMapMode = wgpu::MapMode::Write;
|
||||||
|
|
||||||
|
constexpr std::array<uint32_t, kBufferSize / sizeof(uint32_t)> kExpectedData = {{0, 0, 0, 0}};
|
||||||
|
|
||||||
|
// Map the whole buffer
|
||||||
|
{
|
||||||
|
wgpu::Buffer buffer = CreateBuffer(kBufferSize, kBufferUsage);
|
||||||
|
EXPECT_LAZY_CLEAR(1u, MapAsyncAndWait(buffer, kMapMode, 0, kBufferSize));
|
||||||
|
buffer.Unmap();
|
||||||
|
|
||||||
|
EXPECT_BUFFER_U32_RANGE_EQ(reinterpret_cast<const uint32_t*>(kExpectedData.data()), buffer,
|
||||||
|
0, kExpectedData.size());
|
||||||
|
}
|
||||||
|
|
||||||
|
// Map a range of a buffer
|
||||||
|
{
|
||||||
|
wgpu::Buffer buffer = CreateBuffer(kBufferSize, kBufferUsage);
|
||||||
|
|
||||||
|
constexpr uint64_t kOffset = 4u;
|
||||||
|
constexpr uint64_t kSize = 8u;
|
||||||
|
EXPECT_LAZY_CLEAR(1u, MapAsyncAndWait(buffer, kMapMode, kOffset, kSize));
|
||||||
|
buffer.Unmap();
|
||||||
|
|
||||||
|
EXPECT_BUFFER_U32_RANGE_EQ(reinterpret_cast<const uint32_t*>(kExpectedData.data()), buffer,
|
||||||
|
0, kExpectedData.size());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// Test that the code path of creating a buffer with BufferDescriptor.mappedAtCreation == true
|
||||||
|
// clears the buffer correctly at the creation of the buffer.
|
||||||
|
TEST_P(BufferZeroInitTest, MapAtCreation) {
|
||||||
|
constexpr uint32_t kBufferSize = 16u;
|
||||||
|
constexpr wgpu::BufferUsage kBufferUsage =
|
||||||
|
wgpu::BufferUsage::MapWrite | wgpu::BufferUsage::CopySrc;
|
||||||
|
|
||||||
|
wgpu::Buffer buffer;
|
||||||
|
EXPECT_LAZY_CLEAR(1u, buffer = CreateBuffer(kBufferSize, kBufferUsage, true));
|
||||||
|
buffer.Unmap();
|
||||||
|
|
||||||
|
constexpr std::array<uint32_t, kBufferSize / sizeof(uint32_t)> kExpectedData = {{0, 0, 0, 0}};
|
||||||
|
EXPECT_BUFFER_U32_RANGE_EQ(reinterpret_cast<const uint32_t*>(kExpectedData.data()), buffer, 0,
|
||||||
|
kExpectedData.size());
|
||||||
|
}
|
||||||
|
|
||||||
DAWN_INSTANTIATE_TEST(BufferZeroInitTest,
|
DAWN_INSTANTIATE_TEST(BufferZeroInitTest,
|
||||||
D3D12Backend({"nonzero_clear_resources_on_creation_for_testing",
|
D3D12Backend({"nonzero_clear_resources_on_creation_for_testing",
|
||||||
"lazy_clear_buffer_on_first_use"}),
|
"lazy_clear_buffer_on_first_use"}),
|
||||||
|
|
Loading…
Reference in New Issue