mirror of
https://github.com/encounter/dawn-cmake.git
synced 2025-05-31 11:41:35 +00:00
This patch implements buffer lazy initialization beffor MapAsync() and buffer creation with BufferDescriptor.mappedAtCreation == true. Note that this patch doesn't initialize buffers in MapReadAsyc() and MapWriteAsync() because they are deprecated and will be removed soon. BUG=dawn:414 TEST=dawn_end2end_tests Change-Id: Ifea99833897081f599c45797e0829c57de1ac926 Reviewed-on: https://dawn-review.googlesource.com/c/dawn/+/24687 Reviewed-by: Corentin Wallez <cwallez@chromium.org> Commit-Queue: Jiawei Shao <jiawei.shao@intel.com>
390 lines
16 KiB
C++
390 lines
16 KiB
C++
// Copyright 2020 The Dawn Authors
|
|
//
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
// you may not use this file except in compliance with the License.
|
|
// You may obtain a copy of the License at
|
|
//
|
|
// http://www.apache.org/licenses/LICENSE-2.0
|
|
//
|
|
// Unless required by applicable law or agreed to in writing, software
|
|
// distributed under the License is distributed on an "AS IS" BASIS,
|
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
// See the License for the specific language governing permissions and
|
|
// limitations under the License.
|
|
|
|
#include "tests/DawnTest.h"
|
|
|
|
#include "utils/WGPUHelpers.h"
|
|
|
|
#define EXPECT_LAZY_CLEAR(N, statement) \
|
|
do { \
|
|
if (UsesWire()) { \
|
|
statement; \
|
|
} else { \
|
|
size_t lazyClearsBefore = dawn_native::GetLazyClearCountForTesting(device.Get()); \
|
|
statement; \
|
|
size_t lazyClearsAfter = dawn_native::GetLazyClearCountForTesting(device.Get()); \
|
|
EXPECT_EQ(N, lazyClearsAfter - lazyClearsBefore); \
|
|
} \
|
|
} while (0)
|
|
|
|
class BufferZeroInitTest : public DawnTest {
|
|
public:
|
|
wgpu::Buffer CreateBuffer(uint64_t size,
|
|
wgpu::BufferUsage usage,
|
|
bool mappedAtCreation = false) {
|
|
wgpu::BufferDescriptor descriptor;
|
|
descriptor.size = size;
|
|
descriptor.usage = usage;
|
|
descriptor.mappedAtCreation = mappedAtCreation;
|
|
return device.CreateBuffer(&descriptor);
|
|
}
|
|
|
|
void MapAsyncAndWait(wgpu::Buffer buffer,
|
|
wgpu::MapMode mapMode,
|
|
uint64_t offset,
|
|
uint64_t size) {
|
|
ASSERT(mapMode == wgpu::MapMode::Read || mapMode == wgpu::MapMode::Write);
|
|
|
|
bool done = false;
|
|
buffer.MapAsync(
|
|
mapMode, offset, size,
|
|
[](WGPUBufferMapAsyncStatus status, void* userdata) {
|
|
ASSERT_EQ(WGPUBufferMapAsyncStatus_Success, status);
|
|
*static_cast<bool*>(userdata) = true;
|
|
},
|
|
&done);
|
|
|
|
while (!done) {
|
|
WaitABit();
|
|
}
|
|
}
|
|
};
|
|
|
|
// Test that calling writeBuffer to overwrite the entire buffer doesn't need to lazily initialize
|
|
// the destination buffer.
|
|
TEST_P(BufferZeroInitTest, WriteBufferToEntireBuffer) {
|
|
constexpr uint32_t kBufferSize = 8u;
|
|
constexpr wgpu::BufferUsage kBufferUsage =
|
|
wgpu::BufferUsage::CopySrc | wgpu::BufferUsage::CopyDst;
|
|
wgpu::Buffer buffer = CreateBuffer(kBufferSize, kBufferUsage);
|
|
|
|
constexpr std::array<uint32_t, kBufferSize / sizeof(uint32_t)> kExpectedData = {
|
|
{0x02020202u, 0x02020202u}};
|
|
EXPECT_LAZY_CLEAR(0u, queue.WriteBuffer(buffer, 0, kExpectedData.data(), kBufferSize));
|
|
|
|
EXPECT_BUFFER_U32_RANGE_EQ(kExpectedData.data(), buffer, 0, kBufferSize / sizeof(uint32_t));
|
|
}
|
|
|
|
// Test that calling writeBuffer to overwrite a part of buffer needs to lazily initialize the
|
|
// destination buffer.
|
|
TEST_P(BufferZeroInitTest, WriteBufferToSubBuffer) {
|
|
constexpr uint32_t kBufferSize = 8u;
|
|
constexpr wgpu::BufferUsage kBufferUsage =
|
|
wgpu::BufferUsage::CopySrc | wgpu::BufferUsage::CopyDst;
|
|
|
|
constexpr uint32_t kCopyValue = 0x02020202u;
|
|
|
|
// offset == 0
|
|
{
|
|
wgpu::Buffer buffer = CreateBuffer(kBufferSize, kBufferUsage);
|
|
|
|
constexpr uint32_t kCopyOffset = 0u;
|
|
EXPECT_LAZY_CLEAR(1u,
|
|
queue.WriteBuffer(buffer, kCopyOffset, &kCopyValue, sizeof(kCopyValue)));
|
|
|
|
EXPECT_BUFFER_U32_EQ(kCopyValue, buffer, kCopyOffset);
|
|
EXPECT_BUFFER_U32_EQ(0, buffer, kBufferSize - sizeof(kCopyValue));
|
|
}
|
|
|
|
// offset > 0
|
|
{
|
|
wgpu::Buffer buffer = CreateBuffer(kBufferSize, kBufferUsage);
|
|
|
|
constexpr uint32_t kCopyOffset = 4u;
|
|
EXPECT_LAZY_CLEAR(1u,
|
|
queue.WriteBuffer(buffer, kCopyOffset, &kCopyValue, sizeof(kCopyValue)));
|
|
|
|
EXPECT_BUFFER_U32_EQ(0, buffer, 0);
|
|
EXPECT_BUFFER_U32_EQ(kCopyValue, buffer, kCopyOffset);
|
|
}
|
|
}
|
|
|
|
// Test that the code path of CopyBufferToBuffer clears the source buffer correctly when it is the
|
|
// first use of the source buffer.
|
|
TEST_P(BufferZeroInitTest, CopyBufferToBufferSource) {
|
|
constexpr uint64_t kBufferSize = 16u;
|
|
constexpr wgpu::BufferUsage kBufferUsage =
|
|
wgpu::BufferUsage::CopySrc | wgpu::BufferUsage::CopyDst;
|
|
wgpu::BufferDescriptor bufferDescriptor;
|
|
bufferDescriptor.size = kBufferSize;
|
|
bufferDescriptor.usage = kBufferUsage;
|
|
|
|
constexpr std::array<uint8_t, kBufferSize> kInitialData = {
|
|
{1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16}};
|
|
|
|
wgpu::Buffer dstBuffer =
|
|
utils::CreateBufferFromData(device, kInitialData.data(), kBufferSize, kBufferUsage);
|
|
|
|
constexpr std::array<uint32_t, kBufferSize / sizeof(uint32_t)> kExpectedData = {{0, 0, 0, 0}};
|
|
|
|
// Full copy from the source buffer
|
|
{
|
|
wgpu::Buffer srcBuffer = device.CreateBuffer(&bufferDescriptor);
|
|
wgpu::CommandEncoder encoder = device.CreateCommandEncoder();
|
|
encoder.CopyBufferToBuffer(srcBuffer, 0, dstBuffer, 0, kBufferSize);
|
|
wgpu::CommandBuffer commandBuffer = encoder.Finish();
|
|
|
|
EXPECT_LAZY_CLEAR(1u, queue.Submit(1, &commandBuffer));
|
|
EXPECT_BUFFER_U32_RANGE_EQ(kExpectedData.data(), srcBuffer, 0,
|
|
kBufferSize / sizeof(uint32_t));
|
|
}
|
|
|
|
// Partial copy from the source buffer
|
|
// srcOffset == 0
|
|
{
|
|
constexpr uint64_t kSrcOffset = 0;
|
|
constexpr uint64_t kCopySize = kBufferSize / 2;
|
|
|
|
wgpu::Buffer srcBuffer = device.CreateBuffer(&bufferDescriptor);
|
|
wgpu::CommandEncoder encoder = device.CreateCommandEncoder();
|
|
encoder.CopyBufferToBuffer(srcBuffer, kSrcOffset, dstBuffer, 0, kCopySize);
|
|
wgpu::CommandBuffer commandBuffer = encoder.Finish();
|
|
|
|
EXPECT_LAZY_CLEAR(1u, queue.Submit(1, &commandBuffer));
|
|
EXPECT_BUFFER_U32_RANGE_EQ(kExpectedData.data(), srcBuffer, 0,
|
|
kBufferSize / sizeof(uint32_t));
|
|
}
|
|
|
|
// srcOffset > 0 and srcOffset + copySize == srcBufferSize
|
|
{
|
|
constexpr uint64_t kSrcOffset = kBufferSize / 2;
|
|
constexpr uint64_t kCopySize = kBufferSize - kSrcOffset;
|
|
|
|
wgpu::Buffer srcBuffer = device.CreateBuffer(&bufferDescriptor);
|
|
wgpu::CommandEncoder encoder = device.CreateCommandEncoder();
|
|
encoder.CopyBufferToBuffer(srcBuffer, kSrcOffset, dstBuffer, 0, kCopySize);
|
|
wgpu::CommandBuffer commandBuffer = encoder.Finish();
|
|
|
|
EXPECT_LAZY_CLEAR(1u, queue.Submit(1, &commandBuffer));
|
|
EXPECT_BUFFER_U32_RANGE_EQ(kExpectedData.data(), srcBuffer, 0,
|
|
kBufferSize / sizeof(uint32_t));
|
|
}
|
|
|
|
// srcOffset > 0 and srcOffset + copySize < srcBufferSize
|
|
{
|
|
constexpr uint64_t kSrcOffset = kBufferSize / 4;
|
|
constexpr uint64_t kCopySize = kBufferSize / 2;
|
|
|
|
wgpu::Buffer srcBuffer = device.CreateBuffer(&bufferDescriptor);
|
|
wgpu::CommandEncoder encoder = device.CreateCommandEncoder();
|
|
encoder.CopyBufferToBuffer(srcBuffer, kSrcOffset, dstBuffer, 0, kCopySize);
|
|
wgpu::CommandBuffer commandBuffer = encoder.Finish();
|
|
|
|
EXPECT_LAZY_CLEAR(1u, queue.Submit(1, &commandBuffer));
|
|
EXPECT_BUFFER_U32_RANGE_EQ(kExpectedData.data(), srcBuffer, 0,
|
|
kBufferSize / sizeof(uint32_t));
|
|
}
|
|
}
|
|
|
|
// Test that the code path of CopyBufferToBuffer clears the destination buffer correctly when it is
|
|
// the first use of the destination buffer.
|
|
TEST_P(BufferZeroInitTest, CopyBufferToBufferDestination) {
|
|
constexpr uint64_t kBufferSize = 16u;
|
|
constexpr wgpu::BufferUsage kBufferUsage =
|
|
wgpu::BufferUsage::CopySrc | wgpu::BufferUsage::CopyDst;
|
|
wgpu::BufferDescriptor bufferDescriptor;
|
|
bufferDescriptor.size = kBufferSize;
|
|
bufferDescriptor.usage = kBufferUsage;
|
|
|
|
const std::array<uint8_t, kBufferSize> kInitialData = {
|
|
{1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16}};
|
|
wgpu::Buffer srcBuffer =
|
|
utils::CreateBufferFromData(device, kInitialData.data(), kBufferSize, kBufferUsage);
|
|
|
|
// Full copy from the source buffer doesn't need lazy initialization at all.
|
|
{
|
|
wgpu::Buffer dstBuffer = device.CreateBuffer(&bufferDescriptor);
|
|
wgpu::CommandEncoder encoder = device.CreateCommandEncoder();
|
|
encoder.CopyBufferToBuffer(srcBuffer, 0, dstBuffer, 0, kBufferSize);
|
|
wgpu::CommandBuffer commandBuffer = encoder.Finish();
|
|
|
|
EXPECT_LAZY_CLEAR(0u, queue.Submit(1, &commandBuffer));
|
|
|
|
EXPECT_BUFFER_U32_RANGE_EQ(reinterpret_cast<const uint32_t*>(kInitialData.data()),
|
|
dstBuffer, 0, kBufferSize / sizeof(uint32_t));
|
|
}
|
|
|
|
// Partial copy from the source buffer needs lazy initialization.
|
|
// offset == 0
|
|
{
|
|
constexpr uint32_t kDstOffset = 0;
|
|
constexpr uint32_t kCopySize = kBufferSize / 2;
|
|
|
|
wgpu::Buffer dstBuffer = device.CreateBuffer(&bufferDescriptor);
|
|
wgpu::CommandEncoder encoder = device.CreateCommandEncoder();
|
|
encoder.CopyBufferToBuffer(srcBuffer, 0, dstBuffer, kDstOffset, kCopySize);
|
|
wgpu::CommandBuffer commandBuffer = encoder.Finish();
|
|
|
|
EXPECT_LAZY_CLEAR(1u, queue.Submit(1, &commandBuffer));
|
|
|
|
std::array<uint8_t, kBufferSize> expectedData;
|
|
expectedData.fill(0);
|
|
for (uint32_t index = kDstOffset; index < kDstOffset + kCopySize; ++index) {
|
|
expectedData[index] = kInitialData[index - kDstOffset];
|
|
}
|
|
|
|
EXPECT_BUFFER_U32_RANGE_EQ(reinterpret_cast<uint32_t*>(expectedData.data()), dstBuffer, 0,
|
|
kBufferSize / sizeof(uint32_t));
|
|
}
|
|
|
|
// offset > 0 and dstOffset + CopySize == kBufferSize
|
|
{
|
|
constexpr uint32_t kDstOffset = kBufferSize / 2;
|
|
constexpr uint32_t kCopySize = kBufferSize - kDstOffset;
|
|
|
|
wgpu::Buffer dstBuffer = device.CreateBuffer(&bufferDescriptor);
|
|
wgpu::CommandEncoder encoder = device.CreateCommandEncoder();
|
|
encoder.CopyBufferToBuffer(srcBuffer, 0, dstBuffer, kDstOffset, kCopySize);
|
|
wgpu::CommandBuffer commandBuffer = encoder.Finish();
|
|
|
|
EXPECT_LAZY_CLEAR(1u, queue.Submit(1, &commandBuffer));
|
|
|
|
std::array<uint8_t, kBufferSize> expectedData;
|
|
expectedData.fill(0);
|
|
for (uint32_t index = kDstOffset; index < kDstOffset + kCopySize; ++index) {
|
|
expectedData[index] = kInitialData[index - kDstOffset];
|
|
}
|
|
|
|
EXPECT_BUFFER_U32_RANGE_EQ(reinterpret_cast<uint32_t*>(expectedData.data()), dstBuffer, 0,
|
|
kBufferSize / sizeof(uint32_t));
|
|
}
|
|
|
|
// offset > 0 and dstOffset + CopySize < kBufferSize
|
|
{
|
|
constexpr uint32_t kDstOffset = kBufferSize / 4;
|
|
constexpr uint32_t kCopySize = kBufferSize / 2;
|
|
|
|
wgpu::Buffer dstBuffer = device.CreateBuffer(&bufferDescriptor);
|
|
wgpu::CommandEncoder encoder = device.CreateCommandEncoder();
|
|
encoder.CopyBufferToBuffer(srcBuffer, 0, dstBuffer, kDstOffset, kCopySize);
|
|
wgpu::CommandBuffer commandBuffer = encoder.Finish();
|
|
|
|
EXPECT_LAZY_CLEAR(1u, queue.Submit(1, &commandBuffer));
|
|
|
|
std::array<uint8_t, kBufferSize> expectedData;
|
|
expectedData.fill(0);
|
|
for (uint32_t index = kDstOffset; index < kDstOffset + kCopySize; ++index) {
|
|
expectedData[index] = kInitialData[index - kDstOffset];
|
|
}
|
|
|
|
EXPECT_BUFFER_U32_RANGE_EQ(reinterpret_cast<uint32_t*>(expectedData.data()), dstBuffer, 0,
|
|
kBufferSize / sizeof(uint32_t));
|
|
}
|
|
}
|
|
|
|
// Test that the code path of readable buffer mapping clears the buffer correctly when it is the
|
|
// first use of the buffer.
|
|
TEST_P(BufferZeroInitTest, MapReadAsync) {
|
|
constexpr uint32_t kBufferSize = 16u;
|
|
constexpr wgpu::BufferUsage kBufferUsage =
|
|
wgpu::BufferUsage::MapRead | wgpu::BufferUsage::CopyDst;
|
|
|
|
constexpr wgpu::MapMode kMapMode = wgpu::MapMode::Read;
|
|
|
|
// Map the whole buffer
|
|
{
|
|
wgpu::Buffer buffer = CreateBuffer(kBufferSize, kBufferUsage);
|
|
EXPECT_LAZY_CLEAR(1u, MapAsyncAndWait(buffer, kMapMode, 0, kBufferSize));
|
|
|
|
const uint32_t* mappedDataUint = static_cast<const uint32_t*>(buffer.GetConstMappedRange());
|
|
for (uint32_t i = 0; i < kBufferSize / sizeof(uint32_t); ++i) {
|
|
EXPECT_EQ(0u, mappedDataUint[i]);
|
|
}
|
|
buffer.Unmap();
|
|
}
|
|
|
|
// Map a range of a buffer
|
|
{
|
|
wgpu::Buffer buffer = CreateBuffer(kBufferSize, kBufferUsage);
|
|
|
|
constexpr uint64_t kOffset = 4u;
|
|
constexpr uint64_t kSize = 8u;
|
|
EXPECT_LAZY_CLEAR(1u, MapAsyncAndWait(buffer, kMapMode, kOffset, kSize));
|
|
|
|
const uint32_t* mappedDataUint = static_cast<const uint32_t*>(buffer.GetConstMappedRange());
|
|
for (uint32_t i = 0; i < kSize / sizeof(uint32_t); ++i) {
|
|
EXPECT_EQ(0u, mappedDataUint[i]);
|
|
}
|
|
buffer.Unmap();
|
|
|
|
EXPECT_LAZY_CLEAR(0u, MapAsyncAndWait(buffer, kMapMode, 0, kBufferSize));
|
|
mappedDataUint = static_cast<const uint32_t*>(buffer.GetConstMappedRange());
|
|
for (uint32_t i = 0; i < kBufferSize / sizeof(uint32_t); ++i) {
|
|
EXPECT_EQ(0u, mappedDataUint[i]);
|
|
}
|
|
buffer.Unmap();
|
|
}
|
|
}
|
|
|
|
// Test that the code path of writable buffer mapping clears the buffer correctly when it is the
|
|
// first use of the buffer.
|
|
TEST_P(BufferZeroInitTest, MapWriteAsync) {
|
|
constexpr uint32_t kBufferSize = 16u;
|
|
constexpr wgpu::BufferUsage kBufferUsage =
|
|
wgpu::BufferUsage::MapWrite | wgpu::BufferUsage::CopySrc;
|
|
|
|
constexpr wgpu::MapMode kMapMode = wgpu::MapMode::Write;
|
|
|
|
constexpr std::array<uint32_t, kBufferSize / sizeof(uint32_t)> kExpectedData = {{0, 0, 0, 0}};
|
|
|
|
// Map the whole buffer
|
|
{
|
|
wgpu::Buffer buffer = CreateBuffer(kBufferSize, kBufferUsage);
|
|
EXPECT_LAZY_CLEAR(1u, MapAsyncAndWait(buffer, kMapMode, 0, kBufferSize));
|
|
buffer.Unmap();
|
|
|
|
EXPECT_BUFFER_U32_RANGE_EQ(reinterpret_cast<const uint32_t*>(kExpectedData.data()), buffer,
|
|
0, kExpectedData.size());
|
|
}
|
|
|
|
// Map a range of a buffer
|
|
{
|
|
wgpu::Buffer buffer = CreateBuffer(kBufferSize, kBufferUsage);
|
|
|
|
constexpr uint64_t kOffset = 4u;
|
|
constexpr uint64_t kSize = 8u;
|
|
EXPECT_LAZY_CLEAR(1u, MapAsyncAndWait(buffer, kMapMode, kOffset, kSize));
|
|
buffer.Unmap();
|
|
|
|
EXPECT_BUFFER_U32_RANGE_EQ(reinterpret_cast<const uint32_t*>(kExpectedData.data()), buffer,
|
|
0, kExpectedData.size());
|
|
}
|
|
}
|
|
|
|
// Test that the code path of creating a buffer with BufferDescriptor.mappedAtCreation == true
|
|
// clears the buffer correctly at the creation of the buffer.
|
|
TEST_P(BufferZeroInitTest, MapAtCreation) {
|
|
constexpr uint32_t kBufferSize = 16u;
|
|
constexpr wgpu::BufferUsage kBufferUsage =
|
|
wgpu::BufferUsage::MapWrite | wgpu::BufferUsage::CopySrc;
|
|
|
|
wgpu::Buffer buffer;
|
|
EXPECT_LAZY_CLEAR(1u, buffer = CreateBuffer(kBufferSize, kBufferUsage, true));
|
|
buffer.Unmap();
|
|
|
|
constexpr std::array<uint32_t, kBufferSize / sizeof(uint32_t)> kExpectedData = {{0, 0, 0, 0}};
|
|
EXPECT_BUFFER_U32_RANGE_EQ(reinterpret_cast<const uint32_t*>(kExpectedData.data()), buffer, 0,
|
|
kExpectedData.size());
|
|
}
|
|
|
|
DAWN_INSTANTIATE_TEST(BufferZeroInitTest,
|
|
D3D12Backend({"nonzero_clear_resources_on_creation_for_testing",
|
|
"lazy_clear_buffer_on_first_use"}),
|
|
MetalBackend({"nonzero_clear_resources_on_creation_for_testing",
|
|
"lazy_clear_buffer_on_first_use"}),
|
|
OpenGLBackend({"nonzero_clear_resources_on_creation_for_testing",
|
|
"lazy_clear_buffer_on_first_use"}),
|
|
VulkanBackend({"nonzero_clear_resources_on_creation_for_testing",
|
|
"lazy_clear_buffer_on_first_use"}));
|