Add per-stage and per-pipeline-layout limits and validation

This CL adds per-stage and per-pipeline-layout limits according
to the WebGPU spec. It also slightly increases kMaxBindingsPerGroup
from 16 to 24 so that these limits can be effectively tested
without hitting kMaxBindingsPerGroup. kMaxBindingsPerGroup is not a
real WebGPU limit and will be removed in future patches.

Bug: dawn:443
Change-Id: I72be062cd31dea4ebd851f2d9f8274a77f286846
Reviewed-on: https://dawn-review.googlesource.com/c/dawn/+/24481
Reviewed-by: Stephen White <senorblanco@chromium.org>
Commit-Queue: Austin Eng <enga@chromium.org>
This commit is contained in:
Austin Eng 2020-07-14 00:53:23 +00:00 committed by Commit Bot service account
parent 261b05d3dd
commit 8e316f1177
11 changed files with 388 additions and 155 deletions

View File

@ -19,7 +19,7 @@
static constexpr uint32_t kMaxBindGroups = 4u; static constexpr uint32_t kMaxBindGroups = 4u;
// TODO(cwallez@chromium.org): investigate bindgroup limits // TODO(cwallez@chromium.org): investigate bindgroup limits
static constexpr uint32_t kMaxBindingsPerGroup = 16u; static constexpr uint32_t kMaxBindingsPerGroup = 24u;
static constexpr uint32_t kMaxVertexAttributes = 16u; static constexpr uint32_t kMaxVertexAttributes = 16u;
// Vulkan has a standalone limit named maxVertexInputAttributeOffset (2047u at least) for vertex // Vulkan has a standalone limit named maxVertexInputAttributeOffset (2047u at least) for vertex
// attribute offset. The limit might be meaningless because Vulkan has another limit named // attribute offset. The limit might be meaningless because Vulkan has another limit named
@ -35,15 +35,21 @@ static constexpr uint32_t kMaxColorAttachments = 4u;
static constexpr uint32_t kTextureBytesPerRowAlignment = 256u; static constexpr uint32_t kTextureBytesPerRowAlignment = 256u;
// Dynamic buffer offsets require offset to be divisible by 256 // Dynamic buffer offsets require offset to be divisible by 256
static constexpr uint64_t kMinDynamicBufferOffsetAlignment = 256u; static constexpr uint64_t kMinDynamicBufferOffsetAlignment = 256u;
// Max numbers of dynamic uniform buffers
static constexpr uint32_t kMaxDynamicUniformBufferCount = 8u; // Per stage limits
// Max numbers of dynamic storage buffers static constexpr uint32_t kMaxSampledTexturesPerShaderStage = 16;
static constexpr uint32_t kMaxDynamicStorageBufferCount = 4u; static constexpr uint32_t kMaxSamplersPerShaderStage = 16;
// Max numbers of dynamic buffers static constexpr uint32_t kMaxStorageBuffersPerShaderStage = 4;
static constexpr uint32_t kMaxDynamicBufferCount = static constexpr uint32_t kMaxStorageTexturesPerShaderStage = 4;
kMaxDynamicUniformBufferCount + kMaxDynamicStorageBufferCount; static constexpr uint32_t kMaxUniformBuffersPerShaderStage = 12;
// Per pipeline layout limits
static constexpr uint32_t kMaxDynamicUniformBuffersPerPipelineLayout = 8u;
static constexpr uint32_t kMaxDynamicStorageBuffersPerPipelineLayout = 4u;
// Max size of uniform buffer binding // Max size of uniform buffer binding
static constexpr uint64_t kMaxUniformBufferBindingSize = 16384u; static constexpr uint64_t kMaxUniformBufferBindingSize = 16384u;
// Indirect command sizes // Indirect command sizes
static constexpr uint64_t kDispatchIndirectSize = 3 * sizeof(uint32_t); static constexpr uint64_t kDispatchIndirectSize = 3 * sizeof(uint32_t);
static constexpr uint64_t kDrawIndirectSize = 4 * sizeof(uint32_t); static constexpr uint64_t kDrawIndirectSize = 4 * sizeof(uint32_t);

View File

@ -163,6 +163,7 @@ source_set("dawn_native_sources") {
"BindGroupLayout.cpp", "BindGroupLayout.cpp",
"BindGroupLayout.h", "BindGroupLayout.h",
"BindGroupTracker.h", "BindGroupTracker.h",
"BindingInfo.cpp",
"BindingInfo.h", "BindingInfo.h",
"BuddyAllocator.cpp", "BuddyAllocator.cpp",
"BuddyAllocator.h", "BuddyAllocator.h",

View File

@ -17,6 +17,7 @@
#include "common/BitSetIterator.h" #include "common/BitSetIterator.h"
#include "common/HashUtils.h" #include "common/HashUtils.h"
#include "dawn_native/Device.h" #include "dawn_native/Device.h"
#include "dawn_native/PerStage.h"
#include "dawn_native/ValidationUtils_autogen.h" #include "dawn_native/ValidationUtils_autogen.h"
#include <algorithm> #include <algorithm>
@ -194,8 +195,7 @@ namespace dawn_native {
} }
std::set<BindingNumber> bindingsSet; std::set<BindingNumber> bindingsSet;
uint32_t dynamicUniformBufferCount = 0; BindingCounts bindingCounts = {};
uint32_t dynamicStorageBufferCount = 0;
for (uint32_t i = 0; i < descriptor->entryCount; ++i) { for (uint32_t i = 0; i < descriptor->entryCount; ++i) {
const BindGroupLayoutEntry& entry = descriptor->entries[i]; const BindGroupLayoutEntry& entry = descriptor->entries[i];
BindingNumber bindingNumber = BindingNumber(entry.binding); BindingNumber bindingNumber = BindingNumber(entry.binding);
@ -226,29 +226,35 @@ namespace dawn_native {
switch (entry.type) { switch (entry.type) {
case wgpu::BindingType::UniformBuffer: case wgpu::BindingType::UniformBuffer:
if (entry.hasDynamicOffset) {
++dynamicUniformBufferCount;
}
break;
case wgpu::BindingType::StorageBuffer: case wgpu::BindingType::StorageBuffer:
case wgpu::BindingType::ReadonlyStorageBuffer: case wgpu::BindingType::ReadonlyStorageBuffer:
if (entry.hasDynamicOffset) {
++dynamicStorageBufferCount;
}
break; break;
case wgpu::BindingType::SampledTexture: case wgpu::BindingType::SampledTexture:
if (entry.hasDynamicOffset) {
return DAWN_VALIDATION_ERROR("Sampled textures cannot be dynamic");
}
break;
case wgpu::BindingType::Sampler: case wgpu::BindingType::Sampler:
case wgpu::BindingType::ComparisonSampler: case wgpu::BindingType::ComparisonSampler:
if (entry.hasDynamicOffset) {
return DAWN_VALIDATION_ERROR("Samplers cannot be dynamic");
}
break;
case wgpu::BindingType::ReadonlyStorageTexture: case wgpu::BindingType::ReadonlyStorageTexture:
case wgpu::BindingType::WriteonlyStorageTexture: case wgpu::BindingType::WriteonlyStorageTexture:
if (entry.hasDynamicOffset) { if (entry.hasDynamicOffset) {
return DAWN_VALIDATION_ERROR("Samplers and textures cannot be dynamic"); return DAWN_VALIDATION_ERROR("Storage textures cannot be dynamic");
} }
break; break;
case wgpu::BindingType::StorageTexture: case wgpu::BindingType::StorageTexture:
return DAWN_VALIDATION_ERROR("storage textures aren't supported (yet)"); return DAWN_VALIDATION_ERROR("storage textures aren't supported (yet)");
default:
UNREACHABLE();
break;
} }
IncrementBindingCounts(&bindingCounts, entry);
bindingsSet.insert(bindingNumber); bindingsSet.insert(bindingNumber);
} }
@ -256,15 +262,7 @@ namespace dawn_native {
return DAWN_VALIDATION_ERROR("The number of bindings exceeds kMaxBindingsPerGroup."); return DAWN_VALIDATION_ERROR("The number of bindings exceeds kMaxBindingsPerGroup.");
} }
if (dynamicUniformBufferCount > kMaxDynamicUniformBufferCount) { DAWN_TRY(ValidateBindingCounts(bindingCounts));
return DAWN_VALIDATION_ERROR(
"The number of dynamic uniform buffer exceeds the maximum value");
}
if (dynamicStorageBufferCount > kMaxDynamicStorageBufferCount) {
return DAWN_VALIDATION_ERROR(
"The number of dynamic storage buffer exceeds the maximum value");
}
return {}; return {};
} }
@ -383,13 +381,13 @@ namespace dawn_native {
BindGroupLayoutBase::BindGroupLayoutBase(DeviceBase* device, BindGroupLayoutBase::BindGroupLayoutBase(DeviceBase* device,
const BindGroupLayoutDescriptor* descriptor) const BindGroupLayoutDescriptor* descriptor)
: CachedObject(device), mBindingCount(descriptor->entryCount) { : CachedObject(device) {
std::vector<BindGroupLayoutEntry> sortedBindings( std::vector<BindGroupLayoutEntry> sortedBindings(
descriptor->entries, descriptor->entries + descriptor->entryCount); descriptor->entries, descriptor->entries + descriptor->entryCount);
std::sort(sortedBindings.begin(), sortedBindings.end(), SortBindingsCompare); std::sort(sortedBindings.begin(), sortedBindings.end(), SortBindingsCompare);
for (BindingIndex i{0}; i < mBindingCount; ++i) { for (BindingIndex i{0}; i < BindingIndex(descriptor->entryCount); ++i) {
const BindGroupLayoutEntry& binding = sortedBindings[static_cast<uint32_t>(i)]; const BindGroupLayoutEntry& binding = sortedBindings[static_cast<uint32_t>(i)];
mBindingInfo[i].binding = BindingNumber(binding.binding); mBindingInfo[i].binding = BindingNumber(binding.binding);
mBindingInfo[i].type = binding.type; mBindingInfo[i].type = binding.type;
@ -399,21 +397,6 @@ namespace dawn_native {
mBindingInfo[i].storageTextureFormat = binding.storageTextureFormat; mBindingInfo[i].storageTextureFormat = binding.storageTextureFormat;
mBindingInfo[i].minBufferBindingSize = binding.minBufferBindingSize; mBindingInfo[i].minBufferBindingSize = binding.minBufferBindingSize;
switch (binding.type) {
case wgpu::BindingType::UniformBuffer:
case wgpu::BindingType::StorageBuffer:
case wgpu::BindingType::ReadonlyStorageBuffer:
// Buffers must be contiguously packed at the start of the binding info.
ASSERT(mBufferCount == i);
++mBufferCount;
if (binding.minBufferBindingSize == 0) {
++mUnverifiedBufferCount;
}
break;
default:
break;
}
if (binding.viewDimension == wgpu::TextureViewDimension::Undefined) { if (binding.viewDimension == wgpu::TextureViewDimension::Undefined) {
mBindingInfo[i].viewDimension = wgpu::TextureViewDimension::e2D; mBindingInfo[i].viewDimension = wgpu::TextureViewDimension::e2D;
} else { } else {
@ -422,30 +405,17 @@ namespace dawn_native {
mBindingInfo[i].multisampled = binding.multisampled; mBindingInfo[i].multisampled = binding.multisampled;
mBindingInfo[i].hasDynamicOffset = binding.hasDynamicOffset; mBindingInfo[i].hasDynamicOffset = binding.hasDynamicOffset;
if (binding.hasDynamicOffset) {
switch (binding.type) { if (IsBufferBinding(binding.type)) {
case wgpu::BindingType::UniformBuffer: // Buffers must be contiguously packed at the start of the binding info.
++mDynamicUniformBufferCount; ASSERT(GetBufferCount() == i);
break;
case wgpu::BindingType::StorageBuffer:
case wgpu::BindingType::ReadonlyStorageBuffer:
++mDynamicStorageBufferCount;
break;
case wgpu::BindingType::SampledTexture:
case wgpu::BindingType::Sampler:
case wgpu::BindingType::ComparisonSampler:
case wgpu::BindingType::StorageTexture:
case wgpu::BindingType::ReadonlyStorageTexture:
case wgpu::BindingType::WriteonlyStorageTexture:
UNREACHABLE();
break;
}
} }
IncrementBindingCounts(&mBindingCounts, binding);
const auto& it = mBindingMap.emplace(BindingNumber(binding.binding), i); const auto& it = mBindingMap.emplace(BindingNumber(binding.binding), i);
ASSERT(it.second); ASSERT(it.second);
} }
ASSERT(CheckBufferBindingsFirst({mBindingInfo.data(), mBindingCount})); ASSERT(CheckBufferBindingsFirst({mBindingInfo.data(), GetBindingCount()}));
} }
BindGroupLayoutBase::BindGroupLayoutBase(DeviceBase* device, ObjectBase::ErrorTag tag) BindGroupLayoutBase::BindGroupLayoutBase(DeviceBase* device, ObjectBase::ErrorTag tag)
@ -501,29 +471,26 @@ namespace dawn_native {
} }
BindingIndex BindGroupLayoutBase::GetBindingCount() const { BindingIndex BindGroupLayoutBase::GetBindingCount() const {
return mBindingCount; return BindingIndex(mBindingCounts.totalCount);
} }
BindingIndex BindGroupLayoutBase::GetBufferCount() const { BindingIndex BindGroupLayoutBase::GetBufferCount() const {
return mBufferCount; return BindingIndex(mBindingCounts.bufferCount);
} }
BindingIndex BindGroupLayoutBase::GetDynamicBufferCount() const { BindingIndex BindGroupLayoutBase::GetDynamicBufferCount() const {
// This is a binding index because dynamic buffers are packed at the front of the binding // This is a binding index because dynamic buffers are packed at the front of the binding
// info. // info.
return static_cast<BindingIndex>(mDynamicStorageBufferCount + mDynamicUniformBufferCount); return static_cast<BindingIndex>(mBindingCounts.dynamicStorageBufferCount +
} mBindingCounts.dynamicUniformBufferCount);
uint32_t BindGroupLayoutBase::GetDynamicUniformBufferCount() const {
return mDynamicUniformBufferCount;
}
uint32_t BindGroupLayoutBase::GetDynamicStorageBufferCount() const {
return mDynamicStorageBufferCount;
} }
uint32_t BindGroupLayoutBase::GetUnverifiedBufferCount() const { uint32_t BindGroupLayoutBase::GetUnverifiedBufferCount() const {
return mUnverifiedBufferCount; return mBindingCounts.unverifiedBufferCount;
}
const BindingCounts& BindGroupLayoutBase::GetBindingCountInfo() const {
return mBindingCounts;
} }
size_t BindGroupLayoutBase::GetBindingDataSize() const { size_t BindGroupLayoutBase::GetBindingDataSize() const {
@ -532,31 +499,29 @@ namespace dawn_native {
// Followed by: // Followed by:
// |---------buffer size array--------| // |---------buffer size array--------|
// |-uint64_t[mUnverifiedBufferCount]-| // |-uint64_t[mUnverifiedBufferCount]-|
size_t objectPointerStart = static_cast<uint32_t>(mBufferCount) * sizeof(BufferBindingData); size_t objectPointerStart = mBindingCounts.bufferCount * sizeof(BufferBindingData);
ASSERT(IsAligned(objectPointerStart, alignof(Ref<ObjectBase>))); ASSERT(IsAligned(objectPointerStart, alignof(Ref<ObjectBase>)));
size_t bufferSizeArrayStart = Align( size_t bufferSizeArrayStart =
objectPointerStart + static_cast<uint32_t>(mBindingCount) * sizeof(Ref<ObjectBase>), Align(objectPointerStart + mBindingCounts.totalCount * sizeof(Ref<ObjectBase>),
sizeof(uint64_t)); sizeof(uint64_t));
ASSERT(IsAligned(bufferSizeArrayStart, alignof(uint64_t))); ASSERT(IsAligned(bufferSizeArrayStart, alignof(uint64_t)));
return bufferSizeArrayStart + mUnverifiedBufferCount * sizeof(uint64_t); return bufferSizeArrayStart + mBindingCounts.unverifiedBufferCount * sizeof(uint64_t);
} }
BindGroupLayoutBase::BindingDataPointers BindGroupLayoutBase::ComputeBindingDataPointers( BindGroupLayoutBase::BindingDataPointers BindGroupLayoutBase::ComputeBindingDataPointers(
void* dataStart) const { void* dataStart) const {
BufferBindingData* bufferData = reinterpret_cast<BufferBindingData*>(dataStart); BufferBindingData* bufferData = reinterpret_cast<BufferBindingData*>(dataStart);
auto bindings = auto bindings = reinterpret_cast<Ref<ObjectBase>*>(bufferData + mBindingCounts.bufferCount);
reinterpret_cast<Ref<ObjectBase>*>(bufferData + static_cast<uint32_t>(mBufferCount)); uint64_t* unverifiedBufferSizes = AlignPtr(
uint64_t* unverifiedBufferSizes = reinterpret_cast<uint64_t*>(bindings + mBindingCounts.totalCount), sizeof(uint64_t));
AlignPtr(reinterpret_cast<uint64_t*>(bindings + static_cast<uint32_t>(mBindingCount)),
sizeof(uint64_t));
ASSERT(IsPtrAligned(bufferData, alignof(BufferBindingData))); ASSERT(IsPtrAligned(bufferData, alignof(BufferBindingData)));
ASSERT(IsPtrAligned(bindings, alignof(Ref<ObjectBase>))); ASSERT(IsPtrAligned(bindings, alignof(Ref<ObjectBase>)));
ASSERT(IsPtrAligned(unverifiedBufferSizes, alignof(uint64_t))); ASSERT(IsPtrAligned(unverifiedBufferSizes, alignof(uint64_t)));
return {{bufferData, mBufferCount}, return {{bufferData, GetBufferCount()},
{bindings, mBindingCount}, {bindings, GetBindingCount()},
{unverifiedBufferSizes, mUnverifiedBufferCount}}; {unverifiedBufferSizes, mBindingCounts.unverifiedBufferCount}};
} }
} // namespace dawn_native } // namespace dawn_native

View File

@ -79,13 +79,16 @@ namespace dawn_native {
}; };
BindingIndex GetBindingCount() const; BindingIndex GetBindingCount() const;
// Returns |BindingIndex| because buffers are packed at the front.
BindingIndex GetBufferCount() const; BindingIndex GetBufferCount() const;
// Returns |BindingIndex| because dynamic buffers are packed at the front. // Returns |BindingIndex| because dynamic buffers are packed at the front.
BindingIndex GetDynamicBufferCount() const; BindingIndex GetDynamicBufferCount() const;
uint32_t GetDynamicUniformBufferCount() const;
uint32_t GetDynamicStorageBufferCount() const;
uint32_t GetUnverifiedBufferCount() const; uint32_t GetUnverifiedBufferCount() const;
// Used to get counts and validate them in pipeline layout creation. Other getters
// should be used to get typed integer counts.
const BindingCounts& GetBindingCountInfo() const;
struct BufferBindingData { struct BufferBindingData {
uint64_t offset; uint64_t offset;
uint64_t size; uint64_t size;
@ -120,12 +123,7 @@ namespace dawn_native {
private: private:
BindGroupLayoutBase(DeviceBase* device, ObjectBase::ErrorTag tag); BindGroupLayoutBase(DeviceBase* device, ObjectBase::ErrorTag tag);
BindingIndex mBindingCount; BindingCounts mBindingCounts = {};
BindingIndex mBufferCount{0}; // |BindingIndex| because buffers are packed at the front.
uint32_t mUnverifiedBufferCount = 0; // Buffers with minimum buffer size unspecified
uint32_t mDynamicUniformBufferCount = 0;
uint32_t mDynamicStorageBufferCount = 0;
ityp::array<BindingIndex, BindingInfo, kMaxBindingsPerGroup> mBindingInfo; ityp::array<BindingIndex, BindingInfo, kMaxBindingsPerGroup> mBindingInfo;
// Map from BindGroupLayoutEntry.binding to packed indices. // Map from BindGroupLayoutEntry.binding to packed indices.

View File

@ -0,0 +1,137 @@
// Copyright 2020 The Dawn Authors
//
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
#include "dawn_native/BindingInfo.h"
namespace dawn_native {
void IncrementBindingCounts(BindingCounts* bindingCounts, const BindGroupLayoutEntry& entry) {
bindingCounts->totalCount += 1;
uint32_t PerStageBindingCounts::*perStageBindingCountMember = nullptr;
switch (entry.type) {
case wgpu::BindingType::UniformBuffer:
++bindingCounts->bufferCount;
if (entry.hasDynamicOffset) {
++bindingCounts->dynamicUniformBufferCount;
}
if (entry.minBufferBindingSize == 0) {
++bindingCounts->unverifiedBufferCount;
}
perStageBindingCountMember = &PerStageBindingCounts::uniformBufferCount;
break;
case wgpu::BindingType::StorageBuffer:
case wgpu::BindingType::ReadonlyStorageBuffer:
++bindingCounts->bufferCount;
if (entry.hasDynamicOffset) {
++bindingCounts->dynamicStorageBufferCount;
}
if (entry.minBufferBindingSize == 0) {
++bindingCounts->unverifiedBufferCount;
}
perStageBindingCountMember = &PerStageBindingCounts::storageBufferCount;
break;
case wgpu::BindingType::SampledTexture:
perStageBindingCountMember = &PerStageBindingCounts::sampledTextureCount;
break;
case wgpu::BindingType::Sampler:
case wgpu::BindingType::ComparisonSampler:
perStageBindingCountMember = &PerStageBindingCounts::samplerCount;
break;
case wgpu::BindingType::ReadonlyStorageTexture:
case wgpu::BindingType::WriteonlyStorageTexture:
perStageBindingCountMember = &PerStageBindingCounts::storageTextureCount;
break;
case wgpu::BindingType::StorageTexture:
default:
UNREACHABLE();
break;
}
ASSERT(perStageBindingCountMember != nullptr);
for (SingleShaderStage stage : IterateStages(entry.visibility)) {
++(bindingCounts->perStage[stage].*perStageBindingCountMember);
}
}
void AccumulateBindingCounts(BindingCounts* bindingCounts, const BindingCounts& rhs) {
bindingCounts->totalCount += rhs.totalCount;
bindingCounts->bufferCount += rhs.bufferCount;
bindingCounts->unverifiedBufferCount += rhs.unverifiedBufferCount;
bindingCounts->dynamicUniformBufferCount += rhs.dynamicUniformBufferCount;
bindingCounts->dynamicStorageBufferCount += rhs.dynamicStorageBufferCount;
for (SingleShaderStage stage : IterateStages(kAllStages)) {
bindingCounts->perStage[stage].sampledTextureCount +=
rhs.perStage[stage].sampledTextureCount;
bindingCounts->perStage[stage].samplerCount += rhs.perStage[stage].samplerCount;
bindingCounts->perStage[stage].storageBufferCount +=
rhs.perStage[stage].storageBufferCount;
bindingCounts->perStage[stage].storageTextureCount +=
rhs.perStage[stage].storageTextureCount;
bindingCounts->perStage[stage].uniformBufferCount +=
rhs.perStage[stage].uniformBufferCount;
}
}
MaybeError ValidateBindingCounts(const BindingCounts& bindingCounts) {
if (bindingCounts.dynamicUniformBufferCount > kMaxDynamicUniformBuffersPerPipelineLayout) {
return DAWN_VALIDATION_ERROR(
"The number of dynamic uniform buffers exceeds the maximum per-pipeline-layout "
"limit");
}
if (bindingCounts.dynamicStorageBufferCount > kMaxDynamicStorageBuffersPerPipelineLayout) {
return DAWN_VALIDATION_ERROR(
"The number of dynamic storage buffers exceeds the maximum per-pipeline-layout "
"limit");
}
for (SingleShaderStage stage : IterateStages(kAllStages)) {
if (bindingCounts.perStage[stage].sampledTextureCount >
kMaxSampledTexturesPerShaderStage) {
return DAWN_VALIDATION_ERROR(
"The number of sampled textures exceeds the maximum "
"per-stage limit.");
}
if (bindingCounts.perStage[stage].samplerCount > kMaxSamplersPerShaderStage) {
return DAWN_VALIDATION_ERROR(
"The number of samplers exceeds the maximum per-stage limit.");
}
if (bindingCounts.perStage[stage].storageBufferCount >
kMaxStorageBuffersPerShaderStage) {
return DAWN_VALIDATION_ERROR(
"The number of storage buffers exceeds the maximum per-stage limit.");
}
if (bindingCounts.perStage[stage].storageTextureCount >
kMaxStorageTexturesPerShaderStage) {
return DAWN_VALIDATION_ERROR(
"The number of storage textures exceeds the maximum per-stage limit.");
}
if (bindingCounts.perStage[stage].uniformBufferCount >
kMaxUniformBuffersPerShaderStage) {
return DAWN_VALIDATION_ERROR(
"The number of uniform buffers exceeds the maximum per-stage limit.");
}
}
return {};
}
} // namespace dawn_native

View File

@ -18,7 +18,10 @@
#include "common/Constants.h" #include "common/Constants.h"
#include "common/TypedInteger.h" #include "common/TypedInteger.h"
#include "common/ityp_array.h" #include "common/ityp_array.h"
#include "dawn_native/Error.h"
#include "dawn_native/Format.h" #include "dawn_native/Format.h"
#include "dawn_native/PerStage.h"
#include "dawn_native/dawn_platform.h" #include "dawn_native/dawn_platform.h"
#include <cstdint> #include <cstdint>
@ -48,6 +51,27 @@ namespace dawn_native {
uint64_t minBufferBindingSize = 0; uint64_t minBufferBindingSize = 0;
}; };
struct PerStageBindingCounts {
uint32_t sampledTextureCount;
uint32_t samplerCount;
uint32_t storageBufferCount;
uint32_t storageTextureCount;
uint32_t uniformBufferCount;
};
struct BindingCounts {
uint32_t totalCount;
uint32_t bufferCount;
uint32_t unverifiedBufferCount; // Buffers with minimum buffer size unspecified
uint32_t dynamicUniformBufferCount;
uint32_t dynamicStorageBufferCount;
PerStage<PerStageBindingCounts> perStage;
};
void IncrementBindingCounts(BindingCounts* bindingCounts, const BindGroupLayoutEntry& entry);
void AccumulateBindingCounts(BindingCounts* bindingCounts, const BindingCounts& rhs);
MaybeError ValidateBindingCounts(const BindingCounts& bindingCounts);
// For buffer size validation // For buffer size validation
using RequiredBufferSizes = ityp::array<BindGroupIndex, std::vector<uint64_t>, kMaxBindGroups>; using RequiredBufferSizes = ityp::array<BindGroupIndex, std::vector<uint64_t>, kMaxBindGroups>;

View File

@ -35,6 +35,7 @@ target_sources(dawn_native PRIVATE
"BindGroupLayout.cpp" "BindGroupLayout.cpp"
"BindGroupLayout.h" "BindGroupLayout.h"
"BindGroupTracker.h" "BindGroupTracker.h"
"BindingInfo.cpp"
"BindingInfo.h" "BindingInfo.h"
"BuddyAllocator.cpp" "BuddyAllocator.cpp"
"BuddyAllocator.h" "BuddyAllocator.h"

View File

@ -70,24 +70,14 @@ namespace dawn_native {
return DAWN_VALIDATION_ERROR("too many bind group layouts"); return DAWN_VALIDATION_ERROR("too many bind group layouts");
} }
uint32_t totalDynamicUniformBufferCount = 0; BindingCounts bindingCounts = {};
uint32_t totalDynamicStorageBufferCount = 0;
for (uint32_t i = 0; i < descriptor->bindGroupLayoutCount; ++i) { for (uint32_t i = 0; i < descriptor->bindGroupLayoutCount; ++i) {
DAWN_TRY(device->ValidateObject(descriptor->bindGroupLayouts[i])); DAWN_TRY(device->ValidateObject(descriptor->bindGroupLayouts[i]));
totalDynamicUniformBufferCount += AccumulateBindingCounts(&bindingCounts,
descriptor->bindGroupLayouts[i]->GetDynamicUniformBufferCount(); descriptor->bindGroupLayouts[i]->GetBindingCountInfo());
totalDynamicStorageBufferCount +=
descriptor->bindGroupLayouts[i]->GetDynamicStorageBufferCount();
}
if (totalDynamicUniformBufferCount > kMaxDynamicUniformBufferCount) {
return DAWN_VALIDATION_ERROR("too many dynamic uniform buffers in pipeline layout");
}
if (totalDynamicStorageBufferCount > kMaxDynamicStorageBufferCount) {
return DAWN_VALIDATION_ERROR("too many dynamic storage buffers in pipeline layout");
} }
DAWN_TRY(ValidateBindingCounts(bindingCounts));
return {}; return {};
} }
@ -140,6 +130,7 @@ namespace dawn_native {
// A counter of how many bindings we've populated in |entryData| // A counter of how many bindings we've populated in |entryData|
ityp::array<BindGroupIndex, BindingIndex, kMaxBindGroups> entryCounts = {}; ityp::array<BindGroupIndex, BindingIndex, kMaxBindGroups> entryCounts = {};
BindingCounts bindingCounts = {};
BindGroupIndex bindGroupLayoutCount(0); BindGroupIndex bindGroupLayoutCount(0);
for (uint32_t moduleIndex = 0; moduleIndex < count; ++moduleIndex) { for (uint32_t moduleIndex = 0; moduleIndex < count; ++moduleIndex) {
const ShaderModuleBase* module = modules[moduleIndex]; const ShaderModuleBase* module = modules[moduleIndex];
@ -201,6 +192,7 @@ namespace dawn_native {
} }
} }
IncrementBindingCounts(&bindingCounts, bindingSlot);
BindingIndex currentBindingCount = entryCounts[group]; BindingIndex currentBindingCount = entryCounts[group];
entryData[group][currentBindingCount] = bindingSlot; entryData[group][currentBindingCount] = bindingSlot;
@ -214,6 +206,8 @@ namespace dawn_native {
} }
} }
DAWN_TRY(ValidateBindingCounts(bindingCounts));
ityp::array<BindGroupIndex, BindGroupLayoutBase*, kMaxBindGroups> bindGroupLayouts = {}; ityp::array<BindGroupIndex, BindGroupLayoutBase*, kMaxBindGroups> bindGroupLayouts = {};
for (BindGroupIndex group(0); group < bindGroupLayoutCount; ++group) { for (BindGroupIndex group(0); group < bindGroupLayoutCount; ++group) {
BindGroupLayoutDescriptor desc = {}; BindGroupLayoutDescriptor desc = {};

View File

@ -28,7 +28,7 @@ TEST_P(PipelineLayoutTests, DynamicBuffersOverflow) {
wgpu::BindGroupLayout bglA; wgpu::BindGroupLayout bglA;
{ {
std::vector<wgpu::BindGroupLayoutEntry> entries; std::vector<wgpu::BindGroupLayoutEntry> entries;
for (uint32_t i = 0; i < kMaxDynamicStorageBufferCount; i++) { for (uint32_t i = 0; i < kMaxDynamicStorageBuffersPerPipelineLayout; i++) {
entries.push_back( entries.push_back(
{i, wgpu::ShaderStage::Compute, wgpu::BindingType::StorageBuffer, true}); {i, wgpu::ShaderStage::Compute, wgpu::BindingType::StorageBuffer, true});
} }
@ -40,10 +40,11 @@ TEST_P(PipelineLayoutTests, DynamicBuffersOverflow) {
} }
// Create the second bind group layout that has one non-dynamic buffer binding. // Create the second bind group layout that has one non-dynamic buffer binding.
// It is in the fragment stage to avoid the max per-stage storage buffer limit.
wgpu::BindGroupLayout bglB; wgpu::BindGroupLayout bglB;
{ {
wgpu::BindGroupLayoutDescriptor descriptor; wgpu::BindGroupLayoutDescriptor descriptor;
wgpu::BindGroupLayoutEntry entry = {0, wgpu::ShaderStage::Compute, wgpu::BindGroupLayoutEntry entry = {0, wgpu::ShaderStage::Fragment,
wgpu::BindingType::StorageBuffer, false}; wgpu::BindingType::StorageBuffer, false};
descriptor.entryCount = 1; descriptor.entryCount = 1;
descriptor.entries = &entry; descriptor.entries = &entry;

View File

@ -485,6 +485,13 @@ TEST_F(BindGroupValidationTest, ErrorLayout) {
class BindGroupLayoutValidationTest : public ValidationTest { class BindGroupLayoutValidationTest : public ValidationTest {
public: public:
wgpu::BindGroupLayout MakeBindGroupLayout(wgpu::BindGroupLayoutEntry* binding, uint32_t count) {
wgpu::BindGroupLayoutDescriptor descriptor;
descriptor.entryCount = count;
descriptor.entries = binding;
return device.CreateBindGroupLayout(&descriptor);
}
void TestCreateBindGroupLayout(wgpu::BindGroupLayoutEntry* binding, void TestCreateBindGroupLayout(wgpu::BindGroupLayoutEntry* binding,
uint32_t count, uint32_t count,
bool expected) { bool expected) {
@ -545,8 +552,14 @@ TEST_F(BindGroupLayoutValidationTest, BindGroupLayoutEntryUnbounded) {
TEST_F(BindGroupLayoutValidationTest, BindGroupLayoutMaxBindings) { TEST_F(BindGroupLayoutValidationTest, BindGroupLayoutMaxBindings) {
wgpu::BindGroupLayoutEntry entries[kMaxBindingsPerGroup + 1]; wgpu::BindGroupLayoutEntry entries[kMaxBindingsPerGroup + 1];
wgpu::BindingType bindingsTypes[3] = {wgpu::BindingType::UniformBuffer,
wgpu::BindingType::SampledTexture,
wgpu::BindingType::Sampler};
for (uint32_t i = 0; i < kMaxBindingsPerGroup + 1; i++) { for (uint32_t i = 0; i < kMaxBindingsPerGroup + 1; i++) {
entries[i].type = wgpu::BindingType::UniformBuffer; // Alternate between uniform/sampled tex/sampler to avoid per-stage limits.
// Note: This is a temporary test and will be removed once the kMaxBindingsPerGroup
// limit is lifted.
entries[i].type = bindingsTypes[i % 3];
entries[i].binding = i; entries[i].binding = i;
entries[i].visibility = wgpu::ShaderStage::Compute; entries[i].visibility = wgpu::ShaderStage::Compute;
} }
@ -634,6 +647,96 @@ TEST_F(BindGroupLayoutValidationTest, BindGroupLayoutVisibilityNoneExpectsBindGr
ASSERT_DEVICE_ERROR(utils::MakeBindGroup(device, bgl, {{0, buffer}})); ASSERT_DEVICE_ERROR(utils::MakeBindGroup(device, bgl, {{0, buffer}}));
} }
TEST_F(BindGroupLayoutValidationTest, PerStageLimits) {
struct TestInfo {
uint32_t maxCount;
wgpu::BindingType bindingType;
wgpu::BindingType otherBindingType;
};
constexpr TestInfo kTestInfos[] = {
{kMaxSampledTexturesPerShaderStage, wgpu::BindingType::SampledTexture,
wgpu::BindingType::UniformBuffer},
{kMaxSamplersPerShaderStage, wgpu::BindingType::Sampler, wgpu::BindingType::UniformBuffer},
{kMaxSamplersPerShaderStage, wgpu::BindingType::ComparisonSampler,
wgpu::BindingType::UniformBuffer},
{kMaxStorageBuffersPerShaderStage, wgpu::BindingType::StorageBuffer,
wgpu::BindingType::UniformBuffer},
{kMaxStorageTexturesPerShaderStage, wgpu::BindingType::ReadonlyStorageTexture,
wgpu::BindingType::UniformBuffer},
{kMaxStorageTexturesPerShaderStage, wgpu::BindingType::WriteonlyStorageTexture,
wgpu::BindingType::UniformBuffer},
{kMaxUniformBuffersPerShaderStage, wgpu::BindingType::UniformBuffer,
wgpu::BindingType::SampledTexture},
};
for (TestInfo info : kTestInfos) {
wgpu::BindGroupLayout bgl[2];
std::vector<wgpu::BindGroupLayoutEntry> maxBindings;
auto PopulateEntry = [](wgpu::BindGroupLayoutEntry entry) {
switch (entry.type) {
case wgpu::BindingType::ReadonlyStorageTexture:
case wgpu::BindingType::WriteonlyStorageTexture:
entry.storageTextureFormat = wgpu::TextureFormat::RGBA8Unorm;
break;
default:
break;
}
return entry;
};
for (uint32_t i = 0; i < info.maxCount; ++i) {
maxBindings.push_back(PopulateEntry({i, wgpu::ShaderStage::Compute, info.bindingType}));
}
// Creating with the maxes works.
bgl[0] = MakeBindGroupLayout(maxBindings.data(), maxBindings.size());
// Adding an extra binding of a different type works.
{
std::vector<wgpu::BindGroupLayoutEntry> bindings = maxBindings;
bindings.push_back(
PopulateEntry({info.maxCount, wgpu::ShaderStage::Compute, info.otherBindingType}));
MakeBindGroupLayout(bindings.data(), bindings.size());
}
// Adding an extra binding of the maxed type in a different stage works
{
std::vector<wgpu::BindGroupLayoutEntry> bindings = maxBindings;
bindings.push_back(
PopulateEntry({info.maxCount, wgpu::ShaderStage::Fragment, info.bindingType}));
MakeBindGroupLayout(bindings.data(), bindings.size());
}
// Adding an extra binding of the maxed type and stage exceeds the per stage limit.
{
std::vector<wgpu::BindGroupLayoutEntry> bindings = maxBindings;
bindings.push_back(
PopulateEntry({info.maxCount, wgpu::ShaderStage::Compute, info.bindingType}));
ASSERT_DEVICE_ERROR(MakeBindGroupLayout(bindings.data(), bindings.size()));
}
// Creating a pipeline layout from the valid BGL works.
TestCreatePipelineLayout(bgl, 1, true);
// Adding an extra binding of a different type in a different BGL works
bgl[1] = utils::MakeBindGroupLayout(
device, {PopulateEntry({0, wgpu::ShaderStage::Compute, info.otherBindingType})});
TestCreatePipelineLayout(bgl, 2, true);
// Adding an extra binding of the maxed type in a different stage works
bgl[1] = utils::MakeBindGroupLayout(
device, {PopulateEntry({0, wgpu::ShaderStage::Fragment, info.bindingType})});
TestCreatePipelineLayout(bgl, 2, true);
// Adding an extra binding of the maxed type in a different BGL exceeds the per stage limit.
bgl[1] = utils::MakeBindGroupLayout(
device, {PopulateEntry({0, wgpu::ShaderStage::Compute, info.bindingType})});
TestCreatePipelineLayout(bgl, 2, false);
}
}
// Check that dynamic buffer numbers exceed maximum value in one bind group layout. // Check that dynamic buffer numbers exceed maximum value in one bind group layout.
TEST_F(BindGroupLayoutValidationTest, DynamicBufferNumberLimit) { TEST_F(BindGroupLayoutValidationTest, DynamicBufferNumberLimit) {
wgpu::BindGroupLayout bgl[2]; wgpu::BindGroupLayout bgl[2];
@ -641,49 +744,49 @@ TEST_F(BindGroupLayoutValidationTest, DynamicBufferNumberLimit) {
std::vector<wgpu::BindGroupLayoutEntry> maxStorageDB; std::vector<wgpu::BindGroupLayoutEntry> maxStorageDB;
std::vector<wgpu::BindGroupLayoutEntry> maxReadonlyStorageDB; std::vector<wgpu::BindGroupLayoutEntry> maxReadonlyStorageDB;
for (uint32_t i = 0; i < kMaxDynamicUniformBufferCount; ++i) { // In this test, we use all the same shader stage. Ensure that this does not exceed the
// per-stage limit.
static_assert(kMaxDynamicUniformBuffersPerPipelineLayout <= kMaxUniformBuffersPerShaderStage,
"");
static_assert(kMaxDynamicStorageBuffersPerPipelineLayout <= kMaxStorageBuffersPerShaderStage,
"");
for (uint32_t i = 0; i < kMaxDynamicUniformBuffersPerPipelineLayout; ++i) {
maxUniformDB.push_back( maxUniformDB.push_back(
{i, wgpu::ShaderStage::Compute, wgpu::BindingType::UniformBuffer, true}); {i, wgpu::ShaderStage::Compute, wgpu::BindingType::UniformBuffer, true});
} }
for (uint32_t i = 0; i < kMaxDynamicStorageBufferCount; ++i) { for (uint32_t i = 0; i < kMaxDynamicStorageBuffersPerPipelineLayout; ++i) {
maxStorageDB.push_back( maxStorageDB.push_back(
{i, wgpu::ShaderStage::Compute, wgpu::BindingType::StorageBuffer, true}); {i, wgpu::ShaderStage::Compute, wgpu::BindingType::StorageBuffer, true});
} }
for (uint32_t i = 0; i < kMaxDynamicStorageBufferCount; ++i) { for (uint32_t i = 0; i < kMaxDynamicStorageBuffersPerPipelineLayout; ++i) {
maxReadonlyStorageDB.push_back( maxReadonlyStorageDB.push_back(
{i, wgpu::ShaderStage::Compute, wgpu::BindingType::ReadonlyStorageBuffer, true}); {i, wgpu::ShaderStage::Compute, wgpu::BindingType::ReadonlyStorageBuffer, true});
} }
auto MakeBindGroupLayout = [&](wgpu::BindGroupLayoutEntry* binding, // Test creating with the maxes works
uint32_t count) -> wgpu::BindGroupLayout {
wgpu::BindGroupLayoutDescriptor descriptor;
descriptor.entryCount = count;
descriptor.entries = binding;
return device.CreateBindGroupLayout(&descriptor);
};
{ {
bgl[0] = MakeBindGroupLayout(maxUniformDB.data(), maxUniformDB.size()); bgl[0] = MakeBindGroupLayout(maxUniformDB.data(), maxUniformDB.size());
bgl[1] = MakeBindGroupLayout(maxStorageDB.data(), maxStorageDB.size()); TestCreatePipelineLayout(bgl, 1, true);
TestCreatePipelineLayout(bgl, 2, true); bgl[0] = MakeBindGroupLayout(maxStorageDB.data(), maxStorageDB.size());
TestCreatePipelineLayout(bgl, 1, true);
bgl[0] = MakeBindGroupLayout(maxReadonlyStorageDB.data(), maxReadonlyStorageDB.size());
TestCreatePipelineLayout(bgl, 1, true);
} }
{ // The following tests exceed the per-pipeline layout limits. We use the Fragment stage to
bgl[0] = MakeBindGroupLayout(maxUniformDB.data(), maxUniformDB.size()); // ensure we don't hit the per-stage limit.
bgl[1] = MakeBindGroupLayout(maxReadonlyStorageDB.data(), maxReadonlyStorageDB.size());
TestCreatePipelineLayout(bgl, 2, true);
}
// Check dynamic uniform buffers exceed maximum in pipeline layout. // Check dynamic uniform buffers exceed maximum in pipeline layout.
{ {
bgl[0] = MakeBindGroupLayout(maxUniformDB.data(), maxUniformDB.size()); bgl[0] = MakeBindGroupLayout(maxUniformDB.data(), maxUniformDB.size());
bgl[1] = utils::MakeBindGroupLayout( bgl[1] = utils::MakeBindGroupLayout(
device, { device, {
{0, wgpu::ShaderStage::Compute, wgpu::BindingType::UniformBuffer, true}, {0, wgpu::ShaderStage::Fragment, wgpu::BindingType::UniformBuffer, true},
}); });
TestCreatePipelineLayout(bgl, 2, false); TestCreatePipelineLayout(bgl, 2, false);
@ -694,7 +797,7 @@ TEST_F(BindGroupLayoutValidationTest, DynamicBufferNumberLimit) {
bgl[0] = MakeBindGroupLayout(maxStorageDB.data(), maxStorageDB.size()); bgl[0] = MakeBindGroupLayout(maxStorageDB.data(), maxStorageDB.size());
bgl[1] = utils::MakeBindGroupLayout( bgl[1] = utils::MakeBindGroupLayout(
device, { device, {
{0, wgpu::ShaderStage::Compute, wgpu::BindingType::StorageBuffer, true}, {0, wgpu::ShaderStage::Fragment, wgpu::BindingType::StorageBuffer, true},
}); });
TestCreatePipelineLayout(bgl, 2, false); TestCreatePipelineLayout(bgl, 2, false);
@ -706,7 +809,7 @@ TEST_F(BindGroupLayoutValidationTest, DynamicBufferNumberLimit) {
bgl[1] = utils::MakeBindGroupLayout( bgl[1] = utils::MakeBindGroupLayout(
device, device,
{ {
{0, wgpu::ShaderStage::Compute, wgpu::BindingType::ReadonlyStorageBuffer, true}, {0, wgpu::ShaderStage::Fragment, wgpu::BindingType::ReadonlyStorageBuffer, true},
}); });
TestCreatePipelineLayout(bgl, 2, false); TestCreatePipelineLayout(bgl, 2, false);
@ -719,7 +822,7 @@ TEST_F(BindGroupLayoutValidationTest, DynamicBufferNumberLimit) {
bgl[1] = utils::MakeBindGroupLayout( bgl[1] = utils::MakeBindGroupLayout(
device, device,
{ {
{0, wgpu::ShaderStage::Compute, wgpu::BindingType::ReadonlyStorageBuffer, true}, {0, wgpu::ShaderStage::Fragment, wgpu::BindingType::ReadonlyStorageBuffer, true},
}); });
TestCreatePipelineLayout(bgl, 2, false); TestCreatePipelineLayout(bgl, 2, false);
@ -727,21 +830,24 @@ TEST_F(BindGroupLayoutValidationTest, DynamicBufferNumberLimit) {
// Check dynamic uniform buffers exceed maximum in bind group layout. // Check dynamic uniform buffers exceed maximum in bind group layout.
{ {
maxUniformDB.push_back({kMaxDynamicUniformBufferCount, wgpu::ShaderStage::Compute, maxUniformDB.push_back({kMaxDynamicUniformBuffersPerPipelineLayout,
wgpu::BindingType::UniformBuffer, true}); wgpu::ShaderStage::Fragment, wgpu::BindingType::UniformBuffer,
true});
TestCreateBindGroupLayout(maxUniformDB.data(), maxUniformDB.size(), false); TestCreateBindGroupLayout(maxUniformDB.data(), maxUniformDB.size(), false);
} }
// Check dynamic storage buffers exceed maximum in bind group layout. // Check dynamic storage buffers exceed maximum in bind group layout.
{ {
maxStorageDB.push_back({kMaxDynamicStorageBufferCount, wgpu::ShaderStage::Compute, maxStorageDB.push_back({kMaxDynamicStorageBuffersPerPipelineLayout,
wgpu::BindingType::StorageBuffer, true}); wgpu::ShaderStage::Fragment, wgpu::BindingType::StorageBuffer,
true});
TestCreateBindGroupLayout(maxStorageDB.data(), maxStorageDB.size(), false); TestCreateBindGroupLayout(maxStorageDB.data(), maxStorageDB.size(), false);
} }
// Check dynamic readonly storage buffers exceed maximum in bind group layout. // Check dynamic readonly storage buffers exceed maximum in bind group layout.
{ {
maxReadonlyStorageDB.push_back({kMaxDynamicStorageBufferCount, wgpu::ShaderStage::Compute, maxReadonlyStorageDB.push_back({kMaxDynamicStorageBuffersPerPipelineLayout,
wgpu::ShaderStage::Fragment,
wgpu::BindingType::ReadonlyStorageBuffer, true}); wgpu::BindingType::ReadonlyStorageBuffer, true});
TestCreateBindGroupLayout(maxReadonlyStorageDB.data(), maxReadonlyStorageDB.size(), false); TestCreateBindGroupLayout(maxReadonlyStorageDB.data(), maxReadonlyStorageDB.size(), false);
} }

View File

@ -521,20 +521,18 @@ class MinBufferSizeDefaultLayoutTests : public MinBufferSizeTestsBase {
// Various bindings in std140 have correct minimum size reflection // Various bindings in std140 have correct minimum size reflection
TEST_F(MinBufferSizeDefaultLayoutTests, std140Inferred) { TEST_F(MinBufferSizeDefaultLayoutTests, std140Inferred) {
CheckShaderBindingSizeReflection("std140", {{{0, 0, "float a", 4}, CheckShaderBindingSizeReflection(
{0, 1, "float b[]", 16}, "std140", {{{0, 0, "float a", 4}, {0, 1, "float b[]", 16}, {0, 2, "mat2 c", 32}}});
{0, 2, "mat2 c", 32}, CheckShaderBindingSizeReflection("std140", {{{0, 3, "int d; float e[]", 32},
{0, 3, "int d; float e[]", 32},
{0, 4, "ThreeFloats f", 12}, {0, 4, "ThreeFloats f", 12},
{0, 5, "ThreeFloats g[]", 16}}}); {0, 5, "ThreeFloats g[]", 16}}});
} }
// Various bindings in std430 have correct minimum size reflection // Various bindings in std430 have correct minimum size reflection
TEST_F(MinBufferSizeDefaultLayoutTests, std430Inferred) { TEST_F(MinBufferSizeDefaultLayoutTests, std430Inferred) {
CheckShaderBindingSizeReflection("std430", {{{0, 0, "float a", 4}, CheckShaderBindingSizeReflection(
{0, 1, "float b[]", 4}, "std430", {{{0, 0, "float a", 4}, {0, 1, "float b[]", 4}, {0, 2, "mat2 c", 16}}});
{0, 2, "mat2 c", 16}, CheckShaderBindingSizeReflection("std430", {{{0, 3, "int d; float e[]", 8},
{0, 3, "int d; float e[]", 8},
{0, 4, "ThreeFloats f", 12}, {0, 4, "ThreeFloats f", 12},
{0, 5, "ThreeFloats g[]", 12}}}); {0, 5, "ThreeFloats g[]", 12}}});
} }
@ -558,18 +556,20 @@ TEST_F(MinBufferSizeDefaultLayoutTests, std430BindingTypes) {
TEST_F(MinBufferSizeDefaultLayoutTests, std140MultipleBindGroups) { TEST_F(MinBufferSizeDefaultLayoutTests, std140MultipleBindGroups) {
CheckShaderBindingSizeReflection("std140", CheckShaderBindingSizeReflection("std140",
{{{0, 0, "float a", 4}, {0, 1, "float b[]", 16}}, {{{0, 0, "float a", 4}, {0, 1, "float b[]", 16}},
{{1, 2, "mat2 c", 32}, {1, 3, "int d; float e[]", 32}}, {{1, 2, "mat2 c", 32}, {1, 3, "int d; float e[]", 32}}});
{{2, 4, "ThreeFloats f", 12}}, CheckShaderBindingSizeReflection(
{{3, 5, "ThreeFloats g[]", 16}}}); "std140", {{{0, 4, "ThreeFloats f", 12}, {0, 1, "float b[]", 16}},
{{1, 5, "ThreeFloats g[]", 16}, {1, 3, "int d; float e[]", 32}}});
} }
// Various bindings have correct size across multiple groups // Various bindings have correct size across multiple groups
TEST_F(MinBufferSizeDefaultLayoutTests, std430MultipleBindGroups) { TEST_F(MinBufferSizeDefaultLayoutTests, std430MultipleBindGroups) {
CheckShaderBindingSizeReflection("std430", CheckShaderBindingSizeReflection("std430",
{{{0, 0, "float a", 4}, {0, 1, "float b[]", 4}}, {{{0, 0, "float a", 4}, {0, 1, "float b[]", 4}},
{{1, 2, "mat2 c", 16}, {1, 3, "int d; float e[]", 8}}, {{1, 2, "mat2 c", 16}, {1, 3, "int d; float e[]", 8}}});
{{2, 4, "ThreeFloats f", 12}}, CheckShaderBindingSizeReflection(
{{3, 5, "ThreeFloats g[]", 12}}}); "std430", {{{0, 4, "ThreeFloats f", 12}, {0, 1, "float b[]", 4}},
{{1, 5, "ThreeFloats g[]", 12}, {1, 3, "int d; float e[]", 8}}});
} }
// Minimum size should be the max requirement of both vertex and fragment stages // Minimum size should be the max requirement of both vertex and fragment stages