// Copyright 2018 The Dawn Authors // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #include "dawn_native/vulkan/BindGroupLayoutVk.h" #include "common/BitSetIterator.h" #include "dawn_native/vulkan/BindGroupVk.h" #include "dawn_native/vulkan/DescriptorSetAllocator.h" #include "dawn_native/vulkan/DeviceVk.h" #include "dawn_native/vulkan/FencedDeleter.h" #include "dawn_native/vulkan/VulkanError.h" #include namespace dawn_native { namespace vulkan { namespace { VkShaderStageFlags VulkanShaderStageFlags(wgpu::ShaderStage stages) { VkShaderStageFlags flags = 0; if (stages & wgpu::ShaderStage::Vertex) { flags |= VK_SHADER_STAGE_VERTEX_BIT; } if (stages & wgpu::ShaderStage::Fragment) { flags |= VK_SHADER_STAGE_FRAGMENT_BIT; } if (stages & wgpu::ShaderStage::Compute) { flags |= VK_SHADER_STAGE_COMPUTE_BIT; } return flags; } } // anonymous namespace VkDescriptorType VulkanDescriptorType(wgpu::BindingType type, bool isDynamic) { switch (type) { case wgpu::BindingType::UniformBuffer: if (isDynamic) { return VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER_DYNAMIC; } return VK_DESCRIPTOR_TYPE_UNIFORM_BUFFER; case wgpu::BindingType::Sampler: case wgpu::BindingType::ComparisonSampler: return VK_DESCRIPTOR_TYPE_SAMPLER; case wgpu::BindingType::SampledTexture: return VK_DESCRIPTOR_TYPE_SAMPLED_IMAGE; case wgpu::BindingType::StorageBuffer: case wgpu::BindingType::ReadonlyStorageBuffer: if (isDynamic) { return VK_DESCRIPTOR_TYPE_STORAGE_BUFFER_DYNAMIC; } return VK_DESCRIPTOR_TYPE_STORAGE_BUFFER; case wgpu::BindingType::ReadonlyStorageTexture: case wgpu::BindingType::WriteonlyStorageTexture: return VK_DESCRIPTOR_TYPE_STORAGE_IMAGE; case wgpu::BindingType::StorageTexture: default: UNREACHABLE(); } } // static ResultOrError BindGroupLayout::Create( Device* device, const BindGroupLayoutDescriptor* descriptor) { Ref bgl = AcquireRef(new BindGroupLayout(device, descriptor)); DAWN_TRY(bgl->Initialize()); return bgl.Detach(); } MaybeError BindGroupLayout::Initialize() { // Compute the bindings that will be chained in the DescriptorSetLayout create info. We add // one entry per binding set. This might be optimized by computing continuous ranges of // bindings of the same type. uint32_t numBindings = 0; std::array bindings; for (const auto& it : GetBindingMap()) { BindingNumber bindingNumber = it.first; BindingIndex bindingIndex = it.second; const BindingInfo& bindingInfo = GetBindingInfo(bindingIndex); VkDescriptorSetLayoutBinding* vkBinding = &bindings[numBindings]; vkBinding->binding = bindingNumber; vkBinding->descriptorType = VulkanDescriptorType(bindingInfo.type, bindingInfo.hasDynamicOffset); vkBinding->descriptorCount = 1; vkBinding->stageFlags = VulkanShaderStageFlags(bindingInfo.visibility); vkBinding->pImmutableSamplers = nullptr; numBindings++; } VkDescriptorSetLayoutCreateInfo createInfo; createInfo.sType = VK_STRUCTURE_TYPE_DESCRIPTOR_SET_LAYOUT_CREATE_INFO; createInfo.pNext = nullptr; createInfo.flags = 0; createInfo.bindingCount = numBindings; createInfo.pBindings = bindings.data(); Device* device = ToBackend(GetDevice()); DAWN_TRY(CheckVkSuccess(device->fn.CreateDescriptorSetLayout( device->GetVkDevice(), &createInfo, nullptr, &*mHandle), "CreateDescriptorSetLayout")); // Compute the size of descriptor pools used for this layout. std::map descriptorCountPerType; for (BindingIndex bindingIndex = 0; bindingIndex < GetBindingCount(); ++bindingIndex) { const BindingInfo& bindingInfo = GetBindingInfo(bindingIndex); VkDescriptorType vulkanType = VulkanDescriptorType(bindingInfo.type, bindingInfo.hasDynamicOffset); // map::operator[] will return 0 if the key doesn't exist. descriptorCountPerType[vulkanType]++; } // TODO(enga): Consider deduping allocators for layouts with the same descriptor type // counts. mDescriptorSetAllocator = std::make_unique(this, std::move(descriptorCountPerType)); return {}; } BindGroupLayout::BindGroupLayout(DeviceBase* device, const BindGroupLayoutDescriptor* descriptor) : BindGroupLayoutBase(device, descriptor), mBindGroupAllocator(MakeFrontendBindGroupAllocator(4096)) { } BindGroupLayout::~BindGroupLayout() { Device* device = ToBackend(GetDevice()); // DescriptorSetLayout aren't used by execution on the GPU and can be deleted at any time, // so we can destroy mHandle immediately instead of using the FencedDeleter. // (Swiftshader implements this wrong b/154522740). // In practice, the GPU is done with all descriptor sets because bind group deallocation // refs the bind group layout so that once the bind group is finished being used, we can // recycle its descriptor set. if (mHandle != VK_NULL_HANDLE) { device->fn.DestroyDescriptorSetLayout(device->GetVkDevice(), mHandle, nullptr); mHandle = VK_NULL_HANDLE; } } VkDescriptorSetLayout BindGroupLayout::GetHandle() const { return mHandle; } ResultOrError BindGroupLayout::AllocateBindGroup( Device* device, const BindGroupDescriptor* descriptor) { DescriptorSetAllocation descriptorSetAllocation; DAWN_TRY_ASSIGN(descriptorSetAllocation, mDescriptorSetAllocator->Allocate()); return mBindGroupAllocator.Allocate(device, descriptor, descriptorSetAllocation); } void BindGroupLayout::DeallocateBindGroup(BindGroup* bindGroup, DescriptorSetAllocation* descriptorSetAllocation) { mDescriptorSetAllocator->Deallocate(descriptorSetAllocation); mBindGroupAllocator.Deallocate(bindGroup); } void BindGroupLayout::FinishDeallocation(Serial completedSerial) { mDescriptorSetAllocator->FinishDeallocation(completedSerial); } }} // namespace dawn_native::vulkan