aurora: WIP cache for display lists & static buffers

This commit is contained in:
Luke Street 2022-03-15 02:18:45 -04:00
parent bbdad137af
commit 02a7b85b23
18 changed files with 263 additions and 215 deletions

View File

@ -259,8 +259,8 @@ void CCubeModel::EnableShadowMaps(const CTexture& shadowTex, const zeus::CTransf
void CCubeModel::DisableShadowMaps() { sRenderModelShadow = false; } void CCubeModel::DisableShadowMaps() { sRenderModelShadow = false; }
void CCubeModel::SetArraysCurrent() { void CCubeModel::SetArraysCurrent() {
CGX::SetArray(GX::VA_POS, x0_modelInstance.GetVertexPointer()); CGX::SetArray(GX::VA_POS, x0_modelInstance.GetVertexPointer(), true);
CGX::SetArray(GX::VA_NRM, x0_modelInstance.GetNormalPointer()); CGX::SetArray(GX::VA_NRM, x0_modelInstance.GetNormalPointer(), true);
SetStaticArraysCurrent(); SetStaticArraysCurrent();
} }
@ -280,8 +280,8 @@ void CCubeModel::SetRenderModelBlack(bool v) {
} }
void CCubeModel::SetSkinningArraysCurrent(TConstVectorRef positions, TConstVectorRef normals) { void CCubeModel::SetSkinningArraysCurrent(TConstVectorRef positions, TConstVectorRef normals) {
CGX::SetArray(GX::VA_POS, positions); CGX::SetArray(GX::VA_POS, positions, false);
CGX::SetArray(GX::VA_NRM, normals); CGX::SetArray(GX::VA_NRM, normals, false);
// colors unused // colors unused
SetStaticArraysCurrent(); SetStaticArraysCurrent();
} }
@ -294,21 +294,21 @@ void CCubeModel::SetStaticArraysCurrent() {
sUsingPackedLightmaps = false; sUsingPackedLightmaps = false;
} }
if (sUsingPackedLightmaps) { if (sUsingPackedLightmaps) {
CGX::SetArray(GX::VA_TEX0, packedTexCoords); CGX::SetArray(GX::VA_TEX0, packedTexCoords, true);
} else { } else {
CGX::SetArray(GX::VA_TEX0, texCoords); CGX::SetArray(GX::VA_TEX0, texCoords, true);
} }
// TexCoord1 is currently used for all remaining // TexCoord1 is currently used for all remaining
CGX::SetArray(GX::VA_TEX1, texCoords); CGX::SetArray(GX::VA_TEX1, texCoords, true);
CCubeMaterial::KillCachedViewDepState(); CCubeMaterial::KillCachedViewDepState();
} }
void CCubeModel::SetUsingPackedLightmaps(bool v) { void CCubeModel::SetUsingPackedLightmaps(bool v) {
sUsingPackedLightmaps = v; sUsingPackedLightmaps = v;
if (v) { if (v) {
CGX::SetArray(GX::VA_TEX0, x0_modelInstance.GetPackedTCPointer()); CGX::SetArray(GX::VA_TEX0, x0_modelInstance.GetPackedTCPointer(), true);
} else { } else {
CGX::SetArray(GX::VA_TEX0, x0_modelInstance.GetTCPointer()); CGX::SetArray(GX::VA_TEX0, x0_modelInstance.GetTCPointer(), true);
} }
} }

View File

@ -123,9 +123,9 @@ static inline void SetAlphaCompare(GX::Compare comp0, u8 ref0, GX::AlphaOp op, G
} }
template <typename T> template <typename T>
static inline void SetArray(GX::Attr attr, const std::vector<T>* data) noexcept { static inline void SetArray(GX::Attr attr, const std::vector<T>* data, bool isStatic) noexcept {
if (data != nullptr && sGXState.x0_arrayPtrs[attr - GX::VA_POS] != data) { if (data != nullptr && sGXState.x0_arrayPtrs[attr - GX::VA_POS] != data) {
GXSetArray(attr, data, sizeof(T)); GXSetArray(attr, data, isStatic ? 1 : 0);
} }
} }

View File

@ -19,11 +19,8 @@ add_library(aurora STATIC
target_compile_definitions(aurora PRIVATE IMGUI_USER_CONFIG="imconfig_user.h") # IMGUI_USE_WCHAR32 target_compile_definitions(aurora PRIVATE IMGUI_USER_CONFIG="imconfig_user.h") # IMGUI_USE_WCHAR32
target_include_directories(aurora PUBLIC include ../) target_include_directories(aurora PUBLIC include ../)
target_include_directories(aurora PRIVATE ../imgui ../extern/imgui) target_include_directories(aurora PRIVATE ../imgui ../extern/imgui)
target_include_directories(aurora PRIVATE target_link_libraries(aurora PRIVATE dawn_native dawncpp webgpu_dawn zeus logvisor SDL2-static xxhash
../extern/dawn/src absl::btree absl::flat_hash_map)
../extern/dawn/third_party/abseil-cpp
${CMAKE_CURRENT_BINARY_DIR}/dawn/gen/src) # for hacks :)
target_link_libraries(aurora PRIVATE dawn_native dawncpp webgpu_dawn zeus logvisor SDL2-static xxhash)
if (APPLE) if (APPLE)
target_compile_definitions(aurora PRIVATE DAWN_ENABLE_BACKEND_METAL) target_compile_definitions(aurora PRIVATE DAWN_ENABLE_BACKEND_METAL)
target_sources(aurora PRIVATE lib/dawn/MetalBinding.mm) target_sources(aurora PRIVATE lib/dawn/MetalBinding.mm)

View File

@ -302,9 +302,9 @@ void render(const State& state, const DrawData& data, const wgpu::RenderPassEnco
return; return;
} }
const std::array offsets{data.uniformRange.first}; const std::array offsets{data.uniformRange.offset};
pass.SetBindGroup(0, state.uniformBindGroup, offsets.size(), offsets.data()); pass.SetBindGroup(0, state.uniformBindGroup, offsets.size(), offsets.data());
pass.SetVertexBuffer(0, g_vertexBuffer, data.vertRange.first, data.vertRange.second); pass.SetVertexBuffer(0, g_vertexBuffer, data.vertRange.offset, data.vertRange.size);
pass.Draw(4); pass.Draw(4);
} }
} // namespace aurora::gfx::colored_quad } // namespace aurora::gfx::colored_quad

View File

@ -11,7 +11,7 @@
#include <deque> #include <deque>
#include <logvisor/logvisor.hpp> #include <logvisor/logvisor.hpp>
#include <thread> #include <thread>
#include <unordered_map> #include <absl/container/flat_hash_map.h>
namespace aurora::gfx { namespace aurora::gfx {
static logvisor::Module Log("aurora::gfx"); static logvisor::Module Log("aurora::gfx");
@ -187,10 +187,10 @@ std::mutex g_pipelineMutex;
static std::thread g_pipelineThread; static std::thread g_pipelineThread;
static std::atomic_bool g_pipelineThreadEnd; static std::atomic_bool g_pipelineThreadEnd;
static std::condition_variable g_pipelineCv; static std::condition_variable g_pipelineCv;
static std::unordered_map<PipelineRef, wgpu::RenderPipeline> g_pipelines; static absl::flat_hash_map<PipelineRef, wgpu::RenderPipeline> g_pipelines;
static std::deque<std::pair<PipelineRef, NewPipelineCallback>> g_queuedPipelines; static std::deque<std::pair<PipelineRef, NewPipelineCallback>> g_queuedPipelines;
static std::unordered_map<BindGroupRef, wgpu::BindGroup> g_cachedBindGroups; static absl::flat_hash_map<BindGroupRef, wgpu::BindGroup> g_cachedBindGroups;
static std::unordered_map<SamplerRef, wgpu::Sampler> g_cachedSamplers; static absl::flat_hash_map<SamplerRef, wgpu::Sampler> g_cachedSamplers;
std::atomic_uint32_t queuedPipelines; std::atomic_uint32_t queuedPipelines;
std::atomic_uint32_t createdPipelines; std::atomic_uint32_t createdPipelines;
@ -198,10 +198,12 @@ static ByteBuffer g_verts;
static ByteBuffer g_uniforms; static ByteBuffer g_uniforms;
static ByteBuffer g_indices; static ByteBuffer g_indices;
static ByteBuffer g_storage; static ByteBuffer g_storage;
static ByteBuffer g_staticStorage;
wgpu::Buffer g_vertexBuffer; wgpu::Buffer g_vertexBuffer;
wgpu::Buffer g_uniformBuffer; wgpu::Buffer g_uniformBuffer;
wgpu::Buffer g_indexBuffer; wgpu::Buffer g_indexBuffer;
wgpu::Buffer g_storageBuffer; wgpu::Buffer g_storageBuffer;
size_t g_staticStorageLastSize = 0;
static ShaderState g_state; static ShaderState g_state;
static PipelineRef g_currentPipeline; static PipelineRef g_currentPipeline;
@ -213,7 +215,7 @@ static PipelineRef find_pipeline(PipelineCreateCommand command, NewPipelineCallb
bool found = false; bool found = false;
{ {
std::scoped_lock guard{g_pipelineMutex}; std::scoped_lock guard{g_pipelineMutex};
found = g_pipelines.find(hash) != g_pipelines.end(); found = g_pipelines.contains(hash);
if (!found) { if (!found) {
const auto ref = const auto ref =
std::find_if(g_queuedPipelines.begin(), g_queuedPipelines.end(), [=](auto v) { return v.first == hash; }); std::find_if(g_queuedPipelines.begin(), g_queuedPipelines.end(), [=](auto v) { return v.first == hash; });
@ -364,11 +366,10 @@ static void pipeline_worker() {
// std::this_thread::sleep_for(std::chrono::milliseconds{1500}); // std::this_thread::sleep_for(std::chrono::milliseconds{1500});
{ {
std::scoped_lock lock{g_pipelineMutex}; std::scoped_lock lock{g_pipelineMutex};
if (g_pipelines.contains(cb.first)) { if (!g_pipelines.try_emplace(cb.first, std::move(result)).second) {
Log.report(logvisor::Fatal, FMT_STRING("Duplicate pipeline {}"), cb.first); Log.report(logvisor::Fatal, FMT_STRING("Duplicate pipeline {}"), cb.first);
unreachable(); unreachable();
} }
g_pipelines[cb.first] = result;
g_queuedPipelines.pop_front(); g_queuedPipelines.pop_front();
hasMore = !g_queuedPipelines.empty(); hasMore = !g_queuedPipelines.empty();
} }
@ -384,7 +385,7 @@ void initialize() {
const wgpu::BufferDescriptor descriptor{ const wgpu::BufferDescriptor descriptor{
.label = "Shared Uniform Buffer", .label = "Shared Uniform Buffer",
.usage = wgpu::BufferUsage::Uniform | wgpu::BufferUsage::CopyDst, .usage = wgpu::BufferUsage::Uniform | wgpu::BufferUsage::CopyDst,
.size = 134217728, // 128mb .size = 5242880, // 5mb
}; };
g_uniformBuffer = g_device.CreateBuffer(&descriptor); g_uniformBuffer = g_device.CreateBuffer(&descriptor);
} }
@ -392,7 +393,7 @@ void initialize() {
const wgpu::BufferDescriptor descriptor{ const wgpu::BufferDescriptor descriptor{
.label = "Shared Vertex Buffer", .label = "Shared Vertex Buffer",
.usage = wgpu::BufferUsage::Vertex | wgpu::BufferUsage::CopyDst, .usage = wgpu::BufferUsage::Vertex | wgpu::BufferUsage::CopyDst,
.size = 16777216, // 16mb .size = 5242880, // 5mb
}; };
g_vertexBuffer = g_device.CreateBuffer(&descriptor); g_vertexBuffer = g_device.CreateBuffer(&descriptor);
} }
@ -400,7 +401,7 @@ void initialize() {
const wgpu::BufferDescriptor descriptor{ const wgpu::BufferDescriptor descriptor{
.label = "Shared Index Buffer", .label = "Shared Index Buffer",
.usage = wgpu::BufferUsage::Index | wgpu::BufferUsage::CopyDst, .usage = wgpu::BufferUsage::Index | wgpu::BufferUsage::CopyDst,
.size = 4194304, // 4mb .size = 2097152, // 2mb
}; };
g_indexBuffer = g_device.CreateBuffer(&descriptor); g_indexBuffer = g_device.CreateBuffer(&descriptor);
} }
@ -439,18 +440,33 @@ void shutdown() {
} }
void render(const wgpu::RenderPassEncoder& pass) { void render(const wgpu::RenderPassEncoder& pass) {
const auto writeBuffer = [](ByteBuffer& buf, wgpu::Buffer& out) { const auto writeBuffer = [](ByteBuffer& buf, wgpu::Buffer& out, std::string_view label) {
const auto size = buf.size(); const auto size = buf.size();
// Log.report(logvisor::Info, FMT_STRING("{} buffer usage: {}"), label, size);
if (size > 0) { if (size > 0) {
g_queue.WriteBuffer(out, 0, buf.data(), size); g_queue.WriteBuffer(out, 0, buf.data(), size);
buf.clear(); buf.clear();
buf.reserve_extra(size); // Reserve size from previous frame buf.reserve_extra(size); // Reserve size from previous frame
} }
}; };
writeBuffer(g_verts, g_vertexBuffer); writeBuffer(g_verts, g_vertexBuffer, "Vertex");
writeBuffer(g_uniforms, g_uniformBuffer); writeBuffer(g_uniforms, g_uniformBuffer, "Uniform");
writeBuffer(g_indices, g_indexBuffer); writeBuffer(g_indices, g_indexBuffer, "Index");
writeBuffer(g_storage, g_storageBuffer); {
const auto staticSize = g_staticStorage.size();
if (staticSize > g_staticStorageLastSize) {
g_queue.WriteBuffer(g_storageBuffer, g_staticStorageLastSize, g_staticStorage.data() + g_staticStorageLastSize,
staticSize - g_staticStorageLastSize);
g_staticStorageLastSize = staticSize;
}
const auto size = g_storage.size();
if (size > 0) {
g_queue.WriteBuffer(g_storageBuffer, staticSize, g_storage.data(), size);
g_storage.clear();
g_storage.reserve_extra(size); // Reserve size from previous frame
}
// Log.report(logvisor::Info, FMT_STRING("Static storage: {}, storage: {}"), staticSize, size);
}
g_currentPipeline = UINT64_MAX; g_currentPipeline = UINT64_MAX;
@ -498,10 +514,11 @@ bool bind_pipeline(PipelineRef ref, const wgpu::RenderPassEncoder& pass) {
return true; return true;
} }
std::lock_guard guard{g_pipelineMutex}; std::lock_guard guard{g_pipelineMutex};
if (!g_pipelines.contains(ref)) { const auto it = g_pipelines.find(ref);
if (it == g_pipelines.end()) {
return false; return false;
} }
pass.SetPipeline(g_pipelines[ref]); pass.SetPipeline(it->second);
g_currentPipeline = ref; g_currentPipeline = ref;
return true; return true;
} }
@ -522,7 +539,7 @@ static inline Range push(ByteBuffer& target, const uint8_t* data, size_t length,
target.append_zeroes(padding); target.append_zeroes(padding);
} }
} }
return {begin, begin + length + padding}; return {static_cast<uint32_t>(begin), static_cast<uint32_t>(length + padding)};
} }
static inline Range map(ByteBuffer& target, size_t length, size_t alignment) { static inline Range map(ByteBuffer& target, size_t length, size_t alignment) {
size_t padding = 0; size_t padding = 0;
@ -534,7 +551,7 @@ static inline Range map(ByteBuffer& target, size_t length, size_t alignment) {
} }
auto begin = target.size(); auto begin = target.size();
target.append_zeroes(length + padding); target.append_zeroes(length + padding);
return {begin, begin + length + padding}; return {static_cast<uint32_t>(begin), static_cast<uint32_t>(length + padding)};
} }
Range push_verts(const uint8_t* data, size_t length) { return push(g_verts, data, length, 0 /* TODO? */); } Range push_verts(const uint8_t* data, size_t length) { return push(g_verts, data, length, 0 /* TODO? */); }
Range push_indices(const uint8_t* data, size_t length) { return push(g_indices, data, length, 0 /* TODO? */); } Range push_indices(const uint8_t* data, size_t length) { return push(g_indices, data, length, 0 /* TODO? */); }
@ -548,48 +565,57 @@ Range push_storage(const uint8_t* data, size_t length) {
g_device.GetLimits(&limits); g_device.GetLimits(&limits);
return push(g_storage, data, length, limits.limits.minStorageBufferOffsetAlignment); return push(g_storage, data, length, limits.limits.minStorageBufferOffsetAlignment);
} }
Range push_static_storage(const uint8_t* data, size_t length) {
wgpu::SupportedLimits limits;
g_device.GetLimits(&limits);
auto range = push(g_staticStorage, data, length, limits.limits.minStorageBufferOffsetAlignment);
range.isStatic = true;
return range;
}
std::pair<ByteBuffer, Range> map_verts(size_t length) { std::pair<ByteBuffer, Range> map_verts(size_t length) {
const auto range = map(g_verts, length, 0 /* TODO? */); const auto range = map(g_verts, length, 0 /* TODO? */);
return {ByteBuffer{g_verts.data() + range.first, range.second - range.first}, range}; return {ByteBuffer{g_verts.data() + range.offset, range.size}, range};
} }
std::pair<ByteBuffer, Range> map_indices(size_t length) { std::pair<ByteBuffer, Range> map_indices(size_t length) {
const auto range = map(g_indices, length, 0 /* TODO? */); const auto range = map(g_indices, length, 0 /* TODO? */);
return {ByteBuffer{g_indices.data() + range.first, range.second - range.first}, range}; return {ByteBuffer{g_indices.data() + range.offset, range.size}, range};
} }
std::pair<ByteBuffer, Range> map_uniform(size_t length) { std::pair<ByteBuffer, Range> map_uniform(size_t length) {
wgpu::SupportedLimits limits; wgpu::SupportedLimits limits;
g_device.GetLimits(&limits); g_device.GetLimits(&limits);
const auto range = map(g_uniforms, length, limits.limits.minUniformBufferOffsetAlignment); const auto range = map(g_uniforms, length, limits.limits.minUniformBufferOffsetAlignment);
return {ByteBuffer{g_uniforms.data() + range.first, range.second - range.first}, range}; return {ByteBuffer{g_uniforms.data() + range.offset, range.size}, range};
} }
std::pair<ByteBuffer, Range> map_storage(size_t length) { std::pair<ByteBuffer, Range> map_storage(size_t length) {
wgpu::SupportedLimits limits; wgpu::SupportedLimits limits;
g_device.GetLimits(&limits); g_device.GetLimits(&limits);
const auto range = map(g_storage, length, limits.limits.minStorageBufferOffsetAlignment); const auto range = map(g_storage, length, limits.limits.minStorageBufferOffsetAlignment);
return {ByteBuffer{g_storage.data() + range.first, range.second - range.first}, range}; return {ByteBuffer{g_storage.data() + range.offset, range.size}, range};
} }
BindGroupRef bind_group_ref(const wgpu::BindGroupDescriptor& descriptor) { BindGroupRef bind_group_ref(const wgpu::BindGroupDescriptor& descriptor) {
const auto id = xxh3_hash(descriptor); const auto id = xxh3_hash(descriptor);
if (!g_cachedBindGroups.contains(id)) { if (!g_cachedBindGroups.contains(id)) {
g_cachedBindGroups[id] = g_device.CreateBindGroup(&descriptor); g_cachedBindGroups.try_emplace(id, g_device.CreateBindGroup(&descriptor));
} }
return id; return id;
} }
const wgpu::BindGroup& find_bind_group(BindGroupRef id) { const wgpu::BindGroup& find_bind_group(BindGroupRef id) {
if (!g_cachedBindGroups.contains(id)) { const auto it = g_cachedBindGroups.find(id);
if (it == g_cachedBindGroups.end()) {
Log.report(logvisor::Fatal, FMT_STRING("get_bind_group: failed to locate {}"), id); Log.report(logvisor::Fatal, FMT_STRING("get_bind_group: failed to locate {}"), id);
unreachable(); unreachable();
} }
return g_cachedBindGroups[id]; return it->second;
} }
const wgpu::Sampler& sampler_ref(const wgpu::SamplerDescriptor& descriptor) { const wgpu::Sampler& sampler_ref(const wgpu::SamplerDescriptor& descriptor) {
const auto id = xxh3_hash(descriptor); const auto id = xxh3_hash(descriptor);
if (!g_cachedSamplers.contains(id)) { auto it = g_cachedSamplers.find(id);
g_cachedSamplers[id] = g_device.CreateSampler(&descriptor); if (it == g_cachedSamplers.end()) {
it = g_cachedSamplers.try_emplace(id, g_device.CreateSampler(&descriptor)).first;
} }
return g_cachedSamplers[id]; return it->second;
} }
uint32_t align_uniform(uint32_t value) { uint32_t align_uniform(uint32_t value) {

View File

@ -126,6 +126,7 @@ extern wgpu::Buffer g_vertexBuffer;
extern wgpu::Buffer g_uniformBuffer; extern wgpu::Buffer g_uniformBuffer;
extern wgpu::Buffer g_indexBuffer; extern wgpu::Buffer g_indexBuffer;
extern wgpu::Buffer g_storageBuffer; extern wgpu::Buffer g_storageBuffer;
extern size_t g_staticStorageLastSize;
struct TextureRef { struct TextureRef {
wgpu::Texture texture; wgpu::Texture texture;
@ -149,7 +150,14 @@ using BindGroupRef = uint64_t;
using PipelineRef = uint64_t; using PipelineRef = uint64_t;
using SamplerRef = uint64_t; using SamplerRef = uint64_t;
using ShaderRef = uint64_t; using ShaderRef = uint64_t;
using Range = std::pair<uint32_t, uint32_t>; struct Range {
uint32_t offset;
uint32_t size;
bool isStatic;
};
static inline uint32_t storage_offset(Range range) {
return range.isStatic ? range.offset : range.offset + g_staticStorageLastSize;
}
enum class ShaderType { enum class ShaderType {
Aabb, Aabb,
@ -182,9 +190,22 @@ static inline Range push_uniform(const T& data) {
} }
Range push_storage(const uint8_t* data, size_t length); Range push_storage(const uint8_t* data, size_t length);
template <typename T> template <typename T>
static inline Range push_storage(ArrayRef<T> data) {
return push_storage(reinterpret_cast<const uint8_t*>(data.data()), data.size() * sizeof(T));
}
template <typename T>
static inline Range push_storage(const T& data) { static inline Range push_storage(const T& data) {
return push_storage(reinterpret_cast<const uint8_t*>(&data), sizeof(T)); return push_storage(reinterpret_cast<const uint8_t*>(&data), sizeof(T));
} }
Range push_static_storage(const uint8_t* data, size_t length);
template <typename T>
static inline Range push_static_storage(ArrayRef<T> data) {
return push_static_storage(reinterpret_cast<const uint8_t*>(data.data()), data.size() * sizeof(T));
}
template <typename T>
static inline Range push_static_storage(const T& data) {
return push_static_storage(reinterpret_cast<const uint8_t*>(&data), sizeof(T));
}
std::pair<ByteBuffer, Range> map_verts(size_t length); std::pair<ByteBuffer, Range> map_verts(size_t length);
std::pair<ByteBuffer, Range> map_indices(size_t length); std::pair<ByteBuffer, Range> map_indices(size_t length);
std::pair<ByteBuffer, Range> map_uniform(size_t length); std::pair<ByteBuffer, Range> map_uniform(size_t length);

View File

@ -3,7 +3,7 @@
#include "../gpu.hpp" #include "../gpu.hpp"
#include "common.hpp" #include "common.hpp"
#include <unordered_map> #include <absl/container/flat_hash_map.h>
using aurora::gfx::gx::g_gxState; using aurora::gfx::gx::g_gxState;
static logvisor::Module Log("aurora::gx"); static logvisor::Module Log("aurora::gx");
@ -537,8 +537,8 @@ Range build_uniform(const ShaderInfo& info) noexcept {
return range; return range;
} }
static std::unordered_map<u32, wgpu::BindGroupLayout> sUniformBindGroupLayouts; static absl::flat_hash_map<u32, wgpu::BindGroupLayout> sUniformBindGroupLayouts;
static std::unordered_map<u32, std::pair<wgpu::BindGroupLayout, wgpu::BindGroupLayout>> sTextureBindGroupLayouts; static absl::flat_hash_map<u32, std::pair<wgpu::BindGroupLayout, wgpu::BindGroupLayout>> sTextureBindGroupLayouts;
GXBindGroups build_bind_groups(const ShaderInfo& info, const ShaderConfig& config, GXBindGroups build_bind_groups(const ShaderInfo& info, const ShaderConfig& config,
const BindGroupRanges& ranges) noexcept { const BindGroupRanges& ranges) noexcept {
@ -555,25 +555,25 @@ GXBindGroups build_bind_groups(const ShaderInfo& info, const ShaderConfig& confi
wgpu::BindGroupEntry{ wgpu::BindGroupEntry{
.binding = 1, .binding = 1,
.buffer = g_storageBuffer, .buffer = g_storageBuffer,
.size = ranges.vtxDataRange.second - ranges.vtxDataRange.first, .size = ranges.vtxDataRange.size,
}, },
// Normals // Normals
wgpu::BindGroupEntry{ wgpu::BindGroupEntry{
.binding = 2, .binding = 2,
.buffer = g_storageBuffer, .buffer = g_storageBuffer,
.size = ranges.nrmDataRange.second - ranges.nrmDataRange.first, .size = ranges.nrmDataRange.size,
}, },
// UVs // UVs
wgpu::BindGroupEntry{ wgpu::BindGroupEntry{
.binding = 3, .binding = 3,
.buffer = g_storageBuffer, .buffer = g_storageBuffer,
.size = ranges.tcDataRange.second - ranges.tcDataRange.first, .size = ranges.tcDataRange.size,
}, },
// Packed UVs // Packed UVs
wgpu::BindGroupEntry{ wgpu::BindGroupEntry{
.binding = 4, .binding = 4,
.buffer = g_storageBuffer, .buffer = g_storageBuffer,
.size = ranges.packedTcDataRange.second - ranges.packedTcDataRange.first, .size = ranges.packedTcDataRange.size,
}, },
}; };
std::array<wgpu::BindGroupEntry, MaxTextures> samplerEntries; std::array<wgpu::BindGroupEntry, MaxTextures> samplerEntries;
@ -622,8 +622,9 @@ GXBindGroups build_bind_groups(const ShaderInfo& info, const ShaderConfig& confi
GXBindGroupLayouts build_bind_group_layouts(const ShaderInfo& info, const ShaderConfig& config) noexcept { GXBindGroupLayouts build_bind_group_layouts(const ShaderInfo& info, const ShaderConfig& config) noexcept {
GXBindGroupLayouts out; GXBindGroupLayouts out;
u32 uniformSizeKey = info.uniformSize + (config.denormalizedVertexAttributes ? 0 : 1); u32 uniformSizeKey = info.uniformSize + (config.denormalizedVertexAttributes ? 0 : 1);
if (sUniformBindGroupLayouts.contains(uniformSizeKey)) { const auto uniformIt = sUniformBindGroupLayouts.find(uniformSizeKey);
out.uniformLayout = sUniformBindGroupLayouts[uniformSizeKey]; if (uniformIt != sUniformBindGroupLayouts.end()) {
out.uniformLayout = uniformIt->second;
} else { } else {
const std::array uniformLayoutEntries{ const std::array uniformLayoutEntries{
wgpu::BindGroupLayoutEntry{ wgpu::BindGroupLayoutEntry{
@ -683,8 +684,9 @@ GXBindGroupLayouts build_bind_group_layouts(const ShaderInfo& info, const Shader
} }
u32 textureCount = info.sampledTextures.count(); u32 textureCount = info.sampledTextures.count();
if (sTextureBindGroupLayouts.contains(textureCount)) { const auto textureIt = sTextureBindGroupLayouts.find(textureCount);
const auto& [sl, tl] = sTextureBindGroupLayouts[textureCount]; if (textureIt != sTextureBindGroupLayouts.end()) {
const auto& [sl, tl] = textureIt->second;
out.samplerLayout = sl; out.samplerLayout = sl;
out.textureLayout = tl; out.textureLayout = tl;
} else { } else {
@ -728,7 +730,7 @@ GXBindGroupLayouts build_bind_group_layouts(const ShaderInfo& info, const Shader
} }
// TODO this is awkward // TODO this is awkward
extern std::unordered_map<ShaderRef, std::pair<wgpu::ShaderModule, gx::ShaderInfo>> g_gxCachedShaders; extern absl::flat_hash_map<ShaderRef, std::pair<wgpu::ShaderModule, gx::ShaderInfo>> g_gxCachedShaders;
void shutdown() noexcept { void shutdown() noexcept {
// TODO we should probably store this all in g_state.gx instead // TODO we should probably store this all in g_state.gx instead
sUniformBindGroupLayouts.clear(); sUniformBindGroupLayouts.clear();

View File

@ -20,6 +20,7 @@ struct TevPass {
Arg b = Default; Arg b = Default;
Arg c = Default; Arg c = Default;
Arg d = Default; Arg d = Default;
bool operator==(const TevPass&) const = default;
}; };
struct TevOp { struct TevOp {
GX::TevOp op = GX::TevOp::TEV_ADD; GX::TevOp op = GX::TevOp::TEV_ADD;
@ -27,6 +28,7 @@ struct TevOp {
GX::TevScale scale = GX::TevScale::CS_SCALE_1; GX::TevScale scale = GX::TevScale::CS_SCALE_1;
GX::TevRegID outReg = GX::TevRegID::TEVPREV; GX::TevRegID outReg = GX::TevRegID::TEVPREV;
bool clamp = true; bool clamp = true;
bool operator==(const TevOp&) const = default;
}; };
struct TevStage { struct TevStage {
TevPass<GX::TevColorArg, GX::CC_ZERO> colorPass; TevPass<GX::TevColorArg, GX::CC_ZERO> colorPass;
@ -38,6 +40,7 @@ struct TevStage {
GX::TexCoordID texCoordId = GX::TEXCOORD_NULL; GX::TexCoordID texCoordId = GX::TEXCOORD_NULL;
GX::TexMapID texMapId = GX::TEXMAP_NULL; GX::TexMapID texMapId = GX::TEXMAP_NULL;
GX::ChannelID channelId = GX::COLOR_NULL; GX::ChannelID channelId = GX::COLOR_NULL;
bool operator==(const TevStage&) const = default;
}; };
struct TextureBind { struct TextureBind {
aurora::gfx::TextureHandle handle; aurora::gfx::TextureHandle handle;
@ -56,6 +59,7 @@ struct ColorChannelConfig {
GX::ColorSrc matSrc = GX::SRC_REG; GX::ColorSrc matSrc = GX::SRC_REG;
GX::ColorSrc ambSrc = GX::SRC_REG; GX::ColorSrc ambSrc = GX::SRC_REG;
bool lightingEnabled = false; bool lightingEnabled = false;
bool operator==(const ColorChannelConfig&) const = default;
}; };
// For uniform generation // For uniform generation
struct ColorChannelState { struct ColorChannelState {
@ -72,6 +76,7 @@ struct TcgConfig {
GX::TexMtx mtx = GX::IDENTITY; GX::TexMtx mtx = GX::IDENTITY;
GX::PTTexMtx postMtx = GX::PTIDENTITY; GX::PTTexMtx postMtx = GX::PTIDENTITY;
bool normalize = false; bool normalize = false;
bool operator==(const TcgConfig&) const = default;
}; };
struct FogState { struct FogState {
GX::FogType type = GX::FOG_NONE; GX::FogType type = GX::FOG_NONE;
@ -129,6 +134,7 @@ struct ShaderConfig {
std::optional<float> alphaDiscard; std::optional<float> alphaDiscard;
bool denormalizedVertexAttributes = false; bool denormalizedVertexAttributes = false;
bool denormalizedHasNrm = false; // TODO this is a hack bool denormalizedHasNrm = false; // TODO this is a hack
bool operator==(const ShaderConfig&) const = default;
}; };
struct PipelineConfig { struct PipelineConfig {
ShaderConfig shaderConfig; ShaderConfig shaderConfig;

View File

@ -3,14 +3,17 @@
#include "../gpu.hpp" #include "../gpu.hpp"
#include "gx.hpp" #include "gx.hpp"
#include <unordered_map> #include <absl/container/flat_hash_map.h>
namespace aurora::gfx::gx { namespace aurora::gfx::gx {
using namespace fmt::literals; using namespace fmt::literals;
static logvisor::Module Log("aurora::gfx::gx"); static logvisor::Module Log("aurora::gfx::gx");
std::unordered_map<ShaderRef, std::pair<wgpu::ShaderModule, gx::ShaderInfo>> g_gxCachedShaders; absl::flat_hash_map<ShaderRef, std::pair<wgpu::ShaderModule, gx::ShaderInfo>> g_gxCachedShaders;
#ifndef NDEBUG
static absl::flat_hash_map<ShaderRef, gx::ShaderConfig> g_gxCachedShaderConfigs;
#endif
static std::string color_arg_reg(GX::TevColorArg arg, size_t stageIdx, const TevStage& stage, ShaderInfo& info) { static std::string color_arg_reg(GX::TevColorArg arg, size_t stageIdx, const TevStage& stage, ShaderInfo& info) {
switch (arg) { switch (arg) {
@ -346,8 +349,15 @@ static std::string in_uv(u32 idx) {
std::pair<wgpu::ShaderModule, ShaderInfo> build_shader(const ShaderConfig& config) noexcept { std::pair<wgpu::ShaderModule, ShaderInfo> build_shader(const ShaderConfig& config) noexcept {
const auto hash = xxh3_hash(config); const auto hash = xxh3_hash(config);
if (g_gxCachedShaders.contains(hash)) { const auto it = g_gxCachedShaders.find(hash);
return g_gxCachedShaders[hash]; if (it != g_gxCachedShaders.end()) {
#ifndef NDEBUG
if (g_gxCachedShaderConfigs[hash] != config) {
Log.report(logvisor::Fatal, FMT_STRING("Shader collision!"));
unreachable();
}
#endif
return it->second;
} }
Log.report(logvisor::Info, FMT_STRING("Shader config (hash {:x}):"), hash); Log.report(logvisor::Info, FMT_STRING("Shader config (hash {:x}):"), hash);
@ -791,6 +801,9 @@ fn fs_main(in: VertexOutput) -> @location(0) vec4<f32> {{
info.uniformSize = align_uniform(info.uniformSize); info.uniformSize = align_uniform(info.uniformSize);
auto pair = std::make_pair(std::move(shader), info); auto pair = std::make_pair(std::move(shader), info);
g_gxCachedShaders.emplace(hash, pair); g_gxCachedShaders.emplace(hash, pair);
#ifndef NDEBUG
g_gxCachedShaderConfigs.emplace(hash, config);
#endif
return pair; return pair;
} }

View File

@ -3,6 +3,7 @@
#include "../../gpu.hpp" #include "../../gpu.hpp"
#include "../common.hpp" #include "../common.hpp"
#include <absl/container/flat_hash_map.h>
#include <aurora/model.hpp> #include <aurora/model.hpp>
#include <magic_enum.hpp> #include <magic_enum.hpp>
@ -53,11 +54,10 @@ static const std::vector<zeus::CVector3f>* vtxData;
static const std::vector<zeus::CVector3f>* nrmData; static const std::vector<zeus::CVector3f>* nrmData;
static const std::vector<Vec2<float>>* tex0TcData; static const std::vector<Vec2<float>>* tex0TcData;
static const std::vector<Vec2<float>>* tcData; static const std::vector<Vec2<float>>* tcData;
static std::optional<Range> staticVtxRange;
// void set_vertex_buffer(const std::vector<zeus::CVector3f>* data) noexcept { vtxData = data; } static std::optional<Range> staticNrmRange;
// void set_normal_buffer(const std::vector<zeus::CVector3f>* norm) noexcept { nrmData = norm; } static std::optional<Range> staticPackedTcRange;
// void set_tex0_tc_buffer(const std::vector<Vec2<float>>* tcs) noexcept { tex0TcData = tcs; } static std::optional<Range> staticTcRange;
// void set_tc_buffer(const std::vector<Vec2<float>>* tcs) noexcept { tcData = tcs; }
enum class VertexFormat : u8 { enum class VertexFormat : u8 {
F32F32, F32F32,
@ -110,8 +110,19 @@ static inline std::pair<gx::DlVert, size_t> readVert(const u8* data) noexcept {
return {out, offset}; return {out, offset};
} }
static absl::flat_hash_map<XXH64_hash_t, std::pair<std::vector<gx::DlVert>, std::vector<u32>>> sCachedDisplayLists;
void queue_surface(const u8* dlStart, u32 dlSize) noexcept { void queue_surface(const u8* dlStart, u32 dlSize) noexcept {
// Log.report(logvisor::Info, FMT_STRING("DL size {}"), dlSize); const auto hash = xxh3_hash(dlStart, dlSize, 0);
Range vertRange, idxRange;
uint32_t numIndices;
auto it = sCachedDisplayLists.find(hash);
if (it != sCachedDisplayLists.end()) {
const auto& [verts, indices] = it->second;
numIndices = indices.size();
vertRange = push_verts(ArrayRef{verts});
idxRange = push_indices(ArrayRef{indices});
} else {
std::vector<gx::DlVert> verts; std::vector<gx::DlVert> verts;
std::vector<u32> indices; std::vector<u32> indices;
@ -121,8 +132,6 @@ void queue_surface(const u8* dlStart, u32 dlSize) noexcept {
const auto primitive = static_cast<GX::Primitive>(header & 0xF8); const auto primitive = static_cast<GX::Primitive>(header & 0xF8);
const auto vtxFmt = static_cast<VertexFormat>(header & 0x3); const auto vtxFmt = static_cast<VertexFormat>(header & 0x3);
const auto vtxCount = metaforce::SBig(*reinterpret_cast<const u16*>(dlStart + offset + 1)); const auto vtxCount = metaforce::SBig(*reinterpret_cast<const u16*>(dlStart + offset + 1));
// Log.report(logvisor::Info, FMT_STRING("DL header prim {}, fmt {}, vtx count {}"), primitive,
// magic_enum::enum_name(vtxFmt), vtxCount);
offset += 3; offset += 3;
if (primitive == 0) { if (primitive == 0) {
@ -165,14 +174,31 @@ void queue_surface(const u8* dlStart, u32 dlSize) noexcept {
} }
} }
// Log.report(logvisor::Info, FMT_STRING("Read {} verts, {} indices"), verts.size(), indices.size()); numIndices = indices.size();
const auto vertRange = push_verts(ArrayRef{verts}); vertRange = push_verts(ArrayRef{verts});
const auto idxRange = push_indices(ArrayRef{indices}); idxRange = push_indices(ArrayRef{indices});
const auto sVtxRange = push_storage(reinterpret_cast<const uint8_t*>(vtxData->data()), vtxData->size() * 16); sCachedDisplayLists.try_emplace(hash, std::move(verts), std::move(indices));
const auto sNrmRange = push_storage(reinterpret_cast<const uint8_t*>(nrmData->data()), nrmData->size() * 16); }
const auto sTcRange = push_storage(reinterpret_cast<const uint8_t*>(tcData->data()), tcData->size() * 8);
Range sPackedTcRange; Range sVtxRange, sNrmRange, sTcRange, sPackedTcRange;
if (tcData == tex0TcData) { if (staticVtxRange) {
sVtxRange = *staticVtxRange;
} else {
sVtxRange = push_storage(reinterpret_cast<const uint8_t*>(vtxData->data()), vtxData->size() * 16);
}
if (staticNrmRange) {
sNrmRange = *staticNrmRange;
} else {
sNrmRange = push_storage(reinterpret_cast<const uint8_t*>(nrmData->data()), nrmData->size() * 16);
}
if (staticTcRange) {
sTcRange = *staticTcRange;
} else {
sTcRange = push_storage(reinterpret_cast<const uint8_t*>(tcData->data()), tcData->size() * 8);
}
if (staticPackedTcRange) {
sPackedTcRange = *staticPackedTcRange;
} else if (tcData == tex0TcData) {
sPackedTcRange = sTcRange; sPackedTcRange = sTcRange;
} else { } else {
sPackedTcRange = push_storage(reinterpret_cast<const uint8_t*>(tex0TcData->data()), tex0TcData->size() * 8); sPackedTcRange = push_storage(reinterpret_cast<const uint8_t*>(tex0TcData->data()), tex0TcData->size() * 8);
@ -192,12 +218,9 @@ void queue_surface(const u8* dlStart, u32 dlSize) noexcept {
.pipeline = pipeline, .pipeline = pipeline,
.vertRange = vertRange, .vertRange = vertRange,
.idxRange = idxRange, .idxRange = idxRange,
.sVtxRange = sVtxRange, .dataRanges = ranges,
.sNrmRange = sNrmRange,
.sTcRange = sTcRange,
.sPackedTcRange = sPackedTcRange,
.uniformRange = build_uniform(info), .uniformRange = build_uniform(info),
.indexCount = static_cast<uint32_t>(indices.size()), .indexCount = numIndices,
.bindGroups = info.bindGroups, .bindGroups = info.bindGroups,
}); });
} }
@ -220,36 +243,60 @@ void render(const State& state, const DrawData& data, const wgpu::RenderPassEnco
} }
const std::array offsets{ const std::array offsets{
data.uniformRange.first, data.sVtxRange.first, data.sNrmRange.first, data.uniformRange.offset,
data.sTcRange.first, data.sPackedTcRange.first, storage_offset(data.dataRanges.vtxDataRange),
storage_offset(data.dataRanges.nrmDataRange),
storage_offset(data.dataRanges.tcDataRange),
storage_offset(data.dataRanges.packedTcDataRange),
}; };
pass.SetBindGroup(0, find_bind_group(data.bindGroups.uniformBindGroup), offsets.size(), offsets.data()); pass.SetBindGroup(0, find_bind_group(data.bindGroups.uniformBindGroup), offsets.size(), offsets.data());
if (data.bindGroups.samplerBindGroup && data.bindGroups.textureBindGroup) { if (data.bindGroups.samplerBindGroup && data.bindGroups.textureBindGroup) {
pass.SetBindGroup(1, find_bind_group(data.bindGroups.samplerBindGroup)); pass.SetBindGroup(1, find_bind_group(data.bindGroups.samplerBindGroup));
pass.SetBindGroup(2, find_bind_group(data.bindGroups.textureBindGroup)); pass.SetBindGroup(2, find_bind_group(data.bindGroups.textureBindGroup));
} }
pass.SetVertexBuffer(0, g_vertexBuffer, data.vertRange.first, data.vertRange.second); pass.SetVertexBuffer(0, g_vertexBuffer, data.vertRange.offset, data.vertRange.size);
pass.SetIndexBuffer(g_indexBuffer, wgpu::IndexFormat::Uint32, data.idxRange.first, data.idxRange.second); pass.SetIndexBuffer(g_indexBuffer, wgpu::IndexFormat::Uint32, data.idxRange.offset, data.idxRange.size);
pass.DrawIndexed(data.indexCount); pass.DrawIndexed(data.indexCount);
} }
} // namespace aurora::gfx::model } // namespace aurora::gfx::model
static absl::flat_hash_map<XXH64_hash_t, aurora::gfx::Range> sCachedRanges;
template <typename Vec>
static inline void cache_array(const void* data, Vec*& outPtr, std::optional<aurora::gfx::Range>& outRange, u8 stride) {
Vec* vecPtr = static_cast<Vec*>(data);
outPtr = vecPtr;
if (stride == 1) {
const auto hash = aurora::xxh3_hash(vecPtr->data(), vecPtr->size() * sizeof(typename Vec::value_type), 0);
const auto it = sCachedRanges.find(hash);
if (it != sCachedRanges.end()) {
outRange = it->second;
} else {
const auto range = aurora::gfx::push_static_storage(aurora::ArrayRef{*vecPtr});
sCachedRanges.try_emplace(hash, range);
outRange = range;
}
} else {
outRange.reset();
}
}
void GXSetArray(GX::Attr attr, const void* data, u8 stride) noexcept { void GXSetArray(GX::Attr attr, const void* data, u8 stride) noexcept {
using namespace aurora::gfx::model;
switch (attr) { switch (attr) {
case GX::VA_POS: case GX::VA_POS:
aurora::gfx::model::vtxData = static_cast<const std::vector<zeus::CVector3f>*>(data); cache_array(data, vtxData, staticVtxRange, stride);
break; break;
case GX::VA_NRM: case GX::VA_NRM:
aurora::gfx::model::nrmData = static_cast<const std::vector<zeus::CVector3f>*>(data); cache_array(data, nrmData, staticNrmRange, stride);
break; break;
case GX::VA_TEX0: case GX::VA_TEX0:
aurora::gfx::model::tex0TcData = static_cast<const std::vector<aurora::Vec2<float>>*>(data); cache_array(data, tex0TcData, staticPackedTcRange, stride);
break; break;
case GX::VA_TEX1: case GX::VA_TEX1:
aurora::gfx::model::tcData = static_cast<const std::vector<aurora::Vec2<float>>*>(data); cache_array(data, tcData, staticTcRange, stride);
break; break;
default: default:
aurora::gfx::model::Log.report(logvisor::Fatal, FMT_STRING("GXSetArray: invalid attr {}"), attr); Log.report(logvisor::Fatal, FMT_STRING("GXSetArray: invalid attr {}"), attr);
unreachable(); unreachable();
} }
} }

View File

@ -3,17 +3,12 @@
#include "../common.hpp" #include "../common.hpp"
#include "../gx.hpp" #include "../gx.hpp"
#include <unordered_map>
namespace aurora::gfx::model { namespace aurora::gfx::model {
struct DrawData { struct DrawData {
PipelineRef pipeline; PipelineRef pipeline;
Range vertRange; Range vertRange;
Range idxRange; Range idxRange;
Range sVtxRange; gx::BindGroupRanges dataRanges;
Range sNrmRange;
Range sTcRange;
Range sPackedTcRange;
Range uniformRange; Range uniformRange;
uint32_t indexCount; uint32_t indexCount;
gx::GXBindGroups bindGroups; gx::GXBindGroups bindGroups;

View File

@ -239,7 +239,7 @@ void render(const State& state, const DrawData& data, const wgpu::RenderPassEnco
pass.SetBindGroup(0, state.uniformBindGroup); pass.SetBindGroup(0, state.uniformBindGroup);
pass.SetBindGroup(1, find_bind_group(data.textureBindGroup)); pass.SetBindGroup(1, find_bind_group(data.textureBindGroup));
pass.SetVertexBuffer(0, g_vertexBuffer, data.vertRange.first, data.vertRange.second); pass.SetVertexBuffer(0, g_vertexBuffer, data.vertRange.offset, data.vertRange.size);
pass.Draw(4); pass.Draw(4);
} }
} // namespace aurora::gfx::movie_player } // namespace aurora::gfx::movie_player

View File

@ -4,8 +4,6 @@
#include "common.hpp" #include "common.hpp"
#include "gx.hpp" #include "gx.hpp"
#include <unordered_map>
namespace aurora::gfx { namespace aurora::gfx {
static logvisor::Module Log("aurora::gfx::stream"); static logvisor::Module Log("aurora::gfx::stream");

View File

@ -6,10 +6,6 @@
#include <magic_enum.hpp> #include <magic_enum.hpp>
#include <utility> #include <utility>
namespace aurora::gfx {
extern std::unordered_map<ShaderRef, wgpu::ShaderModule> g_gxCachedShaders;
} // namespace aurora::gfx
namespace aurora::gfx::stream { namespace aurora::gfx::stream {
static logvisor::Module Log("aurora::gfx::stream"); static logvisor::Module Log("aurora::gfx::stream");
@ -66,60 +62,20 @@ wgpu::RenderPipeline create_pipeline(const State& state, [[maybe_unused]] Pipeli
return build_pipeline(config, info, vertexBuffers, shader, "Stream Pipeline"); return build_pipeline(config, info, vertexBuffers, shader, "Stream Pipeline");
} }
State construct_state() { State construct_state() { return {}; }
const auto samplerBinding = wgpu::SamplerBindingLayout{
.type = wgpu::SamplerBindingType::Filtering,
};
const std::array samplerLayoutEntries{
wgpu::BindGroupLayoutEntry{
.binding = 0,
.visibility = wgpu::ShaderStage::Fragment,
.sampler = samplerBinding,
},
};
const auto samplerLayoutDescriptor = wgpu::BindGroupLayoutDescriptor{
.label = "Stream Sampler Bind Group Layout",
.entryCount = samplerLayoutEntries.size(),
.entries = samplerLayoutEntries.data(),
};
auto samplerLayout = g_device.CreateBindGroupLayout(&samplerLayoutDescriptor);
const auto textureBinding = wgpu::TextureBindingLayout{
.sampleType = wgpu::TextureSampleType::Float,
.viewDimension = wgpu::TextureViewDimension::e2D,
};
const std::array textureLayoutEntries{
wgpu::BindGroupLayoutEntry{
.binding = 0,
.visibility = wgpu::ShaderStage::Fragment,
.texture = textureBinding,
},
};
const auto textureLayoutDescriptor = wgpu::BindGroupLayoutDescriptor{
.label = "Stream Texture Bind Group Layout",
.entryCount = textureLayoutEntries.size(),
.entries = textureLayoutEntries.data(),
};
auto textureLayout = g_device.CreateBindGroupLayout(&textureLayoutDescriptor);
return {
.samplerLayout = samplerLayout,
.textureLayout = textureLayout,
};
}
void render(const State& state, const DrawData& data, const wgpu::RenderPassEncoder& pass) { void render(const State& state, const DrawData& data, const wgpu::RenderPassEncoder& pass) {
if (!bind_pipeline(data.pipeline, pass)) { if (!bind_pipeline(data.pipeline, pass)) {
return; return;
} }
const std::array offsets{data.uniformRange.first}; const std::array offsets{data.uniformRange.offset};
pass.SetBindGroup(0, find_bind_group(data.bindGroups.uniformBindGroup), offsets.size(), offsets.data()); pass.SetBindGroup(0, find_bind_group(data.bindGroups.uniformBindGroup), offsets.size(), offsets.data());
if (data.bindGroups.samplerBindGroup && data.bindGroups.textureBindGroup) { if (data.bindGroups.samplerBindGroup && data.bindGroups.textureBindGroup) {
pass.SetBindGroup(1, find_bind_group(data.bindGroups.samplerBindGroup)); pass.SetBindGroup(1, find_bind_group(data.bindGroups.samplerBindGroup));
pass.SetBindGroup(2, find_bind_group(data.bindGroups.textureBindGroup)); pass.SetBindGroup(2, find_bind_group(data.bindGroups.textureBindGroup));
} }
pass.SetVertexBuffer(0, g_vertexBuffer, data.vertRange.first, data.vertRange.second); pass.SetVertexBuffer(0, g_vertexBuffer, data.vertRange.offset, data.vertRange.size);
pass.Draw(data.vertexCount); pass.Draw(data.vertexCount);
} }
} // namespace aurora::gfx::stream } // namespace aurora::gfx::stream

View File

@ -3,8 +3,6 @@
#include "../common.hpp" #include "../common.hpp"
#include "../gx.hpp" #include "../gx.hpp"
#include <unordered_map>
namespace aurora::gfx::stream { namespace aurora::gfx::stream {
struct DrawData { struct DrawData {
PipelineRef pipeline; PipelineRef pipeline;
@ -16,19 +14,7 @@ struct DrawData {
struct PipelineConfig : public gx::PipelineConfig {}; struct PipelineConfig : public gx::PipelineConfig {};
struct CachedBindGroup { struct State {};
wgpu::BindGroupLayout layout;
wgpu::BindGroup bindGroup;
CachedBindGroup(wgpu::BindGroupLayout layout, wgpu::BindGroup&& group)
: layout(std::move(layout)), bindGroup(std::move(group)) {}
};
struct State {
wgpu::BindGroupLayout samplerLayout;
wgpu::BindGroupLayout textureLayout;
mutable std::unordered_map<uint32_t, CachedBindGroup> uniform;
mutable std::unordered_map<uint64_t, wgpu::Sampler> sampler;
mutable std::unordered_map<PipelineRef, gx::ShaderInfo> shaderInfo;
};
State construct_state(); State construct_state();
wgpu::RenderPipeline create_pipeline(const State& state, [[maybe_unused]] PipelineConfig config); wgpu::RenderPipeline create_pipeline(const State& state, [[maybe_unused]] PipelineConfig config);

View File

@ -380,10 +380,10 @@ void render(const State& state, const DrawData& data, const wgpu::RenderPassEnco
return; return;
} }
const std::array offsets{data.uniformRange.first}; const std::array offsets{data.uniformRange.offset};
pass.SetBindGroup(0, state.uniformBindGroup, offsets.size(), offsets.data()); pass.SetBindGroup(0, state.uniformBindGroup, offsets.size(), offsets.data());
pass.SetBindGroup(1, find_bind_group(data.textureBindGroup)); pass.SetBindGroup(1, find_bind_group(data.textureBindGroup));
pass.SetVertexBuffer(0, g_vertexBuffer, data.vertRange.first, data.vertRange.second); pass.SetVertexBuffer(0, g_vertexBuffer, data.vertRange.offset, data.vertRange.size);
pass.Draw(4); pass.Draw(4);
} }
} // namespace aurora::gfx::textured_quad } // namespace aurora::gfx::textured_quad

View File

@ -2,6 +2,7 @@
#include <SDL_haptic.h> #include <SDL_haptic.h>
#include <absl/container/btree_map.h> #include <absl/container/btree_map.h>
#include <absl/container/flat_hash_map.h>
#include <absl/strings/str_split.h> #include <absl/strings/str_split.h>
namespace aurora::input { namespace aurora::input {
@ -13,7 +14,7 @@ struct GameController {
Sint32 m_index = -1; Sint32 m_index = -1;
bool m_hasRumble = false; bool m_hasRumble = false;
}; };
std::unordered_map<Uint32, GameController> g_GameControllers; absl::flat_hash_map<Uint32, GameController> g_GameControllers;
static std::optional<std::string> remap_controller_layout(std::string_view mapping) { static std::optional<std::string> remap_controller_layout(std::string_view mapping) {
std::string newMapping; std::string newMapping;

View File

@ -1,11 +1,11 @@
#pragma once #pragma once
#include <unordered_map>
#include "aurora/aurora.hpp" #include "aurora/aurora.hpp"
#include "SDL_gamecontroller.h" #include "SDL_gamecontroller.h"
#include "SDL_keyboard.h" #include "SDL_keyboard.h"
#include "SDL_keycode.h" #include "SDL_keycode.h"
#include "SDL_mouse.h" #include "SDL_mouse.h"
namespace aurora::input { namespace aurora::input {
Sint32 add_controller(Sint32 which) noexcept; Sint32 add_controller(Sint32 which) noexcept;
void remove_controller(Uint32 instance) noexcept; void remove_controller(Uint32 instance) noexcept;