#include "shader.hpp" #include "../../gpu.hpp" #include "../common.hpp" #include #include namespace aurora::gfx::model { static logvisor::Module Log("aurora::gfx::model"); static const std::vector* vtxData; static const std::vector* nrmData; static const std::vector>* tex0TcData; static const std::vector>* tcData; static std::optional staticVtxRange; static std::optional staticNrmRange; static std::optional staticPackedTcRange; static std::optional staticTcRange; static inline std::pair readVert(const u8* data) noexcept { gx::DlVert out{}; size_t offset = 0; const auto vtxTypes = gx::g_gxState.vtxDesc; const auto read8 = [/*data, &offset*/](GX::AttrType type) -> s8 { // if (type == GX::INDEX8) { // s8 v = static_cast(data[offset]); // ++offset; // return v; // } #ifndef NDEBUG if (type != GX::NONE) { Log.report(logvisor::Fatal, FMT_STRING("unsupported vtx attr")); unreachable(); } #endif return 0; }; const auto read16 = [data, &offset](GX::AttrType type) -> s16 { if (type == GX::INDEX16) { s16 v = metaforce::SBig(*reinterpret_cast(data + offset)); offset += 2; return v; } return 0; }; read8(vtxTypes[GX::VA_PNMTXIDX]); read8(vtxTypes[GX::VA_TEX0MTXIDX]); read8(vtxTypes[GX::VA_TEX1MTXIDX]); read8(vtxTypes[GX::VA_TEX2MTXIDX]); read8(vtxTypes[GX::VA_TEX3MTXIDX]); read8(vtxTypes[GX::VA_TEX4MTXIDX]); read8(vtxTypes[GX::VA_TEX5MTXIDX]); read8(vtxTypes[GX::VA_TEX6MTXIDX]); out.pos = read16(vtxTypes[GX::VA_POS]); out.norm = read16(vtxTypes[GX::VA_NRM]); read16(vtxTypes[GX::VA_CLR0]); read16(vtxTypes[GX::VA_CLR1]); out.uvs[0] = read16(vtxTypes[GX::VA_TEX0]); out.uvs[1] = read16(vtxTypes[GX::VA_TEX1]); out.uvs[2] = read16(vtxTypes[GX::VA_TEX2]); out.uvs[3] = read16(vtxTypes[GX::VA_TEX3]); out.uvs[4] = read16(vtxTypes[GX::VA_TEX4]); out.uvs[5] = read16(vtxTypes[GX::VA_TEX5]); out.uvs[6] = read16(vtxTypes[GX::VA_TEX6]); return {out, offset}; } static absl::flat_hash_map, std::vector>> sCachedDisplayLists; void queue_surface(const u8* dlStart, u32 dlSize) noexcept { const auto hash = xxh3_hash(dlStart, dlSize, 0); Range vertRange, idxRange; uint32_t numIndices; auto it = sCachedDisplayLists.find(hash); if (it != sCachedDisplayLists.end()) { const auto& [verts, indices] = it->second; numIndices = indices.size(); vertRange = push_verts(ArrayRef{verts}); idxRange = push_indices(ArrayRef{indices}); } else { std::vector verts; std::vector indices; size_t offset = 0; while (offset < dlSize - 6) { const auto header = dlStart[offset]; const auto primitive = static_cast(header & 0xF8); const auto vtxCount = metaforce::SBig(*reinterpret_cast(dlStart + offset + 1)); offset += 3; if (primitive == 0) { break; } if (primitive != GX::TRIANGLES && primitive != GX::TRIANGLESTRIP && primitive != GX::TRIANGLEFAN) { Log.report(logvisor::Fatal, FMT_STRING("queue_surface: unsupported primitive type {}"), primitive); unreachable(); } const u32 idxStart = indices.size(); const u16 vertsStart = verts.size(); verts.reserve(vertsStart + vtxCount); if (vtxCount > 3 && (primitive == GX::TRIANGLEFAN || primitive == GX::TRIANGLESTRIP)) { indices.reserve(idxStart + (u32(vtxCount) - 3) * 3 + 3); } else { indices.reserve(idxStart + vtxCount); } auto curVert = vertsStart; for (int v = 0; v < vtxCount; ++v) { const auto [vert, read] = readVert(dlStart + offset); verts.push_back(vert); offset += read; if (primitive == GX::TRIANGLES || v < 3) { // pass } else if (primitive == GX::TRIANGLEFAN) { indices.push_back(vertsStart); indices.push_back(curVert - 1); } else if (primitive == GX::TRIANGLESTRIP) { if ((v & 1) == 0) { indices.push_back(curVert - 2); indices.push_back(curVert - 1); } else { indices.push_back(curVert - 1); indices.push_back(curVert - 2); } } indices.push_back(curVert); ++curVert; } } numIndices = indices.size(); vertRange = push_verts(ArrayRef{verts}); idxRange = push_indices(ArrayRef{indices}); sCachedDisplayLists.try_emplace(hash, std::move(verts), std::move(indices)); } Range sVtxRange, sNrmRange, sTcRange, sPackedTcRange; if (staticVtxRange) { sVtxRange = *staticVtxRange; } else { sVtxRange = push_storage(reinterpret_cast(vtxData->data()), vtxData->size() * 16); } if (staticNrmRange) { sNrmRange = *staticNrmRange; } else { sNrmRange = push_storage(reinterpret_cast(nrmData->data()), nrmData->size() * 16); } if (staticTcRange) { sTcRange = *staticTcRange; } else { sTcRange = push_storage(reinterpret_cast(tcData->data()), tcData->size() * 8); } if (staticPackedTcRange) { sPackedTcRange = *staticPackedTcRange; } else if (tcData == tex0TcData) { sPackedTcRange = sTcRange; } else { sPackedTcRange = push_storage(reinterpret_cast(tex0TcData->data()), tex0TcData->size() * 8); } model::PipelineConfig config{}; const gx::BindGroupRanges ranges{ .vtxDataRange = sVtxRange, .nrmDataRange = sNrmRange, .tcDataRange = sTcRange, .packedTcDataRange = sPackedTcRange, }; const auto info = populate_pipeline_config(config, GX::TRIANGLES, ranges); const auto pipeline = pipeline_ref(config); push_draw_command(model::DrawData{ .pipeline = pipeline, .vertRange = vertRange, .idxRange = idxRange, .dataRanges = ranges, .uniformRange = build_uniform(info), .indexCount = numIndices, .bindGroups = info.bindGroups, }); } State construct_state() { return {}; } wgpu::RenderPipeline create_pipeline(const State& state, [[maybe_unused]] PipelineConfig config) { const auto [shader, info] = build_shader(config.shaderConfig); const auto attributes = gpu::utils::make_vertex_attributes( std::array{wgpu::VertexFormat::Sint16x2, wgpu::VertexFormat::Sint16x4, wgpu::VertexFormat::Sint16x4}); const std::array vertexBuffers{gpu::utils::make_vertex_buffer_layout(sizeof(gx::DlVert), attributes)}; return build_pipeline(config, info, vertexBuffers, shader, "Model Pipeline"); } void render(const State& state, const DrawData& data, const wgpu::RenderPassEncoder& pass) { if (!bind_pipeline(data.pipeline, pass)) { return; } const std::array offsets{ data.uniformRange.offset, storage_offset(data.dataRanges.vtxDataRange), storage_offset(data.dataRanges.nrmDataRange), storage_offset(data.dataRanges.tcDataRange), storage_offset(data.dataRanges.packedTcDataRange), }; pass.SetBindGroup(0, find_bind_group(data.bindGroups.uniformBindGroup), offsets.size(), offsets.data()); if (data.bindGroups.samplerBindGroup && data.bindGroups.textureBindGroup) { pass.SetBindGroup(1, find_bind_group(data.bindGroups.samplerBindGroup)); pass.SetBindGroup(2, find_bind_group(data.bindGroups.textureBindGroup)); } pass.SetVertexBuffer(0, g_vertexBuffer, data.vertRange.offset, data.vertRange.size); pass.SetIndexBuffer(g_indexBuffer, wgpu::IndexFormat::Uint32, data.idxRange.offset, data.idxRange.size); pass.DrawIndexed(data.indexCount); } } // namespace aurora::gfx::model static absl::flat_hash_map sCachedRanges; template static inline void cache_array(const void* data, Vec*& outPtr, std::optional& outRange, u8 stride) { Vec* vecPtr = static_cast(data); outPtr = vecPtr; if (stride == 1) { // const auto hash = aurora::xxh3_hash(vecPtr->data(), vecPtr->size() * sizeof(typename Vec::value_type), 0); // const auto it = sCachedRanges.find(hash); // if (it != sCachedRanges.end()) { // outRange = it->second; // } else { // const auto range = aurora::gfx::push_static_storage(aurora::ArrayRef{*vecPtr}); // sCachedRanges.try_emplace(hash, range); // outRange = range; // } } else { outRange.reset(); } } void GXSetArray(GX::Attr attr, const void* data, u8 stride) noexcept { using namespace aurora::gfx::model; switch (attr) { case GX::VA_POS: cache_array(data, vtxData, staticVtxRange, stride); break; case GX::VA_NRM: cache_array(data, nrmData, staticNrmRange, stride); break; case GX::VA_TEX0: cache_array(data, tex0TcData, staticPackedTcRange, stride); break; case GX::VA_TEX1: cache_array(data, tcData, staticTcRange, stride); break; default: Log.report(logvisor::Fatal, FMT_STRING("GXSetArray: invalid attr {}"), attr); unreachable(); } } void GXCallDisplayList(const void* data, u32 nbytes) noexcept { aurora::gfx::model::queue_surface(static_cast(data), nbytes); }