Ben Clayton 068eb3ebda tint: Fix HLSL texture queries and re-add textureNumLayers()
... overloads for texture cube arrays

FXC is telling porkies, when it says "The array element count
of GetDimensions on TextureCubeArray objects is unavailable
on ps_5_1".

The actual issue, as identified by Teodor Tanasoaia at Mozilla,
is that the argument needs to be unsigned.

In fact, *all* the texture queries should have used an unsigned
scalar or vector for the output value. This has been broken
forever!

This reverts commit bd9f6e66842ed986dfff3264326bb385d467c5f3.

Change-Id: I3e217bec17c6fd203cff618b143ebef3d8a61927
Reviewed-on: https://dawn-review.googlesource.com/c/dawn/+/122980
Auto-Submit: Ben Clayton <bclayton@google.com>
Kokoro: Kokoro <noreply+kokoro@google.com>
Reviewed-by: Dan Sinclair <dsinclair@chromium.org>
Commit-Queue: Dan Sinclair <dsinclair@chromium.org>
2023-03-08 02:48:34 +00:00

25 lines
481 B
WebGPU Shading Language

@group(1) @binding(0) var arg_0 : texture_depth_cube_array;
fn textureNumLayers_48ef47() {
var res : u32 = textureNumLayers(arg_0);
prevent_dce = res;
}
@group(2) @binding(0) var<storage, read_write> prevent_dce : u32;
@vertex
fn vertex_main() -> @builtin(position) vec4<f32> {
textureNumLayers_48ef47();
return vec4<f32>();
}
@fragment
fn fragment_main() {
textureNumLayers_48ef47();
}
@compute @workgroup_size(1)
fn compute_main() {
textureNumLayers_48ef47();
}