// Copyright (c) 2016-2020 NVIDIA Corporation // // SPDX-License-Identifier: CC-BY-4.0 include::{generated}/meta/{refprefix}VK_NV_viewport_swizzle.txt[] === Other Extension Metadata *Last Modified Date*:: 2016-12-22 *Interactions and External Dependencies*:: - This extension requires pname:multiViewport and pname:geometryShader features to be useful. *Contributors*:: - Daniel Koch, NVIDIA - Jeff Bolz, NVIDIA === Description This extension provides a new per-viewport swizzle that can modify the position of primitives sent to each viewport. New viewport swizzle state is added for each viewport, and a new position vector is computed for each vertex by selecting from and optionally negating any of the four components of the original position vector. This new viewport swizzle is useful for a number of algorithms, including single-pass cube map rendering (broadcasting a primitive to multiple faces and reorienting the vertex position for each face) and voxel rasterization. The per-viewport component remapping and negation provided by the swizzle allows application code to re-orient three-dimensional geometry with a view along any of the *X*, *Y*, or *Z* axes. If a perspective projection and depth buffering is required, [eq]#1/W# buffering should be used, as described in the single-pass cube map rendering example in the "`Issues`" section below. include::{generated}/interfaces/VK_NV_viewport_swizzle.txt[] === Issues 1) Where does viewport swizzling occur in the pipeline? *RESOLVED*: Despite being associated with the viewport, viewport swizzling must happen prior to the viewport transform. In particular, it needs to be performed before clipping and perspective division. The viewport mask expansion (`apiext:VK_NV_viewport_array2`) and the viewport swizzle could potentially be performed before or after transform feedback, but feeding back several viewports worth of primitives with different swizzles does not seem particularly useful. This specification applies the viewport mask and swizzle after transform feedback, and makes primitive queries only count each primitive once. 2) Any interesting examples of how this extension, `apiext:VK_NV_viewport_array2`, and `apiext:VK_NV_geometry_shader_passthrough` can be used together in practice? *RESOLVED*: One interesting use case for this extension is for single-pass rendering to a cube map. In this example, the application would attach a cube map texture to a layered FBO where the six cube faces are treated as layers. Vertices are sent through the vertex shader without applying a projection matrix, where the code:gl_Position output is [eq]#(x,y,z,1)# and the center of the cube map is at [eq]#(0,0,0)#. With unextended Vulkan, one could have a conventional instanced geometry shader that looks something like the following: [source,c] --------------------------------------------------- layout(invocations = 6) in; // separate invocation per face layout(triangles) in; layout(triangle_strip) out; layout(max_vertices = 3) out; in Inputs { vec2 texcoord; vec3 normal; vec4 baseColor; } v[]; out Outputs { vec2 texcoord; vec3 normal; vec4 baseColor; }; void main() { int face = gl_InvocationID; // which face am I? // Project gl_Position for each vertex onto the cube map face. vec4 positions[3]; for (int i = 0; i < 3; i++) { positions[i] = rotate(gl_in[i].gl_Position, face); } // If the primitive does not project onto this face, we are done. if (shouldCull(positions)) { return; } // Otherwise, emit a copy of the input primitive to the // appropriate face (using gl_Layer). for (int i = 0; i < 3; i++) { gl_Layer = face; gl_Position = positions[i]; texcoord = v[i].texcoord; normal = v[i].normal; baseColor = v[i].baseColor; EmitVertex(); } } --------------------------------------------------- With passthrough geometry shaders, this can be done using a much simpler shader: [source,c] --------------------------------------------------- layout(triangles) in; layout(passthrough) in Inputs { vec2 texcoord; vec3 normal; vec4 baseColor; } layout(passthrough) in gl_PerVertex { vec4 gl_Position; } gl_in[]; layout(viewport_relative) out int gl_Layer; void main() { // Figure out which faces the primitive projects onto and // generate a corresponding viewport mask. uint mask = 0; for (int i = 0; i < 6; i++) { if (!shouldCull(face)) { mask |= 1U << i; } } gl_ViewportMask = mask; gl_Layer = 0; } --------------------------------------------------- The application code is set up so that each of the six cube faces has a separate viewport (numbered 0 to 5). Each face also has a separate swizzle, programmed via the slink:VkPipelineViewportSwizzleStateCreateInfoNV pipeline state. The viewport swizzle feature performs the coordinate transformation handled by the code:rotate() function in the original shader. The code:viewport_relative layout qualifier says that the viewport number (0 to 5) is added to the base code:gl_Layer value of 0 to determine which layer (cube face) the primitive should be sent to. Note that the use of the passed through input code:normal in this example suggests that the fragment shader in this example would perform an operation like per-fragment lighting. The viewport swizzle would transform the position to be face-relative, but code:normal would remain in the original coordinate system. It seems likely that the fragment shader in either version of the example would want to perform lighting in the original coordinate system. It would likely do this by reconstructing the position of the fragment in the original coordinate system using code:gl_FragCoord, a constant or uniform holding the size of the cube face, and the input code:gl_ViewportIndex (or code:gl_Layer), which identifies the cube face. Since the value of code:normal is in the original coordinate system, it would not need to be modified as part of this coordinate transformation. Note that while the code:rotate() operation in the regular geometry shader above could include an arbitrary post-rotation projection matrix, the viewport swizzle does not support arbitrary math. To get proper projection, [eq]#1/W# buffering should be used. To do this: . Program the viewport swizzles to move the pre-projection [eq]#W# eye coordinate (typically 1.0) into the [eq]#Z# coordinate of the swizzle output and the eye coordinate component used for depth into the [eq]#W# coordinate. For example, the viewport corresponding to the [eq]#+Z# face might use a swizzle of [eq]#(+X, -Y, +W, +Z)#. The [eq]#Z# normalized device coordinate computed after swizzling would then be [eq]#z'/w' = 1/Z~eye~#. . On NVIDIA implementations supporting floating-point depth buffers with values outside [eq]#[0,1]#, prevent unwanted near plane clipping by enabling pname:depthClampEnable. Ensure that the depth clamp does not mess up depth testing by programming the depth range to very large values, such as [eq]#pname:minDepthBounds=-z#, [eq]#pname:maxDepthBounds=+z#, where [eq]#z = 2^127^#. It should be possible to use IEEE infinity encodings also (`0xFF800000` for `-INF`, `0x7F800000` for `+INF`). Even when near/far clipping is disabled, primitives extending behind the eye will still be clipped because one or more vertices will have a negative [eq]#W# coordinate and fail [eq]#X#/[eq]#Y# clipping tests. + -- On other implementations, scale [eq]#X#, [eq]#Y#, and [eq]#Z# eye coordinates so that vertices on the near plane have a post-swizzle [eq]#W# coordinate of 1.0. For example, if the near plane is at [eq]#Z~eye~ = 1/256#, scale [eq]#X#, [eq]#Y#, and [eq]#Z# by 256. -- . Adjust depth testing to reflect the fact that [eq]#1/W# values are large near the eye and small away from the eye. Clear the depth buffer to zero (infinitely far away) and use a depth test of ename:VK_COMPARE_OP_GREATER instead of ename:VK_COMPARE_OP_LESS. === Version History * Revision 1, 2016-12-22 (Piers Daniell) - Internal revisions