/* * Copyright (C) 2005-2007 Brian Paul All Rights Reserved. * Copyright (C) 2008 VMware, Inc. All Rights Reserved. * Copyright © 2014 Intel Corporation * Copyright © 2017 Advanced Micro Devices, Inc. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER * DEALINGS IN THE SOFTWARE. */ /** * \file * * Lower sampler and image references of (non-bindless) uniforms by removing * struct dereferences, and synthesizing new uniform variables without structs * if required. * * This will allow backends to have a simple, uniform treatment of bindless and * non-bindless samplers and images. * * Example: * * struct S { * sampler2D tex[2]; * sampler2D other; * }; * uniform S s[2]; * * tmp = texture(s[n].tex[m], coord); * * Becomes: * * decl_var uniform INTERP_MODE_NONE sampler2D[2][2] lower@s.tex (...) * * vec1 32 ssa_idx = $(2 * n + m) * vec4 32 ssa_out = tex ssa_coord (coord), lower@s.tex[n][m] (texture), lower@s.tex[n][m] (sampler) * * and lower@s.tex has var->data.binding set to the base index as defined by * the opaque uniform mapping. */ #include "compiler/nir/nir.h" #include "compiler/nir/nir_builder.h" #include "compiler/nir/nir_deref.h" #include "gl_nir.h" #include "ir_uniform.h" #include "util/compiler.h" #include "main/mtypes.h" struct lower_samplers_as_deref_state { nir_shader *shader; const struct gl_shader_program *shader_program; struct hash_table *remap_table; }; /* Prepare for removing struct derefs. This pre-pass generates the name * of the lowered deref, and calculates the lowered type and location. * After that, once looking up (or creating if needed) the lowered var, * constructing the new chain of deref instructions is a simple loop * that skips the struct deref's * * path: appended to as we descend down the chain of deref instrs * and remove struct derefs * location: increased as we descend down and remove struct derefs * type: updated as we recurse back up the chain of deref instrs * with the resulting type after removing struct derefs */ static void remove_struct_derefs_prep(nir_deref_instr **p, char **name, unsigned *location, const struct glsl_type **type) { nir_deref_instr *cur = p[0], *next = p[1]; if (!next) { *type = cur->type; return; } switch (next->deref_type) { case nir_deref_type_array: { unsigned length = glsl_get_length(cur->type); remove_struct_derefs_prep(&p[1], name, location, type); *type = glsl_array_type(*type, length, glsl_get_explicit_stride(cur->type)); break; } case nir_deref_type_struct: { *location += glsl_get_struct_location_offset(cur->type, next->strct.index); ralloc_asprintf_append(name, ".%s", glsl_get_struct_elem_name(cur->type, next->strct.index)); remove_struct_derefs_prep(&p[1], name, location, type); break; } default: unreachable("Invalid deref type"); break; } } static void record_images_used(struct shader_info *info, nir_deref_instr *deref) { nir_variable *var = nir_deref_instr_get_variable(deref); /* Structs have been lowered already, so get_aoa_size is sufficient. */ const unsigned size = glsl_type_is_array(var->type) ? glsl_get_aoa_size(var->type) : 1; unsigned mask = ((1ull << MAX2(size, 1)) - 1) << var->data.binding; info->images_used |= mask; } static nir_deref_instr * lower_deref(nir_builder *b, struct lower_samplers_as_deref_state *state, nir_deref_instr *deref) { nir_variable *var = nir_deref_instr_get_variable(deref); gl_shader_stage stage = state->shader->info.stage; if (var->data.bindless || var->data.mode != nir_var_uniform) return NULL; nir_deref_path path; nir_deref_path_init(&path, deref, state->remap_table); assert(path.path[0]->deref_type == nir_deref_type_var); char *name = ralloc_asprintf(state->remap_table, "lower@%s", var->name); unsigned location = var->data.location; const struct glsl_type *type = NULL; unsigned binding; /* * We end up needing to do this in two passes, in order to generate * the name of the lowered var (and detecting whether there even are * any struct deref's), and then the second pass to construct the * actual deref instructions after looking up / generating a new * nir_variable (since we need to construct the deref_var first) */ remove_struct_derefs_prep(path.path, &name, &location, &type); if (state->shader_program && var->data.how_declared != nir_var_hidden) { /* For GLSL programs, look up the bindings in the uniform storage. */ assert(location < state->shader_program->data->NumUniformStorage && state->shader_program->data->UniformStorage[location].opaque[stage].active); binding = state->shader_program->data->UniformStorage[location].opaque[stage].index; } else { /* For ARB programs, built-in shaders, or internally generated sampler * variables in GLSL programs, assume that whoever created the shader * set the bindings correctly already. */ assert(var->data.explicit_binding); binding = var->data.binding; } if (var->type == type) { /* Fast path: We did not encounter any struct derefs. */ var->data.binding = binding; return deref; } uint32_t hash = _mesa_hash_string(name); struct hash_entry *h = _mesa_hash_table_search_pre_hashed(state->remap_table, hash, name); if (h) { var = (nir_variable *)h->data; } else { var = nir_variable_create(state->shader, nir_var_uniform, type, name); var->data.binding = binding; /* Don't set var->data.location. The old structure location could be * used to index into gl_uniform_storage, assuming the full structure * was walked in order. With the new split variables, this invariant * no longer holds and there's no meaningful way to start from a base * location and access a particular array element. Just leave it 0. */ _mesa_hash_table_insert_pre_hashed(state->remap_table, hash, name, var); } /* construct a new deref based on lowered var (skipping the struct deref's * from the original deref: */ nir_deref_instr *new_deref = nir_build_deref_var(b, var); for (nir_deref_instr **p = &path.path[1]; *p; p++) { if ((*p)->deref_type == nir_deref_type_struct) continue; assert((*p)->deref_type == nir_deref_type_array); new_deref = nir_build_deref_array(b, new_deref, nir_ssa_for_src(b, (*p)->arr.index, 1)); } return new_deref; } static void record_textures_used(struct shader_info *info, nir_deref_instr *deref, nir_texop op) { nir_variable *var = nir_deref_instr_get_variable(deref); /* Structs have been lowered already, so get_aoa_size is sufficient. */ const unsigned size = glsl_type_is_array(var->type) ? glsl_get_aoa_size(var->type) : 1; unsigned mask = ((1ull << MAX2(size, 1)) - 1) << var->data.binding; info->textures_used |= mask; if (op == nir_texop_txf || op == nir_texop_txf_ms || op == nir_texop_txf_ms_mcs) info->textures_used_by_txf |= mask; } static bool lower_sampler(nir_tex_instr *instr, struct lower_samplers_as_deref_state *state, nir_builder *b) { int texture_idx = nir_tex_instr_src_index(instr, nir_tex_src_texture_deref); int sampler_idx = nir_tex_instr_src_index(instr, nir_tex_src_sampler_deref); b->cursor = nir_before_instr(&instr->instr); if (texture_idx >= 0) { assert(instr->src[texture_idx].src.is_ssa); nir_deref_instr *texture_deref = lower_deref(b, state, nir_src_as_deref(instr->src[texture_idx].src)); /* only lower non-bindless: */ if (texture_deref) { nir_instr_rewrite_src(&instr->instr, &instr->src[texture_idx].src, nir_src_for_ssa(&texture_deref->dest.ssa)); record_textures_used(&b->shader->info, texture_deref, instr->op); } } if (sampler_idx >= 0) { assert(instr->src[sampler_idx].src.is_ssa); nir_deref_instr *sampler_deref = lower_deref(b, state, nir_src_as_deref(instr->src[sampler_idx].src)); /* only lower non-bindless: */ if (sampler_deref) { nir_instr_rewrite_src(&instr->instr, &instr->src[sampler_idx].src, nir_src_for_ssa(&sampler_deref->dest.ssa)); } } return true; } static bool lower_intrinsic(nir_intrinsic_instr *instr, struct lower_samplers_as_deref_state *state, nir_builder *b) { if (instr->intrinsic == nir_intrinsic_image_deref_load || instr->intrinsic == nir_intrinsic_image_deref_store || instr->intrinsic == nir_intrinsic_image_deref_atomic_add || instr->intrinsic == nir_intrinsic_image_deref_atomic_imin || instr->intrinsic == nir_intrinsic_image_deref_atomic_umin || instr->intrinsic == nir_intrinsic_image_deref_atomic_imax || instr->intrinsic == nir_intrinsic_image_deref_atomic_umax || instr->intrinsic == nir_intrinsic_image_deref_atomic_and || instr->intrinsic == nir_intrinsic_image_deref_atomic_or || instr->intrinsic == nir_intrinsic_image_deref_atomic_xor || instr->intrinsic == nir_intrinsic_image_deref_atomic_exchange || instr->intrinsic == nir_intrinsic_image_deref_atomic_comp_swap || instr->intrinsic == nir_intrinsic_image_deref_atomic_fadd || instr->intrinsic == nir_intrinsic_image_deref_size) { b->cursor = nir_before_instr(&instr->instr); nir_deref_instr *deref = lower_deref(b, state, nir_src_as_deref(instr->src[0])); record_images_used(&state->shader->info, deref); /* don't lower bindless: */ if (!deref) return false; nir_instr_rewrite_src(&instr->instr, &instr->src[0], nir_src_for_ssa(&deref->dest.ssa)); return true; } return false; } static bool lower_impl(nir_function_impl *impl, struct lower_samplers_as_deref_state *state) { nir_builder b; nir_builder_init(&b, impl); bool progress = false; nir_foreach_block(block, impl) { nir_foreach_instr(instr, block) { if (instr->type == nir_instr_type_tex) progress |= lower_sampler(nir_instr_as_tex(instr), state, &b); else if (instr->type == nir_instr_type_intrinsic) progress |= lower_intrinsic(nir_instr_as_intrinsic(instr), state, &b); } } if (progress) { nir_metadata_preserve(impl, nir_metadata_block_index | nir_metadata_dominance); } else { nir_metadata_preserve(impl, nir_metadata_all); } return progress; } bool gl_nir_lower_samplers_as_deref(nir_shader *shader, const struct gl_shader_program *shader_program) { bool progress = false; struct lower_samplers_as_deref_state state; state.shader = shader; state.shader_program = shader_program; state.remap_table = _mesa_hash_table_create(NULL, _mesa_hash_string, _mesa_key_string_equal); nir_foreach_function(function, shader) { if (function->impl) progress |= lower_impl(function->impl, &state); } /* keys are freed automatically by ralloc */ _mesa_hash_table_destroy(state.remap_table, NULL); if (progress) nir_remove_dead_derefs(shader); return progress; }