/* * Copyright © 2016 Intel Corporation * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the "Software"), * to deal in the Software without restriction, including without limitation * the rights to use, copy, modify, merge, publish, distribute, sublicense, * and/or sell copies of the Software, and to permit persons to whom the * Software is furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice (including the next * paragraph) shall be included in all copies or substantial portions of the * Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. */ #include "nir.h" #include "nir_builder.h" static void build_constant_load(nir_builder *b, nir_deref_instr *deref, nir_constant *c) { if (glsl_type_is_vector_or_scalar(deref->type)) { const unsigned num_components = glsl_get_vector_elements(deref->type); const unsigned bit_size = glsl_get_bit_size(deref->type); nir_def *imm = nir_build_imm(b, num_components, bit_size, c->values); nir_store_deref(b, deref, imm, ~0); } else if (glsl_type_is_struct_or_ifc(deref->type)) { unsigned len = glsl_get_length(deref->type); for (unsigned i = 0; i < len; i++) { build_constant_load(b, nir_build_deref_struct(b, deref, i), c->elements[i]); } } else if (glsl_type_is_cmat(deref->type)) { const struct glsl_type *elem_type = glsl_get_cmat_element(deref->type); assert(glsl_type_is_scalar(elem_type)); const unsigned bit_size = glsl_get_bit_size(elem_type); nir_def *elem = nir_build_imm(b, 1, bit_size, c->values); nir_cmat_construct(b, &deref->def, elem); } else { assert(glsl_type_is_array(deref->type) || glsl_type_is_matrix(deref->type)); unsigned len = glsl_get_length(deref->type); for (unsigned i = 0; i < len; i++) { build_constant_load(b, nir_build_deref_array_imm(b, deref, i), c->elements[i]); } } } static bool lower_const_initializer(struct nir_builder *b, struct exec_list *var_list, nir_variable_mode modes) { bool progress = false; b->cursor = nir_before_impl(b->impl); nir_foreach_variable_in_list(var, var_list) { if (!(var->data.mode & modes)) continue; if (var->constant_initializer) { build_constant_load(b, nir_build_deref_var(b, var), var->constant_initializer); progress = true; var->constant_initializer = NULL; } else if (var->pointer_initializer) { nir_deref_instr *src_deref = nir_build_deref_var(b, var->pointer_initializer); nir_deref_instr *dst_deref = nir_build_deref_var(b, var); /* Note that this stores a pointer to src into dst */ nir_store_deref(b, dst_deref, &src_deref->def, ~0); progress = true; var->pointer_initializer = NULL; } } return progress; } bool nir_lower_variable_initializers(nir_shader *shader, nir_variable_mode modes) { bool progress = false; /* Only some variables have initializers that we want to lower. Others * such as uniforms have initializers which are useful later during linking * so we want to skip over those. Restrict to only variable types where * initializers make sense so that callers can use nir_var_all. */ modes &= nir_var_shader_out | nir_var_shader_temp | nir_var_function_temp | nir_var_system_value; nir_foreach_function_with_impl(func, impl, shader) { bool impl_progress = false; nir_builder builder = nir_builder_create(impl); if ((modes & ~nir_var_function_temp) && func->is_entrypoint) { impl_progress |= lower_const_initializer(&builder, &shader->variables, modes); } if (modes & nir_var_function_temp) { impl_progress |= lower_const_initializer(&builder, &impl->locals, nir_var_function_temp); } if (impl_progress) { progress = true; nir_metadata_preserve(impl, nir_metadata_control_flow | nir_metadata_live_defs); } else { nir_metadata_preserve(impl, nir_metadata_all); } } return progress; } /* Zero initialize shared_size bytes of shared memory by splitting work writes * of chunk_size bytes among the invocations. * * Used for implementing VK_KHR_zero_initialize_workgroup_memory. */ bool nir_zero_initialize_shared_memory(nir_shader *shader, const unsigned shared_size, const unsigned chunk_size) { assert(shared_size > 0); assert(chunk_size > 0); assert(chunk_size % 4 == 0); nir_function_impl *impl = nir_shader_get_entrypoint(shader); nir_builder b = nir_builder_at(nir_before_impl(impl)); assert(!shader->info.workgroup_size_variable); const unsigned local_count = shader->info.workgroup_size[0] * shader->info.workgroup_size[1] * shader->info.workgroup_size[2]; const unsigned stride = chunk_size * local_count; /* The initialization logic is simplified if we can always split the memory * in full chunk_size units. */ assert(shared_size % chunk_size == 0); const unsigned chunk_comps = chunk_size / 4; nir_def *local_index = nir_load_local_invocation_index(&b); nir_def *first_offset = nir_imul_imm(&b, local_index, chunk_size); if (stride >= shared_size) { nir_push_if(&b, nir_ult_imm(&b, first_offset, shared_size)); { nir_store_shared(&b, nir_imm_zero(&b, chunk_comps, 32), first_offset, .align_mul = chunk_size, .write_mask = ((1 << chunk_comps) - 1)); } nir_pop_if(&b, NULL); } else { nir_variable *it = nir_local_variable_create(b.impl, glsl_uint_type(), "zero_init_iterator"); nir_store_var(&b, it, first_offset, 0x1); nir_loop *loop = nir_push_loop(&b); { nir_def *offset = nir_load_var(&b, it); nir_push_if(&b, nir_uge_imm(&b, offset, shared_size)); { nir_jump(&b, nir_jump_break); } nir_pop_if(&b, NULL); nir_store_shared(&b, nir_imm_zero(&b, chunk_comps, 32), offset, .align_mul = chunk_size, .write_mask = ((1 << chunk_comps) - 1)); nir_def *new_offset = nir_iadd_imm(&b, offset, stride); nir_store_var(&b, it, new_offset, 0x1); } nir_pop_loop(&b, loop); } nir_barrier(&b, SCOPE_WORKGROUP, SCOPE_WORKGROUP, NIR_MEMORY_ACQ_REL, nir_var_mem_shared); nir_metadata_preserve(nir_shader_get_entrypoint(shader), nir_metadata_none); return true; } /** Clears all shared memory to zero at the end of the shader * * To easily get to the end of the shader it relies on all exits * being lowered. Designed to be called late in the lowering process, * e.g. doesn't need to lower vars to ssa. */ bool nir_clear_shared_memory(nir_shader *shader, const unsigned shared_size, const unsigned chunk_size) { assert(chunk_size > 0); assert(chunk_size % 4 == 0); if (shared_size == 0) return false; nir_function_impl *impl = nir_shader_get_entrypoint(shader); nir_builder b = nir_builder_at(nir_after_impl(impl)); /* The initialization logic is simplified if we can always split the memory * in full chunk_size units. */ assert(shared_size % chunk_size == 0); const unsigned chunk_comps = chunk_size / 4; nir_barrier(&b, SCOPE_WORKGROUP, SCOPE_WORKGROUP, NIR_MEMORY_ACQ_REL, nir_var_mem_shared); nir_def *local_index = nir_load_local_invocation_index(&b); nir_def *first_offset = nir_imul_imm(&b, local_index, chunk_size); unsigned iterations = UINT_MAX; unsigned size_per_iteration = 0; if (!shader->info.workgroup_size_variable) { size_per_iteration = nir_static_workgroup_size(shader) * chunk_size; iterations = DIV_ROUND_UP(shared_size, size_per_iteration); } if (iterations <= shader->options->max_unroll_iterations) { /* Doing a manual inline here because (a) we may not optimize after and * (b) the loop unroll pass doesn't deal well with the potential partial * last iteration.*/ for (unsigned i = 0; i < iterations; ++i) { const unsigned base = size_per_iteration * i; bool use_check = i >= shared_size / size_per_iteration; if (use_check) nir_push_if(&b, nir_ult_imm(&b, first_offset, shared_size - base)); nir_store_shared(&b, nir_imm_zero(&b, chunk_comps, 32), nir_iadd_imm(&b, first_offset, base), .align_mul = chunk_size, .write_mask = ((1 << chunk_comps) - 1)); if (use_check) nir_pop_if(&b, NULL); } } else { nir_phi_instr *offset_phi = nir_phi_instr_create(shader); nir_def_init(&offset_phi->instr, &offset_phi->def, 1, 32); nir_phi_instr_add_src(offset_phi, nir_cursor_current_block(b.cursor), first_offset); nir_def *size_per_iteration_def = shader->info.workgroup_size_variable ? nir_imul_imm(&b, nir_load_workgroup_size(&b), chunk_size) : nir_imm_int(&b, size_per_iteration); nir_def *value = nir_imm_zero(&b, chunk_comps, 32); nir_loop *loop = nir_push_loop(&b); nir_block *loop_block = nir_cursor_current_block(b.cursor); { nir_def *offset = &offset_phi->def; nir_push_if(&b, nir_uge_imm(&b, offset, shared_size)); { nir_jump(&b, nir_jump_break); } nir_pop_if(&b, NULL); nir_store_shared(&b, value, offset, .align_mul = chunk_size, .write_mask = ((1 << chunk_comps) - 1)); nir_def *new_offset = nir_iadd(&b, offset, size_per_iteration_def); nir_phi_instr_add_src(offset_phi, nir_cursor_current_block(b.cursor), new_offset); } nir_pop_loop(&b, loop); b.cursor = nir_before_block(loop_block); nir_builder_instr_insert(&b, &offset_phi->instr); } nir_metadata_preserve(nir_shader_get_entrypoint(shader), nir_metadata_none); return true; }