Home
last modified time | relevance | path

Searched refs:align_mul (Results 1 – 25 of 54) sorted by relevance

123

/third_party/mesa3d/src/amd/vulkan/
Dradv_acceleration_structure.c722 .align_mul = 2, .align_offset = 0); in get_indices()
730 b, 3, 32, nir_iadd(b, addr, nir_u2u64(b, index_id)), .align_mul = 4, .align_offset = 0); in get_indices()
751 .align_mul = 1, .align_offset = 0); in get_indices()
786 .align_mul = 4, .align_offset = 0), in get_vertices()
805 .align_mul = comp_bytes, .align_offset = 0); in get_vertices()
987 .align_mul = 4, .align_offset = 0), in build_leaf_shader()
992 .align_mul = 4, .align_offset = 0), in build_leaf_shader()
997 .align_mul = 4, .align_offset = 0), in build_leaf_shader()
1015 .write_mask = 15, .align_mul = 16, .align_offset = 0); in build_leaf_shader()
1019 nir_build_store_global(&b, node_id, scratch_addr, .write_mask = 1, .align_mul = 4, in build_leaf_shader()
[all …]
Dradv_query.c80 .align_mul = 8); in radv_store_availability()
84 nir_store_ssbo(b, value32, dst_buf, offset, .write_mask = 0x1, .align_mul = 4); in radv_store_availability()
176 nir_ssa_def *load = nir_load_ssbo(&b, 2, 64, src_buf, load_offset, .align_mul = 16); in build_occlusion_query_shader()
210 .align_mul = 8); in build_occlusion_query_shader()
215 .write_mask = 0x1, .align_mul = 8); in build_occlusion_query_shader()
296 nir_ssa_def *available32 = nir_load_ssbo(&b, 1, 32, src_buf, avail_offset, .align_mul = 4); in build_pipeline_statistics_query_shader()
314 nir_ssa_def *start = nir_load_ssbo(&b, 1, 64, src_buf, start_offset, .align_mul = 8); in build_pipeline_statistics_query_shader()
319 nir_ssa_def *end = nir_load_ssbo(&b, 1, 64, src_buf, end_offset, .align_mul = 8); in build_pipeline_statistics_query_shader()
327 .align_mul = 8); in build_pipeline_statistics_query_shader()
332 .write_mask = 0x1, .align_mul = 4); in build_pipeline_statistics_query_shader()
[all …]
Dradv_pipeline_rt.c319 nir_load_scratch(b, 1, 32, nir_load_var(b, vars->stack_ptr), .align_mul = 16), 1); in insert_rt_return()
352 nir_build_load_global(b, 1, 32, load_addr, .align_mul = 4, .align_offset = 0); in load_sbt_entry()
397 .align_mul = 64, .align_offset = offset + i * 16); in nir_build_wto_matrix_load()
425 nir_load_var(&b_shader, vars->stack_ptr), .align_mul = 16, in lower_rt_instructions()
452 nir_load_var(&b_shader, vars->stack_ptr), .align_mul = 16, in lower_rt_instructions()
638 .align_mul = 4, .align_offset = 0); in lower_rt_instructions()
650 .align_mul = 64, .align_offset = 16), in lower_rt_instructions()
654 .align_mul = 64, .align_offset = 32), in lower_rt_instructions()
658 .align_mul = 64, .align_offset = 48)}; in lower_rt_instructions()
1267 .align_mul = 4, .align_offset = 0); in insert_traversal_triangle_case()
[all …]
Dradv_meta_copy_vrs_htile.c122 nir_ssa_def *input_value = nir_load_ssbo(&b, 1, 32, htile_buf, htile_addr, .align_mul = 4); in build_copy_vrs_htile_shader()
138 .access = ACCESS_NON_READABLE, .align_mul = 4); in build_copy_vrs_htile_shader()
Dradv_meta_buffer.c26 .access = ACCESS_NON_READABLE, .align_mul = 16); in build_buffer_fill_shader()
47 nir_ssa_def *load = nir_load_ssbo(&b, 4, 32, src_buf, offset, .align_mul = 16); in build_buffer_copy_shader()
49 .align_mul = 16); in build_buffer_copy_shader()
/third_party/mesa3d/src/compiler/nir/
Dnir_lower_ubo_vec4.c93 unsigned align_mul = nir_intrinsic_align_mul(intr); in nir_lower_ubo_vec4_lower() local
102 align_mul = MIN2(align_mul, 16); in nir_lower_ubo_vec4_lower()
107 bool aligned_mul = (align_mul == 16 && in nir_lower_ubo_vec4_lower()
136 } else if (align_mul == 8 && in nir_lower_ubo_vec4_lower()
Dnir_lower_io.c1198 uint32_t align_mul, uint32_t align_offset, in build_explicit_io_load() argument
1208 align_mul, align_offset, in build_explicit_io_load()
1216 align_mul, align_offset, in build_explicit_io_load()
1222 align_mul, align_offset, in build_explicit_io_load()
1233 align_mul, align_offset, in build_explicit_io_load()
1240 align_mul, align_offset, in build_explicit_io_load()
1377 nir_intrinsic_set_align(load, align_mul, align_offset); in build_explicit_io_load()
1439 uint32_t align_mul, uint32_t align_offset, in build_explicit_io_store() argument
1448 align_mul, align_offset, in build_explicit_io_store()
1455 align_mul, align_offset, in build_explicit_io_store()
[all …]
Dnir_lower_wrmasks.c133 unsigned align_mul = nir_intrinsic_align_mul(intr); in split_wrmask() local
137 align_off = align_off % align_mul; in split_wrmask()
139 nir_intrinsic_set_align(new_intr, align_mul, align_off); in split_wrmask()
Dnir_lower_printf.c68 counter->cast.align_mul = 4; in lower_printf_instr()
91 fmt_str_id_deref->cast.align_mul = 4; in lower_printf_instr()
122 dst_arg_deref->cast.align_mul = 4; in lower_printf_instr()
Dnir_opt_load_store_vectorize.c176 uint32_t align_mul; member
553 uint32_t align_mul = 31; in calc_alignment() local
556 align_mul = MIN2(align_mul, ffsll(entry->key->offset_defs_mul[i])); in calc_alignment()
559 entry->align_mul = 1u << (align_mul - 1); in calc_alignment()
561 if (!has_align || entry->align_mul >= nir_intrinsic_align_mul(entry->intrin)) { in calc_alignment()
562 entry->align_offset = entry->offset % entry->align_mul; in calc_alignment()
564 entry->align_mul = nir_intrinsic_align_mul(entry->intrin); in calc_alignment()
663 if (!ctx->options->callback(low->align_mul, in new_bitsize_acceptable()
804 first->align_mul = low->align_mul; in vectorize_loads()
890 second->align_mul = low->align_mul; in vectorize_stores()
[all …]
Dnir_deref.c883 if (cast->cast.align_mul == 0) in opt_remove_restricting_cast_alignments()
907 if (parent_mul < cast->cast.align_mul) in opt_remove_restricting_cast_alignments()
937 assert(cast->cast.align_mul <= parent_mul); in opt_remove_restricting_cast_alignments()
938 if (parent_offset % cast->cast.align_mul != cast->cast.align_offset) in opt_remove_restricting_cast_alignments()
944 cast->cast.align_mul = 0; in opt_remove_restricting_cast_alignments()
1049 if (cast->cast.align_mul > 0) in opt_replace_struct_wrapper_cast()
1091 if (cast->cast.align_mul > 0) in opt_deref_cast()
1140 parent->cast.align_mul == 0 && in opt_deref_ptr_as_array()
1176 if (cast->cast.align_mul > 0) in is_vector_bitcast_deref()
Dnir_lower_scratch.c57 b, intrin->num_components, bit_size == 1 ? 32 : bit_size, offset, .align_mul=align); in lower_load_store()
70 nir_store_scratch(b, value, offset, .align_mul=align, in lower_load_store()
Dnir_opt_memcpy.c43 if (cast->cast.align_mul > 0) in opt_memcpy_deref_cast()
Dnir_lower_variable_initializers.c186 .align_mul=chunk_size, in nir_zero_initialize_shared_memory()
Dnir_instr_set.c175 hash = HASH(hash, instr->cast.align_mul); in hash_deref()
629 deref1->cast.align_mul != deref2->cast.align_mul || in nir_instrs_equal()
Dnir.h1585 unsigned align_mul; member
1961 unsigned align_mul, unsigned align_offset) in nir_intrinsic_set_align() argument
1963 assert(util_is_power_of_two_nonzero(align_mul)); in nir_intrinsic_set_align()
1964 assert(align_offset < align_mul); in nir_intrinsic_set_align()
1965 nir_intrinsic_set_align_mul(intrin, align_mul); in nir_intrinsic_set_align()
1979 const unsigned align_mul = nir_intrinsic_align_mul(intrin); in nir_intrinsic_align() local
1981 assert(align_offset < align_mul); in nir_intrinsic_align()
1982 return align_offset ? 1 << (ffs(align_offset) - 1) : align_mul; in nir_intrinsic_align()
4932 uint32_t *align_mul,
5614 typedef bool (*nir_should_vectorize_mem_func)(unsigned align_mul,
/third_party/mesa3d/src/gallium/drivers/etnaviv/tests/
Dlower_ubo_tests.cpp143 …nir_load_ubo(&b, 1, 32, index, offset, .align_mul = 16, .align_offset = 0, .range_base = 0, .range… in TEST_F()
158 …nir_load_ubo(&b, 1, 32, index, offset, .align_mul = 16, .align_offset = 0, .range_base = 0, .range… in TEST_F()
175 …nir_load_ubo(&b, 1, 32, index, offset, .align_mul = 16, .align_offset = 0, .range_base = 0, .range… in TEST_F()
/third_party/mesa3d/src/amd/common/
Dac_nir_lower_ngg.c249 …urviving_invocations_in_current_wave), wave_id, .base = lds_addr_base, .align_mul = 1u, .write_mas… in repack_invocations_in_workgroup()
254 …_load_shared(b, 1, num_lds_dwords * 32, nir_imm_int(b, 0), .base = lds_addr_base, .align_mul = 8u); in repack_invocations_in_workgroup()
366 nir_build_store_shared(b, prim_id, addr, .write_mask = 1u, .align_mul = 4u); in emit_ngg_nogs_prim_export()
389 prim_id = nir_build_load_shared(b, 1, 32, addr, .align_mul = 4u); in emit_store_ngg_nogs_es_primitive_id()
697 …(b, es_exporter_tid), es_vertex_lds_addr, .base = lds_es_exporter_tid, .align_mul = 1u, .write_mas… in compact_vertices_after_culling()
701 …nir_build_store_shared(b, pos, exporter_addr, .base = lds_es_pos_x, .align_mul = 4u, .write_mask =… in compact_vertices_after_culling()
706 …store_shared(b, arg_val, exporter_addr, .base = lds_es_arg_0 + 4u * i, .align_mul = 4u, .write_mas… in compact_vertices_after_culling()
724 …d_pos = nir_build_load_shared(b, 4, 32, es_vertex_lds_addr, .base = lds_es_pos_x, .align_mul = 4u); in compact_vertices_after_culling()
729 …ir_build_load_shared(b, 1, 32, es_vertex_lds_addr, .base = lds_es_arg_0 + 4u * i, .align_mul = 4u); in compact_vertices_after_culling()
748 …r_vtx_idx = nir_build_load_shared(b, 1, 8, vtx_addr, .base = lds_es_exporter_tid, .align_mul = 1u); in compact_vertices_after_culling()
[all …]
Dac_nir_lower_tess_io_to_mem.c226 … .align_mul = 16u, .align_offset = (nir_intrinsic_component(intrin) * 4u) % 16u); in lower_ls_output_store()
383 … .align_mul = 16u, .align_offset = (nir_intrinsic_component(intrin) * 4u) % 16u); in lower_hs_per_vertex_input_load()
421 … .align_mul = 16u, .align_offset = (nir_intrinsic_component(intrin) * 4u) % 16u); in lower_hs_output_store()
432 … .align_mul = 16u, .align_offset = (nir_intrinsic_component(intrin) * 4u) % 16u); in lower_hs_output_load()
523 … .align_mul = 16u, .align_offset = st->tcs_tess_lvl_out_loc % 16u); in hs_emit_write_tess_factors()
526 … .align_mul = 16u, .align_offset = st->tcs_tess_lvl_in_loc % 16u) in hs_emit_write_tess_factors()
Dac_nir_lower_esgs_io_to_mem.c143 … .align_mul = 16u, .align_offset = (nir_intrinsic_component(intrin) * 4u) % 16u); in lower_es_output_store()
217 … .align_mul = 16u, .align_offset = (nir_intrinsic_component(intrin) * 4u) % 16u); in lower_gs_per_vertex_input_load()
/third_party/mesa3d/src/gallium/drivers/radeonsi/
Dsi_shaderlib_nir.c92 nir_ssa_def *value = nir_load_ssbo(&b, 1, 8, zero, src_offset, .align_mul=1); in si_create_dcc_retile_cs()
99 nir_store_ssbo(&b, value, zero, dst_offset, .write_mask=0x1, .align_mul=1); in si_create_dcc_retile_cs()
145 nir_store_ssbo(&b, clear_value, zero, offset, .write_mask=0x1, .align_mul=2); in gfx9_create_clear_dcc_msaa_cs()
/third_party/mesa3d/src/intel/vulkan/
Danv_nir_lower_ubo_loads.c97 .align_mul = nir_intrinsic_align_mul(load), in lower_ubo_load_instr()
107 .align_mul = nir_intrinsic_align_mul(load), in lower_ubo_load_instr()
Danv_nir_apply_pipeline_layout.c266 .align_mul = 8, in build_load_descriptor_mem()
277 .align_mul = 8, in build_load_descriptor_mem()
936 cast->cast.align_mul = ANV_UBO_ALIGNMENT; in lower_load_vulkan_descriptor()
942 cast->cast.align_mul = ANV_SSBO_ALIGNMENT; in lower_load_vulkan_descriptor()
1105 .align_mul = intrin->dest.ssa.bit_size / 8, in lower_load_constant()
/third_party/mesa3d/src/freedreno/ir3/
Dir3_nir.c158 ir3_nir_should_vectorize_mem(unsigned align_mul, unsigned align_offset, in ir3_nir_should_vectorize_mem() argument
171 assert(util_is_power_of_two_nonzero(align_mul)); in ir3_nir_should_vectorize_mem()
172 align_mul = MIN2(align_mul, 16); in ir3_nir_should_vectorize_mem()
176 if (align_mul < 4) in ir3_nir_should_vectorize_mem()
179 unsigned worst_start_offset = 16 - align_mul + align_offset; in ir3_nir_should_vectorize_mem()
/third_party/mesa3d/src/intel/compiler/
Dbrw_nir_lower_mem_access_bit_sizes.c169 const unsigned align_mul = nir_intrinsic_align_mul(intrin); in lower_mem_store_bit_size() local
210 (align_mul >= 4 && (align_offset + start) % 4 == 0) || in lower_mem_store_bit_size()

123