/third_party/mesa3d/src/intel/compiler/ |
D | brw_fs_lower_regioning.cpp | 41 return type_sz(inst->dst.type) == 1 && in is_byte_raw_mov() 69 return inst->dst.stride * type_sz(inst->dst.type); in required_dst_byte_stride() 70 } else if (type_sz(inst->dst.type) < get_exec_type_size(inst) && in required_dst_byte_stride() 78 unsigned max_stride = inst->dst.stride * type_sz(inst->dst.type); in required_dst_byte_stride() 79 unsigned min_size = type_sz(inst->dst.type); in required_dst_byte_stride() 80 unsigned max_size = type_sz(inst->dst.type); in required_dst_byte_stride() 84 const unsigned size = type_sz(inst->src[i].type); in required_dst_byte_stride() 150 intel_device_info_is_9lp(devinfo)) && type_sz(t) > 4) in required_exec_type() 153 return brw_int_type(type_sz(t), false); in required_exec_type() 158 if (!has_64bit && type_sz(t) > 4) in required_exec_type() [all …]
|
D | brw_ir_fs.h | 118 return byte_offset(reg, delta * reg.stride * type_sz(reg.type)); in horiz_offset() 125 return byte_offset(reg, delta * stride * type_sz(reg.type)); in horiz_offset() 199 return (MAX2(1, stride) - 1) * type_sz(r.type); in reg_padding() 294 assert((i + 1) * type_sz(type) <= type_sz(reg.type)); in subscript() 300 const int delta = util_logbase2(type_sz(reg.type)) - in subscript() 301 util_logbase2(type_sz(type)); in subscript() 306 unsigned bit_size = type_sz(type) * 8; in subscript() 313 reg.stride *= type_sz(reg.type) / type_sz(type); in subscript() 316 return byte_offset(retype(reg, type), i * type_sz(type)); in subscript() 482 if (type_sz(t) > type_sz(exec_type)) in get_exec_type() [all …]
|
D | brw_fs_cmod_propagation.cpp | 340 if (type_sz(scan_inst->dst.type) > type_sz(inst->dst.type)) in opt_cmod_propagation_local() 359 type_sz(scan_inst->dst.type) != type_sz(inst->src[0].type)) in opt_cmod_propagation_local() 363 if (type_sz(inst->dst.type) < type_sz(scan_inst->dst.type)) in opt_cmod_propagation_local() 387 if (type_sz(scan_inst->dst.type) != type_sz(inst->dst.type)) in opt_cmod_propagation_local()
|
D | brw_ir_vec4.h | 111 return byte_offset(reg, num_components * type_sz(reg.type) * delta); in offset() 117 return byte_offset(reg, delta * type_sz(reg.type)); in horiz_offset() 192 return byte_offset(reg, num_components * type_sz(reg.type) * delta); in offset() 201 return byte_offset(reg, delta * type_sz(reg.type)); in horiz_offset() 448 if (type_sz(t) > type_sz(exec_type)) in get_exec_type() 450 else if (type_sz(t) == type_sz(exec_type) && in get_exec_type() 470 return type_sz(get_exec_type(inst)); in get_exec_type_size()
|
D | brw_fs_copy_propagation.cpp | 385 !(type_sz(inst->src[arg].type) * stride == in can_take_stride() 386 type_sz(dst_type) * inst->dst.stride || in can_take_stride() 401 if (type_sz(inst->src[arg].type) > 4) in can_take_stride() 611 if ((type_sz(entry->dst.type) < type_sz(inst->src[arg].type) || in try_copy_propagate() 632 type_sz(inst->src[arg].type)) % type_sz(entry->src.type) != 0) in try_copy_propagate() 644 type_sz(entry->dst.type) != type_sz(inst->src[arg].type))) in try_copy_propagate() 683 const unsigned reg_width = REG_SIZE / (type_sz(inst->src[arg].type) * in try_copy_propagate() 710 const unsigned component = rel_offset / type_sz(entry->dst.type); in try_copy_propagate() 711 const unsigned suboffset = rel_offset % type_sz(entry->dst.type); in try_copy_propagate() 717 component * entry_stride * type_sz(entry->src.type) + suboffset); in try_copy_propagate() [all …]
|
D | brw_eu_validate.c | 714 (type_sz(dst_type) == 1 || type_sz(src0_type) == 1)) { in is_byte_conversion() 719 (type_sz(dst_type) == 1 || type_sz(src1_type) == 1); in is_byte_conversion() 872 ERROR_IF(type_sz(dst_type) == 1 && in general_restrictions_based_on_operand_types() 873 (type_sz(src0_type) == 8 || in general_restrictions_based_on_operand_types() 874 (num_sources > 1 && type_sz(src1_type) == 8)), in general_restrictions_based_on_operand_types() 877 ERROR_IF(type_sz(dst_type) == 8 && in general_restrictions_based_on_operand_types() 878 (type_sz(src0_type) == 1 || in general_restrictions_based_on_operand_types() 879 (num_sources > 1 && type_sz(src1_type) == 1)), in general_restrictions_based_on_operand_types() 900 (type_sz(src0_type) == 8 || in general_restrictions_based_on_operand_types() 901 (num_sources > 1 && type_sz(src1_type) == 8)), in general_restrictions_based_on_operand_types() [all …]
|
D | brw_reg.h | 315 type_sz(unsigned type) in type_sz() function 423 reg.subnr = subnr * type_sz(type); in brw_reg() 588 return byte_offset(reg, delta * type_sz(reg.type)); in suboffset() 1017 unsigned scale = type_sz(reg.type) / type_sz(type); in subscript() 1021 unsigned bit_size = type_sz(type) * 8; in subscript() 1249 return type_sz(reg.type); in element_sz() 1254 return type_sz(reg.type) << (reg.vstride - 1); in element_sz() 1259 return type_sz(reg.type) << (reg.hstride - 1); in element_sz()
|
D | brw_ir_performance.cpp | 142 sx = DIV_ROUND_UP(inst->exec_size * type_sz(tx), REG_SIZE); in instruction_info() 148 !brw_reg_type_is_floating_point(tx) && type_sz(tx) == 4 && in instruction_info() 149 type_sz(inst->src[0].type) == type_sz(inst->src[1].type)) in instruction_info() 165 sx = DIV_ROUND_UP(inst->exec_size * type_sz(tx), REG_SIZE); in instruction_info() 171 !brw_reg_type_is_floating_point(tx) && type_sz(tx) == 4 && in instruction_info() 172 type_sz(inst->src[0].type) == type_sz(inst->src[1].type)) in instruction_info() 367 if (type_sz(info.tx) > 4) in instruction_desc() 392 if (type_sz(info.tx) > 4) in instruction_desc() 442 if (type_sz(info.tx) > 4) in instruction_desc() 1332 const unsigned offset = (inst->group + i) * type_sz(tx) * in accum_reg_of_channel()
|
D | brw_fs_generator.cpp | 79 const unsigned reg_width = REG_SIZE / (reg->stride * type_sz(reg->type)); in brw_reg_from_fs_reg() 99 assert(reg->stride * type_sz(reg->type) <= REG_SIZE); in brw_reg_from_fs_reg() 130 if (type_sz(reg->type) == 8) { in brw_reg_from_fs_reg() 144 type_sz(inst->dst.type) < 8) { in brw_reg_from_fs_reg() 176 type_sz(reg->type) == 8 && in brw_reg_from_fs_reg() 473 if (type_sz(reg.type) > 4 && !devinfo->has_64bit_float) { in generate_mov_indirect() 552 if (type_sz(reg.type) > 4 && in generate_mov_indirect() 611 type_sz(src.type) <= 4); in generate_shuffle() 650 assert(type_sz(group_idx.type) <= 4); in generate_shuffle() 651 if (type_sz(group_idx.type) == 4) { in generate_shuffle() [all …]
|
D | brw_vec4_copy_propagation.cpp | 148 if (type_sz(value.type) == 8 || type_sz(inst->src[arg].type) == 8) in try_constant_propagate() 336 type_sz(value.type) == 4) in try_copy_propagate() 343 if (type_sz(value.type) != type_sz(inst->src[arg].type)) in try_copy_propagate()
|
D | brw_vec4.cpp | 229 return 4 * type_sz(src[arg].type); in size_read() 232 return exec_size * type_sz(src[arg].type); in size_read() 422 type_sz(inst->src[0].type) < 8 && in opt_vector_float() 729 #define IS_64BIT(reg) (reg.file != BAD_FILE && type_sz(reg.type) == 8) in is_dep_ctrl_unsafe() 1060 if (type_sz(inst->src[0].type) != type_sz(scan_inst->src[0].type)) in opt_register_coalesce() 1082 type_sz(scan_inst->dst.type)) > 8 || in opt_register_coalesce() 1701 const unsigned size_written = type_sz(inst->dst.type); in fixup_3src_null_dest() 1755 if (type_sz(src.type) == 8) { in convert_to_hw_regs() 1801 type_sz(inst->src[i].type) < 8) { in convert_to_hw_regs() 1889 if (inst->opcode == BRW_OPCODE_SEL && type_sz(inst->dst.type) == 8) in get_lowered_simd_width() [all …]
|
D | brw_vec4_reg_allocate.cpp | 341 return type_sz(type) == 8 ? 2.25f : 1.0f; in spill_cost_for_type() 383 if (type_sz(inst->src[i].type) == 8 && inst->exec_size != 8) in evaluate_spill_costs() 390 unsigned type_size = type_sz(inst->src[i].type); in evaluate_spill_costs() 410 if (type_sz(inst->dst.type) == 8 && inst->exec_size != 8) in evaluate_spill_costs() 416 unsigned type_size = type_sz(inst->dst.type); in evaluate_spill_costs()
|
D | brw_vec4_live_variables.h | 118 const unsigned csize = DIV_ROUND_UP(type_sz(reg.type), 4); 132 const unsigned csize = DIV_ROUND_UP(type_sz(reg.type), 4);
|
D | brw_fs_scoreboard.cpp | 87 has_long_src |= type_sz(t) >= 8; in inferred_sync_pipe() 111 MIN2(type_sz(inst->src[0].type), type_sz(inst->src[1].type)) >= 4) || in inferred_exec_pipe() 113 MIN2(type_sz(inst->src[1].type), type_sz(inst->src[2].type)) >= 4)); in inferred_exec_pipe() 120 type_sz(t) >= 8) in inferred_exec_pipe() 123 !devinfo->has_64bit_float && type_sz(t) >= 8) in inferred_exec_pipe() 127 else if (type_sz(inst->dst.type) >= 8 || type_sz(t) >= 8 || in inferred_exec_pipe()
|
D | brw_fs_nir.cpp | 860 if (type_sz(op[0].type) == 2) { in emit_fsign() 880 } else if (type_sz(op[0].type) == 4) { in emit_fsign() 1008 assert(type_sz(op[i].type) > 1); in nir_emit_alu() 1088 assert(type_sz(op[0].type) < 8); /* brw_nir_lower_conversions */ in nir_emit_alu() 1125 assert(type_sz(op[0].type) < 8); /* brw_nir_lower_conversions */ in nir_emit_alu() 1130 assert(type_sz(result.type) < 8); /* brw_nir_lower_conversions */ in nir_emit_alu() 1137 assert(type_sz(op[0].type) < 8); /* brw_nir_lower_conversions */ in nir_emit_alu() 1200 assert(type_sz(result.type) < 8); /* brw_nir_lower_conversions */ in nir_emit_alu() 1458 const uint32_t bit_size = type_sz(op[0].type) * 8; in nir_emit_alu() 2515 assert(type_sz(dst.type) == 4); in emit_gs_input_load() [all …]
|
D | brw_fs_builder.h | 205 DIV_ROUND_UP(n * type_sz(type) * dispatch_width(), 514 if (dispatch_width() * type_sz(tmp.type) > 2 * REG_SIZE) { in emit_scan() 534 if (type_sz(tmp.type) <= 4) { in emit_scan() 774 inst->size_written += dispatch_width() * type_sz(src[i].type) * in LOAD_PAYLOAD()
|
D | brw_fs_validate.cpp | 53 fsv_assert(type_sz(offset(inst->src[URB_LOGICAL_SRC_DATA], bld, j).type) == 4); in validate()
|
D | brw_vec4_cse.cpp | 214 unsigned component_size = width * type_sz(entry->tmp.type); in opt_cse_local() 235 unsigned component_size = width * type_sz(inst->dst.type); in opt_cse_local()
|
D | brw_shader.cpp | 578 const unsigned size = type_sz(type); in brw_saturate_immediate() 763 assert(type_sz(type) > 1); in is_zero() 794 assert(type_sz(type) > 1); in is_one() 825 assert(type_sz(type) > 1); in is_negative_one()
|
D | brw_fs.cpp | 200 (const_offset & 0xf) / type_sz(dst.type), 1); in VARYING_PULL_CONSTANT_LOAD() 438 MIN2(type_sz(src[1].type), type_sz(src[2].type)) : in can_do_source_mods() 439 MIN2(type_sz(src[0].type), type_sz(src[1].type)); in can_do_source_mods() 442 type_sz(exec_type) >= 4 && in can_do_source_mods() 443 type_sz(exec_type) != min_type_sz) in can_do_source_mods() 553 return MAX2(width * stride, 1) * type_sz(type); in component_size() 646 (this->exec_size * type_sz(this->dst.type)) < 32 || in is_partial_write() 949 return components_read(arg) * type_sz(src[arg].type); in size_read() 2033 type_sz(inst->src[i].type); in convert_attr_sources_to_hw_regs() 4203 if (type_sz(inst->src[1].type) < 4 && type_sz(inst->src[0].type) <= 4) in lower_integer_multiplication() [all …]
|
D | brw_vec4_generator.cpp | 1967 assert(type_sz(src[0].type) == 8); in generate_code() 1968 assert(type_sz(dst.type) == 8); in generate_code() 2007 assert(type_sz(src[0].type) == 4); in generate_code() 2008 assert(type_sz(dst.type) == 8); in generate_code() 2023 assert(type_sz(src[0].type) == 8); in generate_code() 2024 assert(type_sz(dst.type) == 4); in generate_code() 2046 assert(type_sz(src[0].type) == 4); in generate_code() 2047 assert(type_sz(dst.type) == 8); in generate_code()
|
D | brw_vec4_builder.h | 178 n * DIV_ROUND_UP(type_sz(type), 4))), 316 inst->size_written = inst->exec_size * type_sz(inst->dst.type); in emit()
|
D | brw_vec4_visitor.cpp | 61 0 : this->exec_size * type_sz(dst.type)); in vec4_instruction() 1090 if (type_sz(inst->dst.type) < 8) { in get_scratch_offset() 1123 if (type_sz(orig_src.type) < 8) { in emit_scratch_read() 1159 bool is_64bit = type_sz(inst->dst.type) == 8; in emit_scratch_write() 1246 dst_reg temp = dst_reg(this, type_sz(src.type) == 8 ? in emit_resolve_reladdr()
|
D | brw_lower_logical_sends.cpp | 301 assert(sample_mask.file != BAD_FILE && type_sz(sample_mask.type) == 4); in lower_fb_write_logical_send() 761 brw_reg_type_from_bit_size(type_sz(src[i].type) * 8, in emit_load_payload_with_padding() 1692 const unsigned src_sz = type_sz(src.type); in lower_lsc_surface_logical_send() 1879 ex_mlen = src_sz * type_sz(src.type) * inst->exec_size / REG_SIZE; in lower_surface_block_logical_send() 1906 assert(type_sz(addr.type) == 8 && addr.stride == 0); in emit_a64_oword_block_header() 1956 const unsigned src_sz = type_sz(src.type); in lower_lsc_a64_logical_send() 2083 ex_mlen = src_comps * type_sz(src.type) * inst->exec_size / REG_SIZE; in lower_a64_logical_send() 2090 ex_mlen = src_comps * type_sz(src.type) * inst->exec_size / REG_SIZE; in lower_a64_logical_send() 2429 assert(type_sz(global_addr.type) == 8 && global_addr.stride == 0); in lower_btd_logical_send()
|
D | brw_fs_sel_peephole.cpp | 208 if (src1.file == IMM && type_sz(src1.type) == 8) { in opt_peephole_sel()
|