• Home
  • Raw
  • Download

Lines Matching refs:nb

107             vec_src[j] = nir_fdot(&b->nb, src0_transpose->elems[j]->def,  in matrix_multiply()
110 dest->elems[i]->def = nir_vec(&b->nb, vec_src, src0_rows); in matrix_multiply()
121 nir_fmul(&b->nb, src0->elems[src0_columns - 1]->def, in matrix_multiply()
122 nir_channel(&b->nb, src1->elems[i]->def, src0_columns - 1)); in matrix_multiply()
125 nir_ffma(&b->nb, src0->elems[j]->def, in matrix_multiply()
126 nir_channel(&b->nb, src1->elems[i]->def, j), in matrix_multiply()
148 dest->elems[i]->def = nir_imul(&b->nb, mat->elems[i]->def, scalar); in mat_times_scalar()
150 dest->elems[i]->def = nir_fmul(&b->nb, mat->elems[i]->def, scalar); in mat_times_scalar()
165 dest->elems[i]->def = nir_fneg(&b->nb, src0->elems[i]->def); in vtn_handle_matrix_alu()
174 nir_fadd(&b->nb, src0->elems[i]->def, src1->elems[i]->def); in vtn_handle_matrix_alu()
183 nir_fsub(&b->nb, src0->elems[i]->def, src1->elems[i]->def); in vtn_handle_matrix_alu()
389 b->nb.exact = true; in handle_no_contraction()
487 b->nb.exact = b->exact; in vtn_handle_alu()
500 dest->def = nir_bany(&b->nb, src[0]); in vtn_handle_alu()
504 dest->def = nir_ball(&b->nb, src[0]); in vtn_handle_alu()
510 nir_fmul(&b->nb, src[0], nir_channel(&b->nb, src[1], i)); in vtn_handle_alu()
516 dest->def = nir_fdot(&b->nb, src[0], src[1]); in vtn_handle_alu()
521 dest->elems[0]->def = nir_iadd(&b->nb, src[0], src[1]); in vtn_handle_alu()
522 dest->elems[1]->def = nir_uadd_carry(&b->nb, src[0], src[1]); in vtn_handle_alu()
527 dest->elems[0]->def = nir_isub(&b->nb, src[0], src[1]); in vtn_handle_alu()
528 dest->elems[1]->def = nir_usub_borrow(&b->nb, src[0], src[1]); in vtn_handle_alu()
533 nir_ssa_def *umul = nir_umul_2x32_64(&b->nb, src[0], src[1]); in vtn_handle_alu()
534 dest->elems[0]->def = nir_unpack_64_2x32_split_x(&b->nb, umul); in vtn_handle_alu()
535 dest->elems[1]->def = nir_unpack_64_2x32_split_y(&b->nb, umul); in vtn_handle_alu()
541 nir_ssa_def *smul = nir_imul_2x32_64(&b->nb, src[0], src[1]); in vtn_handle_alu()
542 dest->elems[0]->def = nir_unpack_64_2x32_split_x(&b->nb, smul); in vtn_handle_alu()
543 dest->elems[1]->def = nir_unpack_64_2x32_split_y(&b->nb, smul); in vtn_handle_alu()
548 dest->def = nir_fadd(&b->nb, in vtn_handle_alu()
549 nir_fabs(&b->nb, nir_fddx(&b->nb, src[0])), in vtn_handle_alu()
550 nir_fabs(&b->nb, nir_fddy(&b->nb, src[0]))); in vtn_handle_alu()
553 dest->def = nir_fadd(&b->nb, in vtn_handle_alu()
554 nir_fabs(&b->nb, nir_fddx_fine(&b->nb, src[0])), in vtn_handle_alu()
555 nir_fabs(&b->nb, nir_fddy_fine(&b->nb, src[0]))); in vtn_handle_alu()
558 dest->def = nir_fadd(&b->nb, in vtn_handle_alu()
559 nir_fabs(&b->nb, nir_fddx_coarse(&b->nb, src[0])), in vtn_handle_alu()
560 nir_fabs(&b->nb, nir_fddy_coarse(&b->nb, src[0]))); in vtn_handle_alu()
565 dest->def = nir_fmul(&b->nb, src[0], src[1]); in vtn_handle_alu()
569 const bool save_exact = b->nb.exact; in vtn_handle_alu()
571 b->nb.exact = true; in vtn_handle_alu()
572 dest->def = nir_fneu(&b->nb, src[0], src[0]); in vtn_handle_alu()
573 b->nb.exact = save_exact; in vtn_handle_alu()
578 const bool save_exact = b->nb.exact; in vtn_handle_alu()
580 b->nb.exact = true; in vtn_handle_alu()
581 dest->def = nir_iand(&b->nb, nir_feq(&b->nb, src[0], src[0]), in vtn_handle_alu()
582 nir_feq(&b->nb, src[1], src[1])); in vtn_handle_alu()
583 b->nb.exact = save_exact; in vtn_handle_alu()
588 const bool save_exact = b->nb.exact; in vtn_handle_alu()
590 b->nb.exact = true; in vtn_handle_alu()
591 dest->def = nir_ior(&b->nb, nir_fneu(&b->nb, src[0], src[0]), in vtn_handle_alu()
592 nir_fneu(&b->nb, src[1], src[1])); in vtn_handle_alu()
593 b->nb.exact = save_exact; in vtn_handle_alu()
598 nir_ssa_def *inf = nir_imm_floatN_t(&b->nb, INFINITY, src[0]->bit_size); in vtn_handle_alu()
599 dest->def = nir_ieq(&b->nb, nir_fabs(&b->nb, src[0]), inf); in vtn_handle_alu()
604 const bool save_exact = b->nb.exact; in vtn_handle_alu()
606 b->nb.exact = true; in vtn_handle_alu()
616 nir_ior(&b->nb, in vtn_handle_alu()
617 nir_feq(&b->nb, src[0], src[1]), in vtn_handle_alu()
618 nir_ior(&b->nb, in vtn_handle_alu()
619 nir_fneu(&b->nb, src[0], src[0]), in vtn_handle_alu()
620 nir_fneu(&b->nb, src[1], src[1]))); in vtn_handle_alu()
622 b->nb.exact = save_exact; in vtn_handle_alu()
644 const bool save_exact = b->nb.exact; in vtn_handle_alu()
646 b->nb.exact = true; in vtn_handle_alu()
656 nir_inot(&b->nb, in vtn_handle_alu()
657 nir_build_alu(&b->nb, op, src[0], src[1], NULL, NULL)); in vtn_handle_alu()
659 b->nb.exact = save_exact; in vtn_handle_alu()
669 const bool save_exact = b->nb.exact; in vtn_handle_alu()
671 b->nb.exact = true; in vtn_handle_alu()
681 nir_iand(&b->nb, in vtn_handle_alu()
682 nir_fneu(&b->nb, src[0], src[1]), in vtn_handle_alu()
683 nir_iand(&b->nb, in vtn_handle_alu()
684 nir_feq(&b->nb, src[0], src[0]), in vtn_handle_alu()
685 nir_feq(&b->nb, src[1], src[1]))); in vtn_handle_alu()
687 b->nb.exact = save_exact; in vtn_handle_alu()
718 dest->def = nir_build_alu(&b->nb, op, src[0], NULL, NULL, NULL); in vtn_handle_alu()
720 dest->def = nir_convert_alu_types(&b->nb, dst_bit_size, src[0], in vtn_handle_alu()
731 dest->def = nir_build_alu(&b->nb, op, src[0], NULL, NULL, NULL); in vtn_handle_alu()
767 src[i] = nir_u2u32(&b->nb, src[i]); in vtn_handle_alu()
770 dest->def = nir_build_alu(&b->nb, op, src[0], src[1], src[2], src[3]); in vtn_handle_alu()
775 dest->def = nir_i2b(&b->nb, in vtn_handle_alu()
776 nir_ushr(&b->nb, src[0], nir_imm_int(&b->nb, src[0]->bit_size - 1))); in vtn_handle_alu()
780 dest->def = nir_umin(&b->nb, in vtn_handle_alu()
781 nir_find_lsb(&b->nb, src[0]), in vtn_handle_alu()
782 nir_imm_int(&b->nb, 32u)); in vtn_handle_alu()
789 dest->def = nir_u2u(&b->nb, nir_bit_count(&b->nb, src[0]), glsl_get_bit_size(dest_type)); in vtn_handle_alu()
821 src[1] = nir_u2u32(&b->nb, src[1]); in vtn_handle_alu()
827 const bool save_exact = b->nb.exact; in vtn_handle_alu()
830 b->nb.exact = true; in vtn_handle_alu()
832 dest->def = nir_build_alu(&b->nb, op, src[0], src[1], src[2], src[3]); in vtn_handle_alu()
834 b->nb.exact = save_exact; in vtn_handle_alu()
856 b->nb.exact = b->exact; in vtn_handle_alu()
929 src[0] = nir_pack_32_4x8(&b->nb, src[0]); in vtn_handle_integer_dot()
930 src[1] = nir_pack_32_4x8(&b->nb, src[1]); in vtn_handle_integer_dot()
936 src[0] = nir_pack_32_2x16(&b->nb, src[0]); in vtn_handle_integer_dot()
937 src[1] = nir_pack_32_2x16(&b->nb, src[1]); in vtn_handle_integer_dot()
1011 nir_build_alu(&b->nb, src0_conversion_op, in vtn_handle_integer_dot()
1012 nir_channel(&b->nb, src[0], i), NULL, NULL, NULL); in vtn_handle_integer_dot()
1015 nir_build_alu(&b->nb, src1_conversion_op, in vtn_handle_integer_dot()
1016 nir_channel(&b->nb, src[1], i), NULL, NULL, NULL); in vtn_handle_integer_dot()
1018 nir_ssa_def *const mul_result = nir_imul(&b->nb, src0, src1); in vtn_handle_integer_dot()
1020 dest = (i == 0) ? mul_result : nir_iadd(&b->nb, dest, mul_result); in vtn_handle_integer_dot()
1041 ? nir_uadd_sat(&b->nb, dest, src[2]) in vtn_handle_integer_dot()
1042 : nir_iadd_sat(&b->nb, dest, src[2]); in vtn_handle_integer_dot()
1048 nir_ssa_def *const zero = nir_imm_zero(&b->nb, 1, 32); in vtn_handle_integer_dot()
1055 dest = nir_sdot_2x16_iadd(&b->nb, src[0], src[1], zero); in vtn_handle_integer_dot()
1058 dest = nir_udot_2x16_uadd(&b->nb, src[0], src[1], zero); in vtn_handle_integer_dot()
1062 dest = nir_sdot_2x16_iadd_sat(&b->nb, src[0], src[1], src[2]); in vtn_handle_integer_dot()
1064 dest = nir_sdot_2x16_iadd(&b->nb, src[0], src[1], zero); in vtn_handle_integer_dot()
1068 dest = nir_udot_2x16_uadd_sat(&b->nb, src[0], src[1], src[2]); in vtn_handle_integer_dot()
1070 dest = nir_udot_2x16_uadd(&b->nb, src[0], src[1], zero); in vtn_handle_integer_dot()
1078 dest = nir_sdot_4x8_iadd(&b->nb, src[0], src[1], zero); in vtn_handle_integer_dot()
1081 dest = nir_udot_4x8_uadd(&b->nb, src[0], src[1], zero); in vtn_handle_integer_dot()
1084 dest = nir_sudot_4x8_iadd(&b->nb, src[0], src[1], zero); in vtn_handle_integer_dot()
1088 dest = nir_sdot_4x8_iadd_sat(&b->nb, src[0], src[1], src[2]); in vtn_handle_integer_dot()
1090 dest = nir_sdot_4x8_iadd(&b->nb, src[0], src[1], zero); in vtn_handle_integer_dot()
1094 dest = nir_udot_4x8_uadd_sat(&b->nb, src[0], src[1], src[2]); in vtn_handle_integer_dot()
1096 dest = nir_udot_4x8_uadd(&b->nb, src[0], src[1], zero); in vtn_handle_integer_dot()
1100 dest = nir_sudot_4x8_iadd_sat(&b->nb, src[0], src[1], src[2]); in vtn_handle_integer_dot()
1102 dest = nir_sudot_4x8_iadd(&b->nb, src[0], src[1], zero); in vtn_handle_integer_dot()
1128 ? nir_iadd_sat(&b->nb, nir_i2i(&b->nb, dest, dest_size), src[2]) in vtn_handle_integer_dot()
1129 : nir_uadd_sat(&b->nb, nir_u2u(&b->nb, dest, dest_size), src[2]); in vtn_handle_integer_dot()
1132 ? nir_i2i(&b->nb, dest, dest_size) in vtn_handle_integer_dot()
1133 : nir_u2u(&b->nb, dest, dest_size); in vtn_handle_integer_dot()
1140 b->nb.exact = b->exact; in vtn_handle_integer_dot()
1174 nir_bitcast_vector(&b->nb, src, glsl_get_bit_size(type->type)); in vtn_handle_bitcast()