/external/libaom/libaom/av1/common/x86/ |
D | highbd_jnt_convolve_sse4.c | 50 const __m128i offset_const = _mm_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_y_sse4_1() local 111 __m128i res_unsigned_lo_0 = _mm_add_epi32(res_a_round0, offset_const); in av1_highbd_dist_wtd_convolve_y_sse4_1() 112 __m128i res_unsigned_lo_1 = _mm_add_epi32(res_a_round1, offset_const); in av1_highbd_dist_wtd_convolve_y_sse4_1() 132 highbd_convolve_rounding_sse2(&comp_avg_res_0, &offset_const, in av1_highbd_dist_wtd_convolve_y_sse4_1() 135 highbd_convolve_rounding_sse2(&comp_avg_res_1, &offset_const, in av1_highbd_dist_wtd_convolve_y_sse4_1() 175 __m128i res_unsigned_hi_0 = _mm_add_epi32(res_b_round0, offset_const); in av1_highbd_dist_wtd_convolve_y_sse4_1() 176 __m128i res_unsigned_hi_1 = _mm_add_epi32(res_b_round1, offset_const); in av1_highbd_dist_wtd_convolve_y_sse4_1() 203 highbd_convolve_rounding_sse2(&comp_avg_res_lo_0, &offset_const, in av1_highbd_dist_wtd_convolve_y_sse4_1() 206 highbd_convolve_rounding_sse2(&comp_avg_res_lo_1, &offset_const, in av1_highbd_dist_wtd_convolve_y_sse4_1() 209 highbd_convolve_rounding_sse2(&comp_avg_res_hi_0, &offset_const, in av1_highbd_dist_wtd_convolve_y_sse4_1() [all …]
|
D | highbd_jnt_convolve_avx2.c | 52 const __m256i offset_const = _mm256_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_2d_copy_avx2() local 79 _mm256_add_epi32(res_32b_lo, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_avx2() 87 _mm256_add_epi32(res_32b_hi, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_avx2() 94 &comp_avg_res_lo, &offset_const, &rounding_const, rounding_shift); in av1_highbd_dist_wtd_convolve_2d_copy_avx2() 96 &comp_avg_res_hi, &offset_const, &rounding_const, rounding_shift); in av1_highbd_dist_wtd_convolve_2d_copy_avx2() 138 _mm256_add_epi32(res_32b, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_avx2() 145 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_highbd_dist_wtd_convolve_2d_copy_avx2() 183 _mm256_add_epi32(res_32b_lo, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_avx2() 191 _mm256_add_epi32(res_32b_hi, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_avx2() 198 highbd_convolve_rounding(&comp_avg_res_lo, &offset_const, in av1_highbd_dist_wtd_convolve_2d_copy_avx2() [all …]
|
D | highbd_convolve_2d_sse4.c | 51 const __m128i offset_const = _mm_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() local 76 _mm_add_epi32(res_32b_lo, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 84 _mm_add_epi32(res_32b_hi, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 91 &comp_avg_res_lo, &offset_const, &rounding_const, rounding_shift); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 93 &comp_avg_res_hi, &offset_const, &rounding_const, rounding_shift); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 130 const __m128i res_unsigned_lo = _mm_add_epi32(res_32b, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 134 _mm_add_epi32(res_32b_hi, offset_const); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 142 &comp_avg_res_lo, &offset_const, &rounding_const, rounding_shift); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 144 &comp_avg_res_hi, &offset_const, &rounding_const, rounding_shift); in av1_highbd_dist_wtd_convolve_2d_copy_sse4_1() 197 const __m128i offset_const = _mm_set1_epi32(offset); in av1_highbd_dist_wtd_convolve_2d_sse4_1() local [all …]
|
D | jnt_convolve_avx2.c | 55 const __m256i offset_const = _mm256_set1_epi16(offset); in av1_dist_wtd_convolve_x_avx2() local 96 const __m256i res_unsigned = _mm256_add_epi16(res, offset_const); in av1_dist_wtd_convolve_x_avx2() 106 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_x_avx2() 151 const __m256i res_unsigned = _mm256_add_epi16(res, offset_const); in av1_dist_wtd_convolve_x_avx2() 161 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_x_avx2() 211 const __m256i offset_const = _mm256_set1_epi16(offset); in av1_dist_wtd_convolve_y_avx2() local 303 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_y_avx2() 368 convolve_rounding(&comp_avg_res_lo, &offset_const, in av1_dist_wtd_convolve_y_avx2() 372 convolve_rounding(&comp_avg_res_hi, &offset_const, in av1_dist_wtd_convolve_y_avx2() 482 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_y_avx2() [all …]
|
D | jnt_convolve_sse2.c | 44 const __m128i offset_const = _mm_set1_epi16(offset); in av1_dist_wtd_convolve_x_sse2() local 73 const __m128i res_unsigned = _mm_add_epi16(res_16b, offset_const); in av1_dist_wtd_convolve_x_sse2() 83 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_x_sse2() 129 const __m128i res_unsigned = _mm_add_epi16(res_16b, offset_const); in av1_dist_wtd_convolve_x_sse2() 140 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_x_sse2() 174 const __m128i offset_const = _mm_set1_epi16(offset); in av1_dist_wtd_convolve_y_sse2() local 221 __m128i res_unsigned = _mm_add_epi16(res_16b, offset_const); in av1_dist_wtd_convolve_y_sse2() 231 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_y_sse2() 250 res_unsigned = _mm_add_epi16(res_16b, offset_const); in av1_dist_wtd_convolve_y_sse2() 260 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_y_sse2() [all …]
|
D | convolve_2d_sse2.c | 387 const __m128i offset_const = _mm_set1_epi16(offset); in av1_dist_wtd_convolve_2d_copy_sse2() local 403 const __m128i res_unsigned_lo = _mm_add_epi16(res_lo, offset_const); in av1_dist_wtd_convolve_2d_copy_sse2() 406 const __m128i res_unsigned_hi = _mm_add_epi16(res_hi, offset_const); in av1_dist_wtd_convolve_2d_copy_sse2() 417 &comp_avg_res_lo, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_2d_copy_sse2() 423 &comp_avg_res_hi, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_2d_copy_sse2() 445 const __m128i res_unsigned = _mm_add_epi16(res, offset_const); in av1_dist_wtd_convolve_2d_copy_sse2() 454 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_2d_copy_sse2()
|
D | jnt_convolve_ssse3.c | 50 const __m128i offset_const = _mm_set1_epi16(offset); in av1_dist_wtd_convolve_2d_ssse3() local 205 const __m128i res_unsigned = _mm_add_epi16(res_16b, offset_const); in av1_dist_wtd_convolve_2d_ssse3() 216 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); in av1_dist_wtd_convolve_2d_ssse3()
|
/external/mesa3d/src/compiler/nir/ |
D | nir_lower_atomics.c | 105 nir_load_const_instr *offset_const = in lower_instr() local 107 offset_const->value.u32[0] = instr->variables[0]->var->data.offset; in lower_instr() 109 nir_instr_insert_before(&instr->instr, &offset_const->instr); in lower_instr() 111 nir_ssa_def *offset_def = &offset_const->def; in lower_instr() 121 offset_const->value.u32[0] += deref_array->base_offset * in lower_instr()
|
/external/libaom/libaom/aom_dsp/x86/ |
D | convolve_sse2.h | 102 const __m128i *const offset_const, in convolve_rounding() argument 105 const __m128i res_signed = _mm_sub_epi16(*res_unsigned, *offset_const); in convolve_rounding() 112 const __m128i *const res_unsigned, const __m128i *const offset_const, in highbd_convolve_rounding_sse2() argument 114 const __m128i res_signed = _mm_sub_epi32(*res_unsigned, *offset_const); in highbd_convolve_rounding_sse2()
|
D | convolve_avx2.h | 179 const __m256i res_unsigned = _mm256_add_epi16(res_16b, offset_const); \ 188 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); \ 207 const __m256i res_unsigned = _mm256_add_epi16(res_16b, offset_const); \ 217 &comp_avg_res, &offset_const, &rounding_const, rounding_shift); \ 401 const __m256i *const offset_const, in convolve_rounding() argument 404 const __m256i res_signed = _mm256_sub_epi16(*res_unsigned, *offset_const); in convolve_rounding() 429 const __m256i *const res_unsigned, const __m256i *const offset_const, in highbd_convolve_rounding() argument 431 const __m256i res_signed = _mm256_sub_epi32(*res_unsigned, *offset_const); in highbd_convolve_rounding()
|
/external/mesa3d/src/intel/compiler/ |
D | brw_nir_analyze_ubo_ranges.c | 134 nir_const_value *offset_const = nir_src_as_const_value(intrin->src[1]); in analyze_ubos_block() local 136 if (block_const && offset_const) { in analyze_ubos_block() 138 const int offset = offset_const->u32[0] / 32; in analyze_ubos_block()
|
D | brw_fs_nir.cpp | 2095 nir_const_value *offset_const = nir_src_as_const_value(offset_src); in emit_gs_input_load() local 2102 offset_const != NULL && vertex_const != NULL && in emit_gs_input_load() 2103 4 * (base_offset + offset_const->u32[0]) < push_reg_count) { in emit_gs_input_load() 2104 int imm_offset = (base_offset + offset_const->u32[0]) * 4 + in emit_gs_input_load() 2213 if (offset_const) { in emit_gs_input_load() 2231 inst->offset = base_offset + offset_const->u32[0]; in emit_gs_input_load() 2269 if(offset_const) { in emit_gs_input_load()
|
/external/libaom/libaom/av1/common/arm/ |
D | convolve_neon.c | 125 const int32x4_t round_shift_vec, const int32x4_t offset_const, in convolve8_vert_4x4_s32() argument 140 sum0 = vaddq_s32(sum0, offset_const); in convolve8_vert_4x4_s32() 154 const int32x4_t round_shift_vec, const int32x4_t offset_const, in convolve8_vert_8x4_s32() argument 178 sum0 = vaddq_s32(sum0, offset_const); in convolve8_vert_8x4_s32() 179 sum1 = vaddq_s32(sum1, offset_const); in convolve8_vert_8x4_s32() 1155 const int32x4_t offset_const = vdupq_n_s32(1 << offset_bits); in av1_convolve_2d_sr_neon() local 1204 round_shift_vec, offset_const, in av1_convolve_2d_sr_neon() 1207 round_shift_vec, offset_const, in av1_convolve_2d_sr_neon() 1210 round_shift_vec, offset_const, in av1_convolve_2d_sr_neon() 1213 round_shift_vec, offset_const, in av1_convolve_2d_sr_neon() [all …]
|
D | convolve_neon.h | 206 const int32x4_t round_shift_vec, const int32x4_t offset_const) { in convolve8_4x4_s32() argument 220 sum0 = vaddq_s32(sum0, offset_const); in convolve8_4x4_s32()
|
D | jnt_convolve_neon.c | 581 const int32x4_t offset_const = vdupq_n_s32(1 << offset); in dist_wtd_convolve_2d_vert_neon() local 638 round_shift_vec, offset_const); in dist_wtd_convolve_2d_vert_neon() 640 round_shift_vec, offset_const); in dist_wtd_convolve_2d_vert_neon() 642 round_shift_vec, offset_const); in dist_wtd_convolve_2d_vert_neon() 644 round_shift_vec, offset_const); in dist_wtd_convolve_2d_vert_neon() 683 round_shift_vec, offset_const); in dist_wtd_convolve_2d_vert_neon()
|