/external/libvpx/libvpx/vpx_dsp/arm/ |
D | fdct16x16_neon.c | 134 a[0] = vshrq_n_s16(vaddq_s16(a[0], one), 2); in partial_round_shift() 135 a[1] = vshrq_n_s16(vaddq_s16(a[1], one), 2); in partial_round_shift() 136 a[2] = vshrq_n_s16(vaddq_s16(a[2], one), 2); in partial_round_shift() 137 a[3] = vshrq_n_s16(vaddq_s16(a[3], one), 2); in partial_round_shift() 138 a[4] = vshrq_n_s16(vaddq_s16(a[4], one), 2); in partial_round_shift() 139 a[5] = vshrq_n_s16(vaddq_s16(a[5], one), 2); in partial_round_shift() 140 a[6] = vshrq_n_s16(vaddq_s16(a[6], one), 2); in partial_round_shift() 141 a[7] = vshrq_n_s16(vaddq_s16(a[7], one), 2); in partial_round_shift() 142 a[8] = vshrq_n_s16(vaddq_s16(a[8], one), 2); in partial_round_shift() 143 a[9] = vshrq_n_s16(vaddq_s16(a[9], one), 2); in partial_round_shift() [all …]
|
D | quantize_neon.c | 61 const int16x8_t coeff_sign = vshrq_n_s16(coeff, 15); in vpx_quantize_b_neon() 70 int16x8_t qcoeff = vshrq_n_s16(vqdmulhq_s16(rounded, quant), 1); in vpx_quantize_b_neon() 75 qcoeff = vshrq_n_s16(vqdmulhq_s16(qcoeff, quant_shift), 1); in vpx_quantize_b_neon() 111 const int16x8_t coeff_sign = vshrq_n_s16(coeff, 15); in vpx_quantize_b_neon() 120 int16x8_t qcoeff = vshrq_n_s16(vqdmulhq_s16(rounded, quant), 1); in vpx_quantize_b_neon() 125 qcoeff = vshrq_n_s16(vqdmulhq_s16(qcoeff, quant_shift), 1); in vpx_quantize_b_neon() 221 const int16x8_t coeff_sign = vshrq_n_s16(coeff, 15); in vpx_quantize_b_32x32_neon() 230 int16x8_t qcoeff = vshrq_n_s16(vqdmulhq_s16(rounded, quant), 1); in vpx_quantize_b_32x32_neon() 269 const int16x8_t coeff_sign = vshrq_n_s16(coeff, 15); in vpx_quantize_b_32x32_neon() 278 int16x8_t qcoeff = vshrq_n_s16(vqdmulhq_s16(rounded, quant), 1); in vpx_quantize_b_32x32_neon()
|
D | fwd_txfm_neon.c | 184 const int16x8_t sign_in0 = vshrq_n_s16(input_0, 15); in vpx_fdct8x8_neon() 185 const int16x8_t sign_in1 = vshrq_n_s16(input_1, 15); in vpx_fdct8x8_neon() 186 const int16x8_t sign_in2 = vshrq_n_s16(input_2, 15); in vpx_fdct8x8_neon() 187 const int16x8_t sign_in3 = vshrq_n_s16(input_3, 15); in vpx_fdct8x8_neon() 188 const int16x8_t sign_in4 = vshrq_n_s16(input_4, 15); in vpx_fdct8x8_neon() 189 const int16x8_t sign_in5 = vshrq_n_s16(input_5, 15); in vpx_fdct8x8_neon() 190 const int16x8_t sign_in6 = vshrq_n_s16(input_6, 15); in vpx_fdct8x8_neon() 191 const int16x8_t sign_in7 = vshrq_n_s16(input_7, 15); in vpx_fdct8x8_neon()
|
D | fdct_neon.c | 86 out_01 = vshrq_n_s16(vaddq_s16(out_01, one), 2); in vpx_fdct4x4_neon() 87 out_23 = vshrq_n_s16(vaddq_s16(out_23, one), 2); in vpx_fdct4x4_neon()
|
D | highbd_loopfilter_neon.c | 262 filter1 = vshrq_n_s16(t, 3); in filter4() 265 filter2 = vshrq_n_s16(t, 3); in filter4()
|
/external/libvpx/libvpx/vp8/encoder/arm/neon/ |
D | fastquantizeb_neon.c | 38 sz0 = vshrq_n_s16(z0, 15); in vp8_fast_quantize_b_neon() 39 sz1 = vshrq_n_s16(z1, 15); in vp8_fast_quantize_b_neon() 54 y0 = vshrq_n_s16(y0, 1); in vp8_fast_quantize_b_neon() 55 y1 = vshrq_n_s16(y1, 1); in vp8_fast_quantize_b_neon()
|
/external/libaom/libaom/aom_dsp/arm/ |
D | fwd_txfm_neon.c | 179 const int16x8_t sign_in0 = vshrq_n_s16(input_0, 15); in aom_fdct8x8_neon() 180 const int16x8_t sign_in1 = vshrq_n_s16(input_1, 15); in aom_fdct8x8_neon() 181 const int16x8_t sign_in2 = vshrq_n_s16(input_2, 15); in aom_fdct8x8_neon() 182 const int16x8_t sign_in3 = vshrq_n_s16(input_3, 15); in aom_fdct8x8_neon() 183 const int16x8_t sign_in4 = vshrq_n_s16(input_4, 15); in aom_fdct8x8_neon() 184 const int16x8_t sign_in5 = vshrq_n_s16(input_5, 15); in aom_fdct8x8_neon() 185 const int16x8_t sign_in6 = vshrq_n_s16(input_6, 15); in aom_fdct8x8_neon() 186 const int16x8_t sign_in7 = vshrq_n_s16(input_7, 15); in aom_fdct8x8_neon()
|
/external/libvpx/libvpx/vp9/encoder/arm/neon/ |
D | vp9_quantize_neon.c | 57 const int16x8_t v_coeff_sign = vshrq_n_s16(v_coeff, 15); in vp9_quantize_fp_neon() 82 const int16x8_t v_coeff_sign = vshrq_n_s16(v_coeff, 15); in vp9_quantize_fp_neon() 138 const int16x8_t dequant_thresh = vshrq_n_s16(vld1q_s16(dequant_ptr), 2); in vp9_quantize_fp_32x32_neon() 144 const int16x8_t coeff_sign = vshrq_n_s16(coeff, 15); in vp9_quantize_fp_32x32_neon() 191 vshrq_n_s16(vmovq_n_s16(dequant_ptr[1]), 2); in vp9_quantize_fp_32x32_neon() 198 const int16x8_t coeff_sign = vshrq_n_s16(coeff, 15); in vp9_quantize_fp_32x32_neon()
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | idct_blk_neon.c | 144 q8 = vshrq_n_s16(q8, 1); in idct_dequant_full_2x_neon() 145 q9 = vshrq_n_s16(q9, 1); in idct_dequant_full_2x_neon() 174 q10 = vshrq_n_s16(q10, 1); in idct_dequant_full_2x_neon() 175 q11 = vshrq_n_s16(q11, 1); in idct_dequant_full_2x_neon()
|
D | iwalsh_neon.c | 62 q0s16 = vshrq_n_s16(q0s16, 3); in vp8_short_inv_walsh4x4_neon() 63 q1s16 = vshrq_n_s16(q1s16, 3); in vp8_short_inv_walsh4x4_neon()
|
D | shortidct4x4llm_neon.c | 49 q4s16 = vshrq_n_s16(q4s16, 1); in vp8_short_idct4x4llm_neon() 78 q4s16 = vshrq_n_s16(q4s16, 1); in vp8_short_idct4x4llm_neon()
|
D | dequant_idct_neon.c | 69 q4 = vshrq_n_s16(q4, 1); in vp8_dequant_idct_add_neon() 97 q4 = vshrq_n_s16(q4, 1); in vp8_dequant_idct_add_neon()
|
/external/libaom/libaom/av1/encoder/arm/neon/ |
D | quantize_neon.c | 56 const int16x8_t v_coeff_sign = vshrq_n_s16(v_coeff, 15); in av1_quantize_fp_neon() 81 const int16x8_t v_coeff_sign = vshrq_n_s16(v_coeff, 15); in av1_quantize_fp_neon()
|
/external/gemmlowp/fixedpoint/ |
D | fixedpoint_neon.h | 280 const int16x8_t fixup = vshrq_n_s16(vandq_s16(x, shift_vec), 15); 307 const int16x8_t fixup = vshrq_n_s16(x, 15);
|
/external/libhevc/encoder/arm/ |
D | ihevce_had_compute_neon.c | 716 a[b16][0][i] = vshrq_n_s16(a[b16][0][i], 2); in ihevce_HAD_32x32_8bit_neon() 717 a[b16][1][i] = vshrq_n_s16(a[b16][1][i], 2); in ihevce_HAD_32x32_8bit_neon() 718 a[b16][2][i] = vshrq_n_s16(a[b16][2][i], 2); in ihevce_HAD_32x32_8bit_neon() 719 a[b16][3][i] = vshrq_n_s16(a[b16][3][i], 2); in ihevce_HAD_32x32_8bit_neon() 1146 a[b16][b8][i] = vshrq_n_s16(a[b16][b8][i], 2); in ihevce_compute_32x32HAD_using_16x16_neon()
|
/external/webp/src/dsp/ |
D | yuv_neon.c | 245 const int16x8_t c0 = vshrq_n_s16(vaddq_s16(a0b1_2, a0a1b0b1), 3); in SharpYUVFilterRow_NEON() 246 const int16x8_t c1 = vshrq_n_s16(vaddq_s16(a1b0_2, a0a1b0b1), 3); in SharpYUVFilterRow_NEON()
|
D | enc_neon.c | 832 const int16x8_t sign = vshrq_n_s16(a, 15); // sign in Quantize_NEON()
|
/external/libhevc/common/arm/ |
D | ihevc_intra_pred_filters_neon_intr.c | 1259 shift_val = vshrq_n_s16(vreinterpretq_s16_u16(sub_val), 1); in ihevc_intra_pred_luma_horz_neonintr() 1319 shift_res = vshrq_n_s16(vreinterpretq_s16_u16(sub_res), 1); in ihevc_intra_pred_luma_horz_neonintr() 1497 shift_val1 = vshrq_n_s16(vreinterpretq_s16_u16(sub_val), 1); in ihevc_intra_pred_luma_ver_neonintr() 1594 subsh_val = vshrq_n_s16(vreinterpretq_s16_u16(sub_val), 1); in ihevc_intra_pred_luma_ver_neonintr()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | quantization_utils.h | 419 const int16x8_t val_sign = vshrq_n_s16(val, 15); in Divide16x8PowRound() 446 const int16x8_t val_sign = vshrq_n_s16(val, 15); in Divide16x8Pow()
|
/external/libaom/libaom/av1/common/arm/ |
D | jnt_convolve_neon.c | 744 filter_x_coef = vshrq_n_s16(filter_x_coef, 1); in av1_dist_wtd_convolve_2d_neon() 916 filter_x_coef = vshrq_n_s16(filter_x_coef, 1); in av1_dist_wtd_convolve_x_neon() 1380 filter_y_coef = vshrq_n_s16(filter_y_coef, 1); in av1_dist_wtd_convolve_y_neon()
|
D | cfl_neon.c | 403 const int16x8_t mask = vshrq_n_s16(b, 15); in vsignq_s16()
|
/external/boringssl/src/crypto/hrss/ |
D | hrss.c | 252 a = (vec_t)vshrq_n_s16(((int16x8_t)a) << 15, 15); in vec_broadcast_bit() 257 a = (vec_t)vshrq_n_s16((int16x8_t)a, 15); in vec_broadcast_bit15()
|
/external/libaom/libaom/aom_dsp/simd/ |
D | v128_intrinsics_arm.h | 824 return vreinterpretq_s64_s16(vshrq_n_s16(vreinterpretq_s16_s64(a), c)); in v128_shr_n_s16()
|
/external/skia/src/core/ |
D | SkBlitter_ARGB32.cpp | 453 dst_wide += vshrq_n_s16(src_wide, 5); in blend_32_neon()
|
/external/skqp/src/core/ |
D | SkBlitter_ARGB32.cpp | 453 dst_wide += vshrq_n_s16(src_wide, 5); in blend_32_neon()
|