/external/libhevc/encoder/arm/ |
D | ihevce_itrans_recon_neon.c | 71 a3 = vqmovun_s16(a1); in ihevce_itrans_recon_dc_4x4_luma_neon() 72 a4 = vqmovun_s16(a2); in ihevce_itrans_recon_dc_4x4_luma_neon() 99 a2 = vqmovun_s16(vaddq_s16(a0, a1)); in ihevce_itrans_recon_dc_4x4_chroma_neon() 118 a0 = (dc_value >= 0) ? vqmovun_s16(vdupq_n_s16(dc_value)) in ihevce_itrans_recon_dc_8x8_luma_neon() 119 : vqmovun_s16(vabsq_s16(vdupq_n_s16(dc_value))); in ihevce_itrans_recon_dc_8x8_luma_neon() 146 a1 = (dc_value >= 0) ? vqmovun_s16(vdupq_n_s16(dc_value)) in ihevce_itrans_recon_dc_8x8_chroma_neon() 147 : vqmovun_s16(vabsq_s16(vdupq_n_s16(dc_value))); in ihevce_itrans_recon_dc_8x8_chroma_neon() 176 a0 = (dc_value >= 0) ? vqmovun_s16(vdupq_n_s16(dc_value)) in ihevce_itrans_recon_dc_16x16_luma_neon() 177 : vqmovun_s16(vabsq_s16(vdupq_n_s16(dc_value))); in ihevce_itrans_recon_dc_16x16_luma_neon() 201 a0 = (dc_value >= 0) ? vqmovun_s16(vdupq_n_s16(dc_value)) in ihevce_itrans_recon_dc_16x16_chroma_neon() [all …]
|
D | ihevce_hme_utils_neon.c | 217 vst1_u8((pu1_dst + 0 * dst_stride), vqmovun_s16(src4_8x16b)); in ihevce_get_wt_inp_4x8_neon() 218 vst1_u8((pu1_dst + 1 * dst_stride), vqmovun_s16(src5_8x16b)); in ihevce_get_wt_inp_4x8_neon() 219 vst1_u8((pu1_dst + 2 * dst_stride), vqmovun_s16(src6_8x16b)); in ihevce_get_wt_inp_4x8_neon() 220 vst1_u8((pu1_dst + 3 * dst_stride), vqmovun_s16(src7_8x16b)); in ihevce_get_wt_inp_4x8_neon() 640 vst1_u8((pu1_dst + 0 * dst_stride), vqmovun_s16(src4_8x16b)); in hme_get_wt_inp_ctb_neon() 641 vst1_u8((pu1_dst + 1 * dst_stride), vqmovun_s16(src5_8x16b)); in hme_get_wt_inp_ctb_neon() 642 vst1_u8((pu1_dst + 2 * dst_stride), vqmovun_s16(src6_8x16b)); in hme_get_wt_inp_ctb_neon() 643 vst1_u8((pu1_dst + 3 * dst_stride), vqmovun_s16(src7_8x16b)); in hme_get_wt_inp_ctb_neon()
|
D | ihevce_scale_by_2_neon.c | 103 c = vqmovun_s16(sum); in ihevce_horz_scale_neon_w16() 177 vst1_u8(pu1_dst_tmp, vqmovun_s16(sum)); in ihevce_vert_scale_neon_w16() 195 vst1_u8(pu1_dst_tmp + 8, vqmovun_s16(sum)); in ihevce_vert_scale_neon_w16()
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | distance_weighted_blend_neon.cc | 74 const uint8x8_t result0 = vqmovun_s16(res0); in DistanceWeightedBlendSmall_NEON() 75 const uint8x8_t result1 = vqmovun_s16(res1); in DistanceWeightedBlendSmall_NEON() 118 vcombine_u8(vqmovun_s16(res_lo), vqmovun_s16(res_hi)); in DistanceWeightedBlendLarge_NEON()
|
D | warp_neon.cc | 290 vst1_u8(reinterpret_cast<uint8_t*>(dst_row), vqmovun_s16(sum)); in Warp_NEON() 317 vst1_u8(reinterpret_cast<uint8_t*>(dst_row), vqmovun_s16(sum)); in Warp_NEON() 419 vst1_u8(reinterpret_cast<uint8_t*>(dst_row), vqmovun_s16(sum)); in Warp_NEON()
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | idct_blk_neon.c | 45 d2u8 = vqmovun_s16(vreinterpretq_s16_u16(q1u16)); in idct_dequant_0_2x_neon() 46 d4u8 = vqmovun_s16(vreinterpretq_s16_u16(q2u16)); in idct_dequant_0_2x_neon() 208 d28 = vreinterpret_s32_u8(vqmovun_s16(q4)); in idct_dequant_full_2x_neon() 209 d29 = vreinterpret_s32_u8(vqmovun_s16(q5)); in idct_dequant_full_2x_neon() 210 d30 = vreinterpret_s32_u8(vqmovun_s16(q6)); in idct_dequant_full_2x_neon() 211 d31 = vreinterpret_s32_u8(vqmovun_s16(q7)); in idct_dequant_full_2x_neon()
|
D | dequant_idct_neon.c | 129 d14 = vreinterpret_s32_u8(vqmovun_s16(q1)); in vp8_dequant_idct_add_neon() 130 d15 = vreinterpret_s32_u8(vqmovun_s16(q2)); in vp8_dequant_idct_add_neon()
|
D | dc_only_idct_add_neon.c | 34 d2u8 = vqmovun_s16(vreinterpretq_s16_u16(q1u16)); in vp8_dc_only_idct_add_neon()
|
/external/XNNPACK/src/requantization/ |
D | q31-neon.c | 84 const uint8x16_t xyzw_packed = vqmovun_high_s16(vqmovun_s16(xy_packed), zw_packed); in xnn_requantize_q31__neon() 88 const uint8x16_t xyzw_packed = vcombine_u8(vqmovun_s16(xy_packed), vqmovun_s16(zw_packed)); in xnn_requantize_q31__neon()
|
D | gemmlowp-neon.c | 92 const uint8x16_t xyzw_packed = vqmovun_high_s16(vqmovun_s16(xy_packed), zw_packed); in xnn_requantize_gemmlowp__neon() 96 const uint8x16_t xyzw_packed = vcombine_u8(vqmovun_s16(xy_packed), vqmovun_s16(zw_packed)); in xnn_requantize_gemmlowp__neon()
|
D | precise-neon.c | 116 const uint8x16_t xyzw_packed = vqmovun_high_s16(vqmovun_s16(xy_packed), zw_packed); in xnn_requantize_precise__neon() 125 const uint8x16_t xyzw_packed = vcombine_u8(vqmovun_s16(xy_packed), vqmovun_s16(zw_packed)); in xnn_requantize_precise__neon()
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | idct4x4_add_neon.c | 50 d[0] = vqmovun_s16(vreinterpretq_s16_u16(sum[0])); in vpx_idct4x4_16_add_neon() 51 d[1] = vqmovun_s16(vreinterpretq_s16_u16(sum[1])); in vpx_idct4x4_16_add_neon()
|
D | idct_neon.h | 259 b[0] = vqmovun_s16(c[0]); in add_and_store_u8_s16() 260 b[1] = vqmovun_s16(c[1]); in add_and_store_u8_s16() 261 b[2] = vqmovun_s16(c[2]); in add_and_store_u8_s16() 262 b[3] = vqmovun_s16(c[3]); in add_and_store_u8_s16() 263 b[4] = vqmovun_s16(c[4]); in add_and_store_u8_s16() 264 b[5] = vqmovun_s16(c[5]); in add_and_store_u8_s16() 265 b[6] = vqmovun_s16(c[6]); in add_and_store_u8_s16() 266 b[7] = vqmovun_s16(c[7]); in add_and_store_u8_s16() 749 const uint8x8_t d = vqmovun_s16(vreinterpretq_s16_u16(q)); in idct8x8_add8x1() 771 const uint8x8_t d = vqmovun_s16(vreinterpretq_s16_u16(q)); in idct16x16_add8x1() [all …]
|
D | idct32x32_add_neon.c | 65 d[0] = vqmovun_s16(q0); in store_combine_results() 66 d[1] = vqmovun_s16(q1); in store_combine_results() 67 d[2] = vqmovun_s16(q2); in store_combine_results() 68 d[3] = vqmovun_s16(q3); in store_combine_results() 102 d[0] = vmovl_u8(vqmovun_s16(q0)); in highbd_store_combine_results_bd8() 103 d[1] = vmovl_u8(vqmovun_s16(q1)); in highbd_store_combine_results_bd8() 104 d[2] = vmovl_u8(vqmovun_s16(q2)); in highbd_store_combine_results_bd8() 105 d[3] = vmovl_u8(vqmovun_s16(q3)); in highbd_store_combine_results_bd8()
|
D | idct4x4_1_add_neon.c | 26 b = vqmovun_s16(vreinterpretq_s16_u16(a)); in idct4x4_1_add_kernel()
|
/external/XNNPACK/src/q8-vadd/ |
D | neon.c | 92 uint8x16_t vy01 = vqmovun_high_s16(vqmovun_s16(vacc0), vacc1); in xnn_q8_vadd_ukernel__neon() 93 uint8x16_t vy23 = vqmovun_high_s16(vqmovun_s16(vacc2), vacc3); in xnn_q8_vadd_ukernel__neon() 143 uint8x16_t vy01 = vcombine_u8(vqmovun_s16(vacc0), vqmovun_s16(vacc1)); in xnn_q8_vadd_ukernel__neon() 187 uint8x8_t vy = vqmovun_s16(vacc); in xnn_q8_vadd_ukernel__neon() 230 uint8x8_t vy = vqmovun_s16(vacc); in xnn_q8_vadd_ukernel__neon()
|
/external/libvpx/libvpx/vp9/common/arm/neon/ |
D | vp9_iht4x4_add_neon.c | 72 d[0] = vqmovun_s16(vreinterpretq_s16_u16(sum[0])); in vp9_iht4x4_16_add_neon() 73 d[1] = vqmovun_s16(vreinterpretq_s16_u16(sum[1])); in vp9_iht4x4_16_add_neon()
|
/external/libaom/libaom/av1/common/arm/ |
D | cfl_neon.c | 484 vsth_u8(dst, vqmovun_s16(vcombine_s16(pred, pred))); in cfl_predict_lbd_neon() 492 vst1_u8(dst, vqmovun_s16(predict_w8(pred_buf_q3, alpha_sign, in cfl_predict_lbd_neon() 497 const uint8x8x2_t predun = { { vqmovun_s16(pred.val[0]), in cfl_predict_lbd_neon() 498 vqmovun_s16(pred.val[1]) } }; in cfl_predict_lbd_neon() 504 { vqmovun_s16(pred.val[0]), vqmovun_s16(pred.val[1]), in cfl_predict_lbd_neon() 505 vqmovun_s16(pred.val[2]), vqmovun_s16(pred.val[3]) } in cfl_predict_lbd_neon()
|
/external/libaom/libaom/aom_dsp/arm/ |
D | blend_a64_mask_neon.c | 74 vst1_u8((dst + 0 * dst_stride), vqmovun_s16(im_res_0)); in blend_8x4() 75 vst1_u8((dst + 1 * dst_stride), vqmovun_s16(im_res_1)); in blend_8x4() 76 vst1_u8((dst + 2 * dst_stride), vqmovun_s16(im_res_2)); in blend_8x4() 77 vst1_u8((dst + 3 * dst_stride), vqmovun_s16(im_res_3)); in blend_8x4() 117 uint8x8_t res_0 = vqmovun_s16(src0_0); in blend_4x4() 118 uint8x8_t res_1 = vqmovun_s16(src0_1); in blend_4x4()
|
/external/XNNPACK/src/q8-igemm/ |
D | 8x8-neon.c | 561 …uint8x16_t vout0x01234567_1x01234567 = vqmovun_high_s16(vqmovun_s16(vacc0x01234567), vacc1x0123456… in xnn_q8_igemm_ukernel_8x8__neon() 562 …uint8x16_t vout2x01234567_3x01234567 = vqmovun_high_s16(vqmovun_s16(vacc2x01234567), vacc3x0123456… in xnn_q8_igemm_ukernel_8x8__neon() 563 …uint8x16_t vout4x01234567_5x01234567 = vqmovun_high_s16(vqmovun_s16(vacc4x01234567), vacc5x0123456… in xnn_q8_igemm_ukernel_8x8__neon() 564 …uint8x16_t vout6x01234567_7x01234567 = vqmovun_high_s16(vqmovun_s16(vacc6x01234567), vacc7x0123456… in xnn_q8_igemm_ukernel_8x8__neon() 583 …uint8x16_t vout0x01234567_1x01234567 = vcombine_u8(vqmovun_s16(vacc0x01234567), vqmovun_s16(vacc1x… in xnn_q8_igemm_ukernel_8x8__neon() 584 …uint8x16_t vout2x01234567_3x01234567 = vcombine_u8(vqmovun_s16(vacc2x01234567), vqmovun_s16(vacc3x… in xnn_q8_igemm_ukernel_8x8__neon() 585 …uint8x16_t vout4x01234567_5x01234567 = vcombine_u8(vqmovun_s16(vacc4x01234567), vqmovun_s16(vacc5x… in xnn_q8_igemm_ukernel_8x8__neon() 586 …uint8x16_t vout6x01234567_7x01234567 = vcombine_u8(vqmovun_s16(vacc6x01234567), vqmovun_s16(vacc7x… in xnn_q8_igemm_ukernel_8x8__neon()
|
D | 4x8-neon.c | 357 …uint8x16_t vout0x01234567_1x01234567 = vqmovun_high_s16(vqmovun_s16(vacc0x01234567), vacc1x0123456… in xnn_q8_igemm_ukernel_4x8__neon() 358 …uint8x16_t vout2x01234567_3x01234567 = vqmovun_high_s16(vqmovun_s16(vacc2x01234567), vacc3x0123456… in xnn_q8_igemm_ukernel_4x8__neon() 365 …uint8x16_t vout0x01234567_1x01234567 = vcombine_u8(vqmovun_s16(vacc0x01234567), vqmovun_s16(vacc1x… in xnn_q8_igemm_ukernel_4x8__neon() 366 …uint8x16_t vout2x01234567_3x01234567 = vcombine_u8(vqmovun_s16(vacc2x01234567), vqmovun_s16(vacc3x… in xnn_q8_igemm_ukernel_4x8__neon()
|
/external/XNNPACK/src/q8-gemm/ |
D | 8x8-neon.c | 516 …uint8x16_t vout0x01234567_1x01234567 = vqmovun_high_s16(vqmovun_s16(vacc0x01234567), vacc1x0123456… in xnn_q8_gemm_ukernel_8x8__neon() 517 …uint8x16_t vout2x01234567_3x01234567 = vqmovun_high_s16(vqmovun_s16(vacc2x01234567), vacc3x0123456… in xnn_q8_gemm_ukernel_8x8__neon() 518 …uint8x16_t vout4x01234567_5x01234567 = vqmovun_high_s16(vqmovun_s16(vacc4x01234567), vacc5x0123456… in xnn_q8_gemm_ukernel_8x8__neon() 519 …uint8x16_t vout6x01234567_7x01234567 = vqmovun_high_s16(vqmovun_s16(vacc6x01234567), vacc7x0123456… in xnn_q8_gemm_ukernel_8x8__neon() 538 …uint8x16_t vout0x01234567_1x01234567 = vcombine_u8(vqmovun_s16(vacc0x01234567), vqmovun_s16(vacc1x… in xnn_q8_gemm_ukernel_8x8__neon() 539 …uint8x16_t vout2x01234567_3x01234567 = vcombine_u8(vqmovun_s16(vacc2x01234567), vqmovun_s16(vacc3x… in xnn_q8_gemm_ukernel_8x8__neon() 540 …uint8x16_t vout4x01234567_5x01234567 = vcombine_u8(vqmovun_s16(vacc4x01234567), vqmovun_s16(vacc5x… in xnn_q8_gemm_ukernel_8x8__neon() 541 …uint8x16_t vout6x01234567_7x01234567 = vcombine_u8(vqmovun_s16(vacc6x01234567), vqmovun_s16(vacc7x… in xnn_q8_gemm_ukernel_8x8__neon()
|
D | 4x8-neon.c | 320 …uint8x16_t vout0x01234567_1x01234567 = vqmovun_high_s16(vqmovun_s16(vacc0x01234567), vacc1x0123456… in xnn_q8_gemm_ukernel_4x8__neon() 321 …uint8x16_t vout2x01234567_3x01234567 = vqmovun_high_s16(vqmovun_s16(vacc2x01234567), vacc3x0123456… in xnn_q8_gemm_ukernel_4x8__neon() 328 …uint8x16_t vout0x01234567_1x01234567 = vcombine_u8(vqmovun_s16(vacc0x01234567), vqmovun_s16(vacc1x… in xnn_q8_gemm_ukernel_4x8__neon() 329 …uint8x16_t vout2x01234567_3x01234567 = vcombine_u8(vqmovun_s16(vacc2x01234567), vqmovun_s16(vacc3x… in xnn_q8_gemm_ukernel_4x8__neon()
|
/external/libhevc/common/arm/ |
D | ihevc_inter_pred_filters_luma_vert_w16inp.s | 197 vqrshrun.s16 d8,q4,#6 @sto_res = vqmovun_s16(sto_res_tmp)@ 215 vqrshrun.s16 d10,q5,#6 @sto_res = vqmovun_s16(sto_res_tmp)@ 281 vqrshrun.s16 d8,q4,#6 @sto_res = vqmovun_s16(sto_res_tmp)@ 303 vqrshrun.s16 d10,q5,#6 @sto_res = vqmovun_s16(sto_res_tmp)@ 348 vqrshrun.s16 d8,q4,#6 @sto_res = vqmovun_s16(sto_res_tmp)@ 360 vqrshrun.s16 d10,q5,#6 @sto_res = vqmovun_s16(sto_res_tmp)@
|
D | ihevc_inter_pred_luma_vert_w16inp_w16out.s | 208 @vqrshrun.s16 d8,q4,#6 @sto_res = vqmovun_s16(sto_res_tmp)@ 227 @vqrshrun.s16 d10,q5,#6 @sto_res = vqmovun_s16(sto_res_tmp)@ 297 @vqrshrun.s16 d8,q4,#6 @sto_res = vqmovun_s16(sto_res_tmp)@ 320 @vqrshrun.s16 d10,q5,#6 @sto_res = vqmovun_s16(sto_res_tmp)@ 368 @vqrshrun.s16 d8,q4,#6 @sto_res = vqmovun_s16(sto_res_tmp)@ 381 @vqrshrun.s16 d10,q5,#6 @sto_res = vqmovun_s16(sto_res_tmp)@
|