/external/libvpx/libvpx/vpx_dsp/arm/ |
D | loopfilter_8_neon.c | 19 uint8x8_t d3u8, // p2 in mbloop_filter_neon() argument 40 d19u8 = vabd_u8(d3u8, d4u8); in mbloop_filter_neon() 59 d27u8 = vabd_u8(d3u8, d6u8); in mbloop_filter_neon() 109 q14u16 = vmlal_u8(q14u16, d3u8, d27u8); in mbloop_filter_neon() 114 q14u16 = vsubw_u8(q14u16, d3u8); in mbloop_filter_neon() 120 q14u16 = vsubw_u8(q14u16, d3u8); in mbloop_filter_neon() 126 q14u16 = vsubw_u8(q14u16, d3u8); in mbloop_filter_neon() 197 q14u16 = vmlal_u8(q14u16, d3u8, d27u8); in mbloop_filter_neon() 208 q14u16 = vsubw_u8(q14u16, d3u8); in mbloop_filter_neon() 217 q14u16 = vsubw_u8(q14u16, d3u8); in mbloop_filter_neon() [all …]
|
D | loopfilter_4_neon.c | 19 uint8x8_t d3u8, // p3 in loop_filter_neon() argument 35 d19u8 = vabd_u8(d3u8, d4u8); in loop_filter_neon() 39 d3u8 = vabd_u8(d17u8, d16u8); in loop_filter_neon() 44 d3u8 = vmax_u8(d3u8, d4u8); in loop_filter_neon() 51 d23u8 = vmax_u8(d23u8, d3u8); in loop_filter_neon() 123 uint8x8_t d3u8, d4u8, d5u8, d6u8, d7u8, d16u8, d17u8, d18u8; in vpx_lpf_horizontal_4_neon() local 136 d3u8 = vld1_u8(s); in vpx_lpf_horizontal_4_neon() 153 d3u8, d4u8, d5u8, d6u8, d7u8, d16u8, d17u8, d18u8, in vpx_lpf_horizontal_4_neon() 178 uint8x8_t d3u8, d4u8, d5u8, d6u8, d7u8, d16u8, d17u8, d18u8; in vpx_lpf_vertical_4_neon() local 195 d3u8 = vld1_u8(s); in vpx_lpf_vertical_4_neon() [all …]
|
D | idct8x8_1_add_neon.c | 20 uint8x8_t d2u8, d3u8, d30u8, d31u8; in vpx_idct8x8_1_add_neon() local 50 d3u8 = vqmovun_s16(vreinterpretq_s16_u16(q10u16)); in vpx_idct8x8_1_add_neon() 56 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); in vpx_idct8x8_1_add_neon()
|
D | idct16x16_1_add_neon.c | 20 uint8x8_t d2u8, d3u8, d30u8, d31u8; in vpx_idct16x16_1_add_neon() local 48 d3u8 = vqmovun_s16(vreinterpretq_s16_u16(q10u16)); in vpx_idct16x16_1_add_neon() 53 vst1_u64((uint64_t *)(d2 + 8), vreinterpret_u64_u8(d3u8)); in vpx_idct16x16_1_add_neon()
|
D | intrapred_neon.c | 667 uint8x8_t d2u8, d3u8, d18u8, d22u8, d23u8; in vpx_tm_predictor_16x16_neon() local 689 d3u8 = vqmovun_s16(q0s16); in vpx_tm_predictor_16x16_neon() 693 vst1_u64((uint64_t *)(dst + 8), vreinterpret_u64_u8(d3u8)); in vpx_tm_predictor_16x16_neon() 710 d3u8 = vqmovun_s16(q0s16); in vpx_tm_predictor_16x16_neon() 714 vst1_u64((uint64_t *)(dst + 8), vreinterpret_u64_u8(d3u8)); in vpx_tm_predictor_16x16_neon() 730 uint8x8_t d0u8, d1u8, d2u8, d3u8, d26u8; in vpx_tm_predictor_32x32_neon() local 756 d3u8 = vqmovun_s16(q15s16); in vpx_tm_predictor_32x32_neon() 758 q1u8 = vcombine_u8(d2u8, d3u8); in vpx_tm_predictor_32x32_neon() 775 d3u8 = vqmovun_s16(q15s16); in vpx_tm_predictor_32x32_neon() 777 q1u8 = vcombine_u8(d2u8, d3u8); in vpx_tm_predictor_32x32_neon() [all …]
|
D | idct8x8_add_neon.c | 246 uint8x8_t d0u8, d1u8, d2u8, d3u8; in vpx_idct8x8_64_add_neon() local 304 d3u8 = vqmovun_s16(vreinterpretq_s16_u16(q11u16)); in vpx_idct8x8_64_add_neon() 312 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); in vpx_idct8x8_64_add_neon() 341 d3u8 = vqmovun_s16(vreinterpretq_s16_u16(q11u16)); in vpx_idct8x8_64_add_neon() 349 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); in vpx_idct8x8_64_add_neon() 359 uint8x8_t d0u8, d1u8, d2u8, d3u8; in vpx_idct8x8_12_add_neon() local 492 d3u8 = vqmovun_s16(vreinterpretq_s16_u16(q11u16)); in vpx_idct8x8_12_add_neon() 500 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); in vpx_idct8x8_12_add_neon() 529 d3u8 = vqmovun_s16(vreinterpretq_s16_u16(q11u16)); in vpx_idct8x8_12_add_neon() 537 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); in vpx_idct8x8_12_add_neon()
|
D | vpx_convolve_avg_neon.c | 28 uint8x8_t d0u8, d1u8, d2u8, d3u8; in vpx_convolve_avg_neon() local 113 d3u8 = vld1_u8(d); in vpx_convolve_avg_neon() 117 q1u8 = vcombine_u8(d2u8, d3u8); in vpx_convolve_avg_neon()
|
D | vpx_convolve8_avg_neon.c | 60 uint8x8_t d2u8, d3u8, d24u8, d25u8, d26u8, d27u8, d28u8, d29u8; in vpx_convolve8_avg_horiz_neon() local 196 d3u8 = vqmovn_u16(q2u16); in vpx_convolve8_avg_horiz_neon() 199 vreinterpret_u16_u8(d3u8)); in vpx_convolve8_avg_horiz_neon() 248 uint8x8_t d2u8, d3u8; in vpx_convolve8_avg_vert_neon() local 346 d3u8 = vqmovn_u16(q2u16); in vpx_convolve8_avg_vert_neon() 348 q1u8 = vcombine_u8(d2u8, d3u8); in vpx_convolve8_avg_vert_neon()
|
D | vpx_convolve8_neon.c | 60 uint8x8_t d2u8, d3u8, d24u8, d25u8, d26u8, d27u8, d28u8, d29u8; in vpx_convolve8_horiz_neon() local 189 d3u8 = vqmovn_u16(q2u16); in vpx_convolve8_horiz_neon() 192 vreinterpret_u16_u8(d3u8)); in vpx_convolve8_horiz_neon()
|
D | variance_neon.c | 373 uint8x8_t d0u8, d1u8, d2u8, d3u8, d4u8, d5u8, d6u8, d7u8; in vpx_get4x4sse_cs_neon() local 390 d3u8 = vld1_u8(src_ptr); in vpx_get4x4sse_cs_neon() 398 q14u16 = vsubl_u8(d3u8, d7u8); in vpx_get4x4sse_cs_neon()
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | bilinearpredict_neon.c | 31 uint8x8_t d0u8, d1u8, d2u8, d3u8, d4u8, d5u8, d6u8; in vp8_bilinear_predict4x4_neon() local 59 d3u8 = vld1_u8(src_ptr); src_ptr += src_pixels_per_line; in vp8_bilinear_predict4x4_neon() 64 q1u8 = vcombine_u8(d2u8, d3u8); in vp8_bilinear_predict4x4_neon() 119 d3u8 = vqrshrn_n_u16(q2u16, 7); in vp8_bilinear_predict4x4_neon() 125 vst1_lane_u32((uint32_t *)dst_ptr, vreinterpret_u32_u8(d3u8), 0); in vp8_bilinear_predict4x4_neon() 127 vst1_lane_u32((uint32_t *)dst_ptr, vreinterpret_u32_u8(d3u8), 1); in vp8_bilinear_predict4x4_neon() 139 uint8x8_t d0u8, d1u8, d2u8, d3u8, d4u8, d5u8; in vp8_bilinear_predict8x4_neon() local 167 d3u8 = vext_u8(vget_low_u8(q1u8), vget_high_u8(q1u8), 1); in vp8_bilinear_predict8x4_neon() 173 q6u16 = vmlal_u8(q6u16, d3u8, d1u8); in vp8_bilinear_predict8x4_neon() 207 d3u8 = vqrshrn_n_u16(q2u16, 7); in vp8_bilinear_predict8x4_neon() [all …]
|
D | sixtappredict_neon.c | 33 uint8x8_t d0u8, d1u8, d2u8, d3u8, d4u8, d5u8, d18u8, d19u8, d20u8, d21u8; in vp8_sixtap_predict4x4_neon() local 62 d3u8 = vreinterpret_u8_s8(vabs_s8(d3s8)); in vp8_sixtap_predict4x4_neon() 109 q5u16 = vmlal_u8(q5u16, d24u8, d3u8); in vp8_sixtap_predict4x4_neon() 110 q6u16 = vmlal_u8(q6u16, d25u8, d3u8); in vp8_sixtap_predict4x4_neon() 120 d3u8 = vqrshrun_n_s16(q5s16, 7); in vp8_sixtap_predict4x4_neon() 123 vst1_lane_u32((uint32_t *)dst_ptr, vreinterpret_u32_u8(d3u8), 0); in vp8_sixtap_predict4x4_neon() 125 vst1_lane_u32((uint32_t *)dst_ptr, vreinterpret_u32_u8(d3u8), 1); in vp8_sixtap_predict4x4_neon() 144 d3u8 = vreinterpret_u8_s8(vabs_s8(d3s8)); in vp8_sixtap_predict4x4_neon() 224 q9u16 = vmull_u8(vreinterpret_u8_u32(d0u32x2.val[0]), d3u8); in vp8_sixtap_predict4x4_neon() 225 q10u16 = vmull_u8(vreinterpret_u8_u32(d1u32x2.val[0]), d3u8); in vp8_sixtap_predict4x4_neon() [all …]
|
/external/libvpx/libvpx/vp9/common/arm/neon/ |
D | vp9_iht8x8_add_neon.c | 493 uint8x8_t d0u8, d1u8, d2u8, d3u8; in vp9_iht8x8_64_add_neon() local 612 d3u8 = vqmovun_s16(vreinterpretq_s16_u16(q11u16)); in vp9_iht8x8_64_add_neon() 620 vst1_u64((uint64_t *)d2, vreinterpret_u64_u8(d3u8)); in vp9_iht8x8_64_add_neon()
|