/external/libaom/libaom/aom_dsp/arm/ |
D | loopfilter_neon.c | 37 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(mask_8x8))); in lpf_mask() 40 p0q0_p1q1 = vtrn_u32(vreinterpret_u32_u8(p0q0), vreinterpret_u32_u8(p1q1)); in lpf_mask() 67 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(mask_8x8))); in lpf_mask2() 70 p0q0_p1q1 = vtrn_u32(vreinterpret_u32_u8(p0q0), vreinterpret_u32_u8(p1q1)); in lpf_mask2() 95 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(flat_8x8))); in lpf_flat_mask4() 110 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(flat_8x8))); in lpf_flat_mask3() 131 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(mask_8x8))); in lpf_mask3_chroma() 134 p0q0_p1q1 = vtrn_u32(vreinterpret_u32_u8(p0q0), vreinterpret_u32_u8(p1q1)); in lpf_mask3_chroma() 192 temp1_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(temp0_8x8))); in lpf_14_neon() 225 q0p0 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(*p0q0))); in lpf_14_neon() [all …]
|
D | intrapred_neon.c | 58 vst1_lane_u32((uint32_t *)(dst + i * stride), vreinterpret_u32_u8(dc), 0); in dc_4x4() 128 vst1_u32((uint32_t *)(dst + i * stride), vreinterpret_u32_u8(dc)); in dc_8x8() 336 const uint32x2_t r3 = vreinterpret_u32_u8(avg2); in aom_d135_predictor_4x4_neon() 401 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(d0u8), 0); in aom_h_predictor_4x4_neon() 404 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(d0u8), 0); in aom_h_predictor_4x4_neon() 407 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(d0u8), 0); in aom_h_predictor_4x4_neon() 410 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(d0u8), 0); in aom_h_predictor_4x4_neon()
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | vpx_convolve_avg_neon.c | 35 s01 = vzip_u32(vreinterpret_u32_u8(s0), vreinterpret_u32_u8(s1)); in vpx_convolve_avg_neon() 37 vld1_lane_u32((const uint32_t *)dst, vreinterpret_u32_u8(dd0), 0)); in vpx_convolve_avg_neon() 39 (const uint32_t *)(dst + dst_stride), vreinterpret_u32_u8(dd0), 1)); in vpx_convolve_avg_neon() 41 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(dd0), 0); in vpx_convolve_avg_neon() 43 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(dd0), 1); in vpx_convolve_avg_neon()
|
D | vpx_scaled_convolve8_neon.c | 66 vst1_lane_u32((uint32_t *)&temp[4 * z], vreinterpret_u32_u8(d), 0); in scaledconvolve_horiz_w4() 80 vreinterpret_u32_u8(d4.val[0]), 0); in scaledconvolve_horiz_w4() 82 vreinterpret_u32_u8(d4.val[1]), 0); in scaledconvolve_horiz_w4() 84 vreinterpret_u32_u8(d4.val[2]), 0); in scaledconvolve_horiz_w4() 86 vreinterpret_u32_u8(d4.val[3]), 0); in scaledconvolve_horiz_w4() 189 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(d), 0); in scaledconvolve_vert_w4()
|
D | idct4x4_1_add_neon.c | 27 vst1_lane_u32((uint32_t *)*dest, vreinterpret_u32_u8(b), 0); in idct4x4_1_add_kernel() 29 vst1_lane_u32((uint32_t *)*dest, vreinterpret_u32_u8(b), 1); in idct4x4_1_add_kernel()
|
D | idct4x4_add_neon.c | 56 vst1_lane_u32((uint32_t *)dest, vreinterpret_u32_u8(d[1]), 1); in vpx_idct4x4_16_add_neon() 58 vst1_lane_u32((uint32_t *)dest, vreinterpret_u32_u8(d[1]), 0); in vpx_idct4x4_16_add_neon()
|
D | vpx_convolve8_neon.c | 130 vreinterpret_u32_u8(d01), 0); in vpx_convolve8_horiz_neon() 132 vreinterpret_u32_u8(d23), 0); in vpx_convolve8_horiz_neon() 134 vreinterpret_u32_u8(d01), 1); in vpx_convolve8_horiz_neon() 136 vreinterpret_u32_u8(d23), 1); in vpx_convolve8_horiz_neon() 204 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(t0), 0); in vpx_convolve8_horiz_neon() 206 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(t1), 0); in vpx_convolve8_horiz_neon() 208 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(t2), 0); in vpx_convolve8_horiz_neon() 210 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(t3), 0); in vpx_convolve8_horiz_neon() 212 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(t0), 1); in vpx_convolve8_horiz_neon() 214 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(t1), 1); in vpx_convolve8_horiz_neon() [all …]
|
D | mem_neon.h | 113 const uint32x2_t a_u32 = vreinterpret_u32_u8(a); in store_unaligned_u8() 175 uint32x2_t a_u32 = vreinterpret_u32_u8(a); in store_u8()
|
D | intrapred_neon.c | 31 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(dc_dup), 0); in dc_store_4x4() 273 const uint32x2_t r0 = vreinterpret_u32_u8(avg2); in vpx_d45_predictor_4x4_neon() 398 const uint32x2_t r3 = vreinterpret_u32_u8(avg2); in vpx_d135_predictor_4x4_neon() 734 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(d), 0); in vpx_h_predictor_4x4_neon() 737 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(d), 0); in vpx_h_predictor_4x4_neon() 740 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(d), 0); in vpx_h_predictor_4x4_neon() 743 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(d), 0); in vpx_h_predictor_4x4_neon() 894 d = vreinterpret_u32_u8(vqmovun_s16(sum)); in vpx_tm_predictor_4x4_neon() 902 d = vreinterpret_u32_u8(vqmovun_s16(sum)); in vpx_tm_predictor_4x4_neon()
|
D | transpose_neon.h | 211 vtrn_u32(vreinterpret_u32_u8(*a0), vreinterpret_u32_u8(a4)); in transpose_u8_4x8() 213 vtrn_u32(vreinterpret_u32_u8(*a1), vreinterpret_u32_u8(a5)); in transpose_u8_4x8() 215 vtrn_u32(vreinterpret_u32_u8(*a2), vreinterpret_u32_u8(a6)); in transpose_u8_4x8() 217 vtrn_u32(vreinterpret_u32_u8(*a3), vreinterpret_u32_u8(a7)); in transpose_u8_4x8()
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | loopfiltersimpleverticaledge_neon.c | 107 vtrn_u32(vreinterpret_u32_u8(a), vreinterpret_u32_u8(e)); in read_4x8() 109 vtrn_u32(vreinterpret_u32_u8(b), vreinterpret_u32_u8(f)); in read_4x8() 111 vtrn_u32(vreinterpret_u32_u8(c), vreinterpret_u32_u8(g)); in read_4x8() 113 vtrn_u32(vreinterpret_u32_u8(d), vreinterpret_u32_u8(h)); in read_4x8()
|
D | dc_only_idct_add_neon.c | 36 vst1_lane_u32((uint32_t *)dst_ptr, vreinterpret_u32_u8(d2u8), 0); in vp8_dc_only_idct_add_neon() 38 vst1_lane_u32((uint32_t *)dst_ptr, vreinterpret_u32_u8(d2u8), 1); in vp8_dc_only_idct_add_neon()
|
D | shortidct4x4llm_neon.c | 115 vst1_lane_u32((uint32_t *)dst_ptr, vreinterpret_u32_u8(d1u8), 0); in vp8_short_idct4x4llm_neon() 117 vst1_lane_u32((uint32_t *)dst_ptr, vreinterpret_u32_u8(d1u8), 1); in vp8_short_idct4x4llm_neon()
|
D | bilinearpredict_neon.c | 75 c0 = vzip_u32(vreinterpret_u32_u8(vget_low_u8(a01)), in vp8_bilinear_predict4x4_neon() 76 vreinterpret_u32_u8(vget_high_u8(a01))); in vp8_bilinear_predict4x4_neon() 77 c1 = vzip_u32(vreinterpret_u32_u8(vget_low_u8(a23)), in vp8_bilinear_predict4x4_neon() 78 vreinterpret_u32_u8(vget_high_u8(a23))); in vp8_bilinear_predict4x4_neon() 79 c2 = vzip_u32(vreinterpret_u32_u8(vget_low_u8(b01)), in vp8_bilinear_predict4x4_neon() 80 vreinterpret_u32_u8(vget_high_u8(b01))); in vp8_bilinear_predict4x4_neon() 81 c3 = vzip_u32(vreinterpret_u32_u8(vget_low_u8(b23)), in vp8_bilinear_predict4x4_neon() 82 vreinterpret_u32_u8(vget_high_u8(b23))); in vp8_bilinear_predict4x4_neon()
|
D | sixtappredict_neon.c | 50 const uint32x2x2_t a_shuf = vzip_u32(vreinterpret_u32_u8(vget_low_u8(a)), in filter_add_accumulate() 51 vreinterpret_u32_u8(vget_high_u8(a))); in filter_add_accumulate() 52 const uint32x2x2_t b_shuf = vzip_u32(vreinterpret_u32_u8(vget_low_u8(b)), in filter_add_accumulate() 53 vreinterpret_u32_u8(vget_high_u8(b))); in filter_add_accumulate() 61 const uint32x2x2_t a_shuf = vzip_u32(vreinterpret_u32_u8(vget_low_u8(a)), in filter_sub_accumulate() 62 vreinterpret_u32_u8(vget_high_u8(a))); in filter_sub_accumulate() 63 const uint32x2x2_t b_shuf = vzip_u32(vreinterpret_u32_u8(vget_low_u8(b)), in filter_sub_accumulate() 64 vreinterpret_u32_u8(vget_high_u8(b))); in filter_sub_accumulate() 224 s01_f5_q = vzip_u32(vreinterpret_u32_u8(s0_f5), vreinterpret_u32_u8(s1_f5)); in vp8_sixtap_predict4x4_neon() 225 s23_f5_q = vzip_u32(vreinterpret_u32_u8(s2_f5), vreinterpret_u32_u8(s3_f5)); in vp8_sixtap_predict4x4_neon() [all …]
|
/external/skqp/src/core/ |
D | SkBlitRow_D32.cpp | 174 vst1_u32(dst, vreinterpret_u32_u8(vres)); in blit_row_s32_blend() 185 vsrc = vreinterpret_u8_u32(vld1_lane_u32(src, vreinterpret_u32_u8(vsrc), 0)); in blit_row_s32_blend() 186 vdst = vreinterpret_u8_u32(vld1_lane_u32(dst, vreinterpret_u32_u8(vdst), 0)); in blit_row_s32_blend() 194 vst1_lane_u32(dst, vreinterpret_u32_u8(vres), 0); in blit_row_s32_blend() 208 vsrc = vreinterpret_u8_u32(vld1_lane_u32(src, vreinterpret_u32_u8(vsrc), 0)); in blit_row_s32a_blend() 209 vdst = vreinterpret_u8_u32(vld1_lane_u32(dst, vreinterpret_u32_u8(vdst), 0)); in blit_row_s32a_blend() 223 vst1_lane_u32(dst, vreinterpret_u32_u8(vres), 0); in blit_row_s32a_blend() 266 vst1_u32(dst, vreinterpret_u32_u8(vres)); in blit_row_s32a_blend()
|
/external/skia/src/core/ |
D | SkBlitRow_D32.cpp | 174 vst1_u32(dst, vreinterpret_u32_u8(vres)); in blit_row_s32_blend() 185 vsrc = vreinterpret_u8_u32(vld1_lane_u32(src, vreinterpret_u32_u8(vsrc), 0)); in blit_row_s32_blend() 186 vdst = vreinterpret_u8_u32(vld1_lane_u32(dst, vreinterpret_u32_u8(vdst), 0)); in blit_row_s32_blend() 194 vst1_lane_u32(dst, vreinterpret_u32_u8(vres), 0); in blit_row_s32_blend() 208 vsrc = vreinterpret_u8_u32(vld1_lane_u32(src, vreinterpret_u32_u8(vsrc), 0)); in blit_row_s32a_blend() 209 vdst = vreinterpret_u8_u32(vld1_lane_u32(dst, vreinterpret_u32_u8(vdst), 0)); in blit_row_s32a_blend() 223 vst1_lane_u32(dst, vreinterpret_u32_u8(vres), 0); in blit_row_s32a_blend() 266 vst1_u32(dst, vreinterpret_u32_u8(vres)); in blit_row_s32a_blend()
|
/external/libhevc/common/arm/ |
D | ihevc_weighted_pred_neon_intr.c | 170 vst1_lane_u32((uint32_t *)pu1_dst, vreinterpret_u32_u8(sto_res), 0); in ihevc_weighted_pred_uni_neonintr() 174 vst1_lane_u32((uint32_t *)pu1_dst_tmp, vreinterpret_u32_u8(sto_res), 0); in ihevc_weighted_pred_uni_neonintr() 317 vst1_lane_u32((uint32_t *)pu1_dst, vreinterpret_u32_u8(sto_res), 0); in ihevc_weighted_pred_chroma_uni_neonintr() 321 vst1_lane_u32((uint32_t *)pu1_dst_tmp, vreinterpret_u32_u8(sto_res), 0); in ihevc_weighted_pred_chroma_uni_neonintr() 481 vst1_lane_u32((uint32_t *)pu1_dst, vreinterpret_u32_u8(sto_res), 0); in ihevc_weighted_pred_bi_neonintr() 485 vst1_lane_u32((uint32_t *)pu1_dst_tmp, vreinterpret_u32_u8(sto_res), 0); in ihevc_weighted_pred_bi_neonintr() 667 vst1_lane_u32((uint32_t *)pu1_dst, vreinterpret_u32_u8(sto_res), 0); in ihevc_weighted_pred_chroma_bi_neonintr() 671 vst1_lane_u32((uint32_t *)pu1_dst_tmp, vreinterpret_u32_u8(sto_res), 0); in ihevc_weighted_pred_chroma_bi_neonintr() 816 vst1_lane_u32((uint32_t *)pu1_dst, vreinterpret_u32_u8(sto_res), 0); in ihevc_weighted_pred_bi_default_neonintr() 820 vst1_lane_u32((uint32_t *)pu1_dst_tmp, vreinterpret_u32_u8(sto_res), 0); in ihevc_weighted_pred_bi_default_neonintr() [all …]
|
D | ihevc_intra_pred_filters_neon_intr.c | 235 … vst1_lane_u32((uint32_t *)pu1_dst_const_nt, vreinterpret_u32_u8(dup_pu1_dst1), 0); in ihevc_intra_pred_luma_ref_substitution_neonintr() 286 … vst1_lane_u32((uint32_t *)pu1_dst_const_three_nt_1, vreinterpret_u32_u8(dup_pu1_dst3), 0); in ihevc_intra_pred_luma_ref_substitution_neonintr() 849 vst1_lane_u32((uint32_t *)pu1_dst, vreinterpret_u32_u8(sto_res), 0); in ihevc_intra_pred_luma_planar_neonintr() 852 vst1_lane_u32((uint32_t *)pu1_dst_tmp, vreinterpret_u32_u8(sto_res), 1); in ihevc_intra_pred_luma_planar_neonintr() 1114 vst1_lane_u32((uint32_t *)pu1_dst_tmp, vreinterpret_u32_u8(dc_val_t), 0); in ihevc_intra_pred_luma_dc_neonintr() 1134 vst1_lane_u32((uint32_t *)pu1_dst_tmp, vreinterpret_u32_u8(dc_val_t), 0); in ihevc_intra_pred_luma_dc_neonintr() 1263 vst1_lane_u32((uint32_t *)pu1_dst_4, vreinterpret_u32_u8(round_val), 0); in ihevc_intra_pred_luma_horz_neonintr() 1268 vst1_lane_u32((uint32_t *)pu1_dst_4, vreinterpret_u32_u8(dup_val), 0); in ihevc_intra_pred_luma_horz_neonintr() 1272 vst1_lane_u32((uint32_t *)pu1_dst_4, vreinterpret_u32_u8(dup_val), 0); in ihevc_intra_pred_luma_horz_neonintr() 1276 vst1_lane_u32((uint32_t *)pu1_dst_4, vreinterpret_u32_u8(dup_val), 0); in ihevc_intra_pred_luma_horz_neonintr() [all …]
|
/external/libaom/libaom/av1/common/arm/ |
D | convolve_neon.c | 280 vreinterpret_u32_u8(d01), 0); in av1_convolve_x_sr_neon() 282 vreinterpret_u32_u8(d23), 0); in av1_convolve_x_sr_neon() 284 vreinterpret_u32_u8(d01), 1); in av1_convolve_x_sr_neon() 286 vreinterpret_u32_u8(d23), 1); in av1_convolve_x_sr_neon() 372 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(t0), in av1_convolve_x_sr_neon() 375 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(t1), in av1_convolve_x_sr_neon() 378 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(t2), in av1_convolve_x_sr_neon() 381 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(t3), in av1_convolve_x_sr_neon() 384 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(t0), in av1_convolve_x_sr_neon() 387 vst1_lane_u32((uint32_t *)dst, vreinterpret_u32_u8(t1), in av1_convolve_x_sr_neon() [all …]
|
D | transpose_neon.h | 152 vtrn_u32(vreinterpret_u32_u8(*a0), vreinterpret_u32_u8(a4)); in transpose_u8_4x8() 154 vtrn_u32(vreinterpret_u32_u8(*a1), vreinterpret_u32_u8(a5)); in transpose_u8_4x8() 156 vtrn_u32(vreinterpret_u32_u8(*a2), vreinterpret_u32_u8(a6)); in transpose_u8_4x8() 158 vtrn_u32(vreinterpret_u32_u8(*a3), vreinterpret_u32_u8(a7)); in transpose_u8_4x8()
|
D | jnt_convolve_neon.c | 654 vst1_lane_u32((uint32_t *)d_u8, vreinterpret_u32_u8(t0), 0); in dist_wtd_convolve_2d_vert_neon() 656 vst1_lane_u32((uint32_t *)d_u8, vreinterpret_u32_u8(t0), 1); in dist_wtd_convolve_2d_vert_neon() 658 vst1_lane_u32((uint32_t *)d_u8, vreinterpret_u32_u8(t1), 0); in dist_wtd_convolve_2d_vert_neon() 660 vst1_lane_u32((uint32_t *)d_u8, vreinterpret_u32_u8(t1), 1); in dist_wtd_convolve_2d_vert_neon() 692 vst1_lane_u32((uint32_t *)d_u8, vreinterpret_u32_u8(t0), 0); in dist_wtd_convolve_2d_vert_neon() 855 vst1_lane_u32((uint32_t *)(dst8_1), vreinterpret_u32_u8(tmp_shift0), 0); in av1_dist_wtd_convolve_2d_copy_neon() 857 vst1_lane_u32((uint32_t *)(dst8_1), vreinterpret_u32_u8(tmp_shift0), 1); in av1_dist_wtd_convolve_2d_copy_neon() 859 vst1_lane_u32((uint32_t *)(dst8_1), vreinterpret_u32_u8(tmp_shift1), 0); in av1_dist_wtd_convolve_2d_copy_neon() 861 vst1_lane_u32((uint32_t *)(dst8_1), vreinterpret_u32_u8(tmp_shift1), 1); in av1_dist_wtd_convolve_2d_copy_neon() 1032 vst1_lane_u32((uint32_t *)d_u8, vreinterpret_u32_u8(t0), in av1_dist_wtd_convolve_x_neon() [all …]
|
D | blend_a64_hmask_neon.c | 101 vreinterpret_u32_u8(vrshrn_n_u16(res, AOM_BLEND_A64_ROUND_BITS)), 0); in aom_blend_a64_hmask_neon() 104 vreinterpret_u32_u8(vrshrn_n_u16(res, AOM_BLEND_A64_ROUND_BITS)), 1); in aom_blend_a64_hmask_neon()
|
D | blend_a64_vmask_neon.c | 100 vreinterpret_u32_u8(vrshrn_n_u16(res, AOM_BLEND_A64_ROUND_BITS)), 0); in aom_blend_a64_vmask_neon() 103 vreinterpret_u32_u8(vrshrn_n_u16(res, AOM_BLEND_A64_ROUND_BITS)), 1); in aom_blend_a64_vmask_neon()
|
/external/libvpx/libvpx/vp9/encoder/arm/neon/ |
D | vp9_frame_scale_neon.c | 226 vst1_lane_u32((uint32_t *)(t + 0 * width_hor), vreinterpret_u32_u8(d[0]), in scale_plane_2_to_1_general() 228 vst1_lane_u32((uint32_t *)(t + 1 * width_hor), vreinterpret_u32_u8(d[1]), in scale_plane_2_to_1_general() 230 vst1_lane_u32((uint32_t *)(t + 2 * width_hor), vreinterpret_u32_u8(d[2]), in scale_plane_2_to_1_general() 232 vst1_lane_u32((uint32_t *)(t + 3 * width_hor), vreinterpret_u32_u8(d[3]), in scale_plane_2_to_1_general() 234 vst1_lane_u32((uint32_t *)(t + 4 * width_hor), vreinterpret_u32_u8(d[0]), in scale_plane_2_to_1_general() 236 vst1_lane_u32((uint32_t *)(t + 5 * width_hor), vreinterpret_u32_u8(d[1]), in scale_plane_2_to_1_general() 238 vst1_lane_u32((uint32_t *)(t + 6 * width_hor), vreinterpret_u32_u8(d[2]), in scale_plane_2_to_1_general() 240 vst1_lane_u32((uint32_t *)(t + 7 * width_hor), vreinterpret_u32_u8(d[3]), in scale_plane_2_to_1_general()
|