/external/libaom/libaom/aom_dsp/arm/ |
D | loopfilter_neon.c | 37 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(mask_8x8))); in lpf_mask() 41 temp_8x8 = vabd_u8(vreinterpret_u8_u32(p0q0_p1q1.val[0]), in lpf_mask() 42 vreinterpret_u8_u32(p0q0_p1q1.val[1])); in lpf_mask() 67 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(mask_8x8))); in lpf_mask2() 71 temp_8x8 = vabd_u8(vreinterpret_u8_u32(p0q0_p1q1.val[0]), in lpf_mask2() 72 vreinterpret_u8_u32(p0q0_p1q1.val[1])); in lpf_mask2() 95 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(flat_8x8))); in lpf_flat_mask4() 110 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(flat_8x8))); in lpf_flat_mask3() 131 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(mask_8x8))); in lpf_mask3_chroma() 135 temp_8x8 = vabd_u8(vreinterpret_u8_u32(p0q0_p1q1.val[0]), in lpf_mask3_chroma() [all …]
|
D | blend_a64_mask_neon.c | 206 mask0 = vreinterpretq_s16_u16(vmovl_u8(vreinterpret_u8_u32(tu0))); in aom_lowbd_blend_a64_d16_mask_neon() 207 mask1 = vreinterpretq_s16_u16(vmovl_u8(vreinterpret_u8_u32(tu1))); in aom_lowbd_blend_a64_d16_mask_neon() 423 s0 = vreinterpret_u8_u32(tu0); in aom_lowbd_blend_a64_d16_mask_neon() 424 s1 = vreinterpret_u8_u32(tu1); in aom_lowbd_blend_a64_d16_mask_neon() 425 s2 = vreinterpret_u8_u32(tu2); in aom_lowbd_blend_a64_d16_mask_neon() 426 s3 = vreinterpret_u8_u32(tu3); in aom_lowbd_blend_a64_d16_mask_neon()
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | common_neon.h | 220 return vreinterpret_u8_u32(vld1_lane_u32(&temp, zero, 0)); in Load4() 228 return vreinterpret_u8_u32( in Load4() 343 return vreinterpret_u8_u32( in InterleaveLow32() 347 return vreinterpret_u8_u32( in InterleaveLow32() 365 return vreinterpret_u8_u32( in InterleaveHigh32() 369 return vreinterpret_u8_u32( in InterleaveHigh32() 431 const uint8x8x2_t d = {vreinterpret_u8_u32(c.val[0]), in Interleave32() 432 vreinterpret_u8_u32(c.val[1])}; in Interleave32() 439 return vreinterpret_u8_u32(b); in Transpose32() 510 vtrn_u8(vreinterpret_u8_u32(d.val[0]), vreinterpret_u8_u32(d.val[1])); in Transpose4x4() [all …]
|
D | intrapred_directional_neon.cc | 375 StoreLo4(dst, vreinterpret_u8_u32(vget_low_u32(d0.val[0]))); in DirectionalZone3_WxH() 377 StoreLo4(dst, vreinterpret_u8_u32(vget_high_u32(d0.val[0]))); in DirectionalZone3_WxH() 379 StoreLo4(dst, vreinterpret_u8_u32(vget_low_u32(d1.val[0]))); in DirectionalZone3_WxH() 381 StoreLo4(dst, vreinterpret_u8_u32(vget_high_u32(d1.val[0]))); in DirectionalZone3_WxH() 384 StoreLo4(dst, vreinterpret_u8_u32(vget_low_u32(d0.val[1]))); in DirectionalZone3_WxH() 386 StoreLo4(dst, vreinterpret_u8_u32(vget_high_u32(d0.val[1]))); in DirectionalZone3_WxH() 388 StoreLo4(dst, vreinterpret_u8_u32(vget_low_u32(d1.val[1]))); in DirectionalZone3_WxH() 390 StoreLo4(dst, vreinterpret_u8_u32(vget_high_u32(d1.val[1]))); in DirectionalZone3_WxH() 392 vst1_u8(dst, vreinterpret_u8_u32(vget_low_u32(d0.val[0]))); in DirectionalZone3_WxH() 394 vst1_u8(dst, vreinterpret_u8_u32(vget_high_u32(d0.val[0]))); in DirectionalZone3_WxH() [all …]
|
/external/skqp/src/core/ |
D | SkBlitRow_D32.cpp | 163 vsrc = vreinterpret_u8_u32(vld1_u32(src)); in blit_row_s32_blend() 164 vdst = vreinterpret_u8_u32(vld1_u32(dst)); in blit_row_s32_blend() 185 vsrc = vreinterpret_u8_u32(vld1_lane_u32(src, vreinterpret_u32_u8(vsrc), 0)); in blit_row_s32_blend() 186 vdst = vreinterpret_u8_u32(vld1_lane_u32(dst, vreinterpret_u32_u8(vdst), 0)); in blit_row_s32_blend() 208 vsrc = vreinterpret_u8_u32(vld1_lane_u32(src, vreinterpret_u32_u8(vsrc), 0)); in blit_row_s32a_blend() 209 vdst = vreinterpret_u8_u32(vld1_lane_u32(dst, vreinterpret_u32_u8(vdst), 0)); in blit_row_s32a_blend() 241 vsrc = vreinterpret_u8_u32(vld1_u32(src)); in blit_row_s32a_blend() 242 vdst = vreinterpret_u8_u32(vld1_u32(dst)); in blit_row_s32a_blend()
|
/external/skia/src/core/ |
D | SkBlitRow_D32.cpp | 162 vsrc = vreinterpret_u8_u32(vld1_u32(src)); in blit_row_s32_blend() 163 vdst = vreinterpret_u8_u32(vld1_u32(dst)); in blit_row_s32_blend() 184 vsrc = vreinterpret_u8_u32(vld1_lane_u32(src, vreinterpret_u32_u8(vsrc), 0)); in blit_row_s32_blend() 185 vdst = vreinterpret_u8_u32(vld1_lane_u32(dst, vreinterpret_u32_u8(vdst), 0)); in blit_row_s32_blend() 207 vsrc = vreinterpret_u8_u32(vld1_lane_u32(src, vreinterpret_u32_u8(vsrc), 0)); in blit_row_s32a_blend() 208 vdst = vreinterpret_u8_u32(vld1_lane_u32(dst, vreinterpret_u32_u8(vdst), 0)); in blit_row_s32a_blend() 240 vsrc = vreinterpret_u8_u32(vld1_u32(src)); in blit_row_s32a_blend() 241 vdst = vreinterpret_u8_u32(vld1_u32(dst)); in blit_row_s32a_blend()
|
/external/libaom/libaom/av1/common/arm/ |
D | transpose_neon.h | 60 *a0 = vreinterpret_u8_u32(vget_low_u32(d0.val[0])); in transpose_u8_8x8() 61 *a1 = vreinterpret_u8_u32(vget_high_u32(d0.val[0])); in transpose_u8_8x8() 62 *a2 = vreinterpret_u8_u32(vget_low_u32(d1.val[0])); in transpose_u8_8x8() 63 *a3 = vreinterpret_u8_u32(vget_high_u32(d1.val[0])); in transpose_u8_8x8() 64 *a4 = vreinterpret_u8_u32(vget_low_u32(d0.val[1])); in transpose_u8_8x8() 65 *a5 = vreinterpret_u8_u32(vget_high_u32(d0.val[1])); in transpose_u8_8x8() 66 *a6 = vreinterpret_u8_u32(vget_low_u32(d1.val[1])); in transpose_u8_8x8() 67 *a7 = vreinterpret_u8_u32(vget_high_u32(d1.val[1])); in transpose_u8_8x8() 126 vtrn_u8(vreinterpret_u8_u32(c0.val[0]), vreinterpret_u8_u32(c0.val[1])); in transpose_u8_4x4()
|
D | blend_a64_hmask_neon.c | 86 const uint8x8_t m = vreinterpret_u8_u32(vld1_dup_u32((uint32_t *)mask)); in aom_blend_a64_hmask_neon() 94 tmp0 = vreinterpret_u8_u32(tmp0_32); in aom_blend_a64_hmask_neon() 96 tmp1 = vreinterpret_u8_u32(tmp1_32); in aom_blend_a64_hmask_neon()
|
D | blend_a64_vmask_neon.c | 93 tmp0 = vreinterpret_u8_u32(tmp0_32); in aom_blend_a64_vmask_neon() 95 tmp1 = vreinterpret_u8_u32(tmp1_32); in aom_blend_a64_vmask_neon()
|
/external/libhevc/encoder/arm/ |
D | ihevce_coarse_layer_sad_neon.c | 164 src2[0] = vreinterpret_u8_u32(a[0]); in hme_store_4x4_sads_high_speed_neon() 165 src2[1] = vreinterpret_u8_u32(a[1]); in hme_store_4x4_sads_high_speed_neon() 166 src2[2] = vreinterpret_u8_u32(a[2]); in hme_store_4x4_sads_high_speed_neon() 167 src2[3] = vreinterpret_u8_u32(a[3]); in hme_store_4x4_sads_high_speed_neon() 174 src = vcombine_u8(vreinterpret_u8_u32(l.val[0]), vreinterpret_u8_u32(m.val[0])); in hme_store_4x4_sads_high_speed_neon() 315 src2[0] = vreinterpret_u8_u32(a[0]); in hme_store_4x4_sads_high_quality_neon() 316 src2[1] = vreinterpret_u8_u32(a[1]); in hme_store_4x4_sads_high_quality_neon() 317 src2[2] = vreinterpret_u8_u32(a[2]); in hme_store_4x4_sads_high_quality_neon() 318 src2[3] = vreinterpret_u8_u32(a[3]); in hme_store_4x4_sads_high_quality_neon() 325 src = vcombine_u8(vreinterpret_u8_u32(l.val[0]), vreinterpret_u8_u32(m.val[0])); in hme_store_4x4_sads_high_quality_neon()
|
/external/libvpx/libvpx/vpx_dsp/arm/ |
D | vpx_convolve_avg_neon.c | 36 dd0 = vreinterpret_u8_u32( in vpx_convolve_avg_neon() 38 dd0 = vreinterpret_u8_u32(vld1_lane_u32( in vpx_convolve_avg_neon() 40 dd0 = vrhadd_u8(vreinterpret_u8_u32(s01.val[0]), dd0); in vpx_convolve_avg_neon()
|
D | transpose_neon.h | 51 b0.val[0] = vcombine_u8(vreinterpret_u8_u32(vget_low_u32(a0)), in vpx_vtrnq_u64_to_u8() 52 vreinterpret_u8_u32(vget_low_u32(a1))); in vpx_vtrnq_u64_to_u8() 53 b0.val[1] = vcombine_u8(vreinterpret_u8_u32(vget_high_u32(a0)), in vpx_vtrnq_u64_to_u8() 54 vreinterpret_u8_u32(vget_high_u32(a1))); in vpx_vtrnq_u64_to_u8() 90 vtrn_u8(vreinterpret_u8_u32(c0.val[0]), vreinterpret_u8_u32(c0.val[1])); in transpose_u8_4x4() 561 *a0 = vreinterpret_u8_u32(vget_low_u32(d0.val[0])); in transpose_u8_8x8() 562 *a1 = vreinterpret_u8_u32(vget_high_u32(d0.val[0])); in transpose_u8_8x8() 563 *a2 = vreinterpret_u8_u32(vget_low_u32(d1.val[0])); in transpose_u8_8x8() 564 *a3 = vreinterpret_u8_u32(vget_high_u32(d1.val[0])); in transpose_u8_8x8() 565 *a4 = vreinterpret_u8_u32(vget_low_u32(d0.val[1])); in transpose_u8_8x8() [all …]
|
D | vpx_convolve8_avg_vert_filter_type1_neon.asm | 399 ; vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1); 406 ; vreinterpret_u8_u32(src_tmp1), coeffabs_0); 412 ; vreinterpret_u8_u32(src_tmp3), coeffabs_2); 418 ; vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3); 423 ; vreinterpret_u8_u32(src_tmp1), coeffabs_4); 430 ; vreinterpret_u8_u32(src_tmp2), coeffabs_5); 437 ; vreinterpret_u8_u32(src_tmp3), coeffabs_6); 444 ; vreinterpret_u8_u32(src_tmp4), coeffabs_7);
|
D | vpx_convolve8_avg_vert_filter_type2_neon.asm | 400 ; vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1); 407 ; vreinterpret_u8_u32(src_tmp1), coeffabs_0); 413 ; vreinterpret_u8_u32(src_tmp3), coeffabs_2); 419 ; vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3); 424 ; vreinterpret_u8_u32(src_tmp1), coeffabs_4); 431 ; vreinterpret_u8_u32(src_tmp2), coeffabs_5); 438 ; vreinterpret_u8_u32(src_tmp3), coeffabs_6); 445 ; vreinterpret_u8_u32(src_tmp4), coeffabs_7);
|
D | vpx_convolve8_vert_filter_type1_neon.asm | 376 ; vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1); 384 ; vreinterpret_u8_u32(src_tmp1), coeffabs_0); 390 ; vreinterpret_u8_u32(src_tmp3), coeffabs_2); 396 ; vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3); 401 ; vreinterpret_u8_u32(src_tmp1), coeffabs_4); 408 ; vreinterpret_u8_u32(src_tmp2), coeffabs_5); 415 ; vreinterpret_u8_u32(src_tmp3), coeffabs_6); 422 ; vreinterpret_u8_u32(src_tmp4), coeffabs_7);
|
D | vpx_convolve8_vert_filter_type2_neon.asm | 375 ; vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1); 382 ; vreinterpret_u8_u32(src_tmp1), coeffabs_0); 388 ; vreinterpret_u8_u32(src_tmp3), coeffabs_2); 394 ; vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3); 399 ; vreinterpret_u8_u32(src_tmp1), coeffabs_4); 406 ; vreinterpret_u8_u32(src_tmp2), coeffabs_5); 413 ; vreinterpret_u8_u32(src_tmp3), coeffabs_6); 420 ; vreinterpret_u8_u32(src_tmp4), coeffabs_7);
|
D | mem_neon.h | 107 return vreinterpret_u8_u32(a_u32); in load_unaligned_u8() 170 return vreinterpret_u8_u32(a); in load_u8()
|
D | idct4x4_1_add_neon.c | 25 a = vaddw_u8(vreinterpretq_u16_s16(res), vreinterpret_u8_u32(*d)); in idct4x4_1_add_kernel()
|
/external/libhevc/common/arm/ |
D | ihevc_intra_pred_filters_neon_intr.c | 1250 sub_val = vsubl_u8(vreinterpret_u8_u32(pu1_ref_val1), dup_sub); in ihevc_intra_pred_luma_horz_neonintr() 1259 dup_val = vdup_lane_u8(vreinterpret_u8_u32(pu1_ref_val2), 2); in ihevc_intra_pred_luma_horz_neonintr() 1263 dup_val = vdup_lane_u8(vreinterpret_u8_u32(pu1_ref_val2), 1); in ihevc_intra_pred_luma_horz_neonintr() 1267 dup_val = vdup_lane_u8(vreinterpret_u8_u32(pu1_ref_val2), 0); in ihevc_intra_pred_luma_horz_neonintr() 1488 sub_val = vsubl_u8(vreinterpret_u8_u32(src_val1), dup_2_sub); in ihevc_intra_pred_luma_ver_neonintr() 1495 vext_val = vext_u8(round_val, vreinterpret_u8_u32(src_val2), 7); in ihevc_intra_pred_luma_ver_neonintr() 1501 … vext_val = vext_u8(vreinterpret_u8_u64(shift_val2), vreinterpret_u8_u32(src_val2), 7); in ihevc_intra_pred_luma_ver_neonintr() 1507 … vext_val = vext_u8(vreinterpret_u8_u64(shift_val2), vreinterpret_u8_u32(src_val2), 7); in ihevc_intra_pred_luma_ver_neonintr() 1513 … vext_val = vext_u8(vreinterpret_u8_u64(shift_val2), vreinterpret_u8_u32(src_val2), 7); in ihevc_intra_pred_luma_ver_neonintr() 1526 vst1_u8(pu1_dst_val2, vreinterpret_u8_u32(src_val3)); in ihevc_intra_pred_luma_ver_neonintr() [all …]
|
D | ihevc_inter_pred_filters_luma_vert.s | 440 …vmull.u8 q0,d5,d23 @mul_res1 = vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffa… 445 …vmlsl.u8 q0,d4,d22 @mul_res1 = vmlsl_u8(mul_res1, vreinterpret_u8_u32(src_tmp… 449 …vmlsl.u8 q0,d6,d24 @mul_res1 = vmlsl_u8(mul_res1, vreinterpret_u8_u32(src_tmp… 453 …vmull.u8 q1,d7,d25 @mul_res2 = vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffa… 456 …vmlal.u8 q0,d4,d26 @mul_res1 = vmlal_u8(mul_res1, vreinterpret_u8_u32(src_tmp… 461 …vmlsl.u8 q1,d5,d27 @mul_res2 = vmlsl_u8(mul_res2, vreinterpret_u8_u32(src_tmp… 466 …vmlal.u8 q0,d6,d28 @mul_res1 = vmlal_u8(mul_res1, vreinterpret_u8_u32(src_tmp… 471 …vmlsl.u8 q1,d7,d29 @mul_res2 = vmlsl_u8(mul_res2, vreinterpret_u8_u32(src_tmp… 888 …vmull.u8 q0,d5,d23 @mul_res1 = vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffa… 893 …vmlsl.u8 q0,d4,d22 @mul_res1 = vmlsl_u8(mul_res1, vreinterpret_u8_u32(src_tmp… [all …]
|
/external/libvpx/config/arm-neon/vpx_dsp/arm/ |
D | vpx_convolve8_avg_vert_filter_type2_neon.asm.S | 406 @ vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1); 413 @ vreinterpret_u8_u32(src_tmp1), coeffabs_0); 419 @ vreinterpret_u8_u32(src_tmp3), coeffabs_2); 425 @ vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3); 430 @ vreinterpret_u8_u32(src_tmp1), coeffabs_4); 437 @ vreinterpret_u8_u32(src_tmp2), coeffabs_5); 444 @ vreinterpret_u8_u32(src_tmp3), coeffabs_6); 451 @ vreinterpret_u8_u32(src_tmp4), coeffabs_7);
|
D | vpx_convolve8_avg_vert_filter_type1_neon.asm.S | 405 @ vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1); 412 @ vreinterpret_u8_u32(src_tmp1), coeffabs_0); 418 @ vreinterpret_u8_u32(src_tmp3), coeffabs_2); 424 @ vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3); 429 @ vreinterpret_u8_u32(src_tmp1), coeffabs_4); 436 @ vreinterpret_u8_u32(src_tmp2), coeffabs_5); 443 @ vreinterpret_u8_u32(src_tmp3), coeffabs_6); 450 @ vreinterpret_u8_u32(src_tmp4), coeffabs_7);
|
D | vpx_convolve8_vert_filter_type1_neon.asm.S | 382 @ vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1); 390 @ vreinterpret_u8_u32(src_tmp1), coeffabs_0); 396 @ vreinterpret_u8_u32(src_tmp3), coeffabs_2); 402 @ vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3); 407 @ vreinterpret_u8_u32(src_tmp1), coeffabs_4); 414 @ vreinterpret_u8_u32(src_tmp2), coeffabs_5); 421 @ vreinterpret_u8_u32(src_tmp3), coeffabs_6); 428 @ vreinterpret_u8_u32(src_tmp4), coeffabs_7);
|
D | vpx_convolve8_vert_filter_type2_neon.asm.S | 381 @ vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1); 388 @ vreinterpret_u8_u32(src_tmp1), coeffabs_0); 394 @ vreinterpret_u8_u32(src_tmp3), coeffabs_2); 400 @ vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3); 405 @ vreinterpret_u8_u32(src_tmp1), coeffabs_4); 412 @ vreinterpret_u8_u32(src_tmp2), coeffabs_5); 419 @ vreinterpret_u8_u32(src_tmp3), coeffabs_6); 426 @ vreinterpret_u8_u32(src_tmp4), coeffabs_7);
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | dc_only_idct_add_neon.c | 33 q1u16 = vaddw_u8(qAdd, vreinterpret_u8_u32(d2u32)); in vp8_dc_only_idct_add_neon()
|