Home
last modified time | relevance | path

Searched refs:vreinterpret_u8_u32 (Results 1 – 25 of 51) sorted by relevance

123

/external/libaom/libaom/aom_dsp/arm/
Dloopfilter_neon.c37 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(mask_8x8))); in lpf_mask()
41 temp_8x8 = vabd_u8(vreinterpret_u8_u32(p0q0_p1q1.val[0]), in lpf_mask()
42 vreinterpret_u8_u32(p0q0_p1q1.val[1])); in lpf_mask()
67 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(mask_8x8))); in lpf_mask2()
71 temp_8x8 = vabd_u8(vreinterpret_u8_u32(p0q0_p1q1.val[0]), in lpf_mask2()
72 vreinterpret_u8_u32(p0q0_p1q1.val[1])); in lpf_mask2()
95 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(flat_8x8))); in lpf_flat_mask4()
110 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(flat_8x8))); in lpf_flat_mask3()
131 temp_8x8 = vreinterpret_u8_u32(vrev64_u32(vreinterpret_u32_u8(mask_8x8))); in lpf_mask3_chroma()
135 temp_8x8 = vabd_u8(vreinterpret_u8_u32(p0q0_p1q1.val[0]), in lpf_mask3_chroma()
[all …]
Dblend_a64_mask_neon.c206 mask0 = vreinterpretq_s16_u16(vmovl_u8(vreinterpret_u8_u32(tu0))); in aom_lowbd_blend_a64_d16_mask_neon()
207 mask1 = vreinterpretq_s16_u16(vmovl_u8(vreinterpret_u8_u32(tu1))); in aom_lowbd_blend_a64_d16_mask_neon()
423 s0 = vreinterpret_u8_u32(tu0); in aom_lowbd_blend_a64_d16_mask_neon()
424 s1 = vreinterpret_u8_u32(tu1); in aom_lowbd_blend_a64_d16_mask_neon()
425 s2 = vreinterpret_u8_u32(tu2); in aom_lowbd_blend_a64_d16_mask_neon()
426 s3 = vreinterpret_u8_u32(tu3); in aom_lowbd_blend_a64_d16_mask_neon()
/external/libgav1/libgav1/src/dsp/arm/
Dcommon_neon.h220 return vreinterpret_u8_u32(vld1_lane_u32(&temp, zero, 0)); in Load4()
228 return vreinterpret_u8_u32( in Load4()
343 return vreinterpret_u8_u32( in InterleaveLow32()
347 return vreinterpret_u8_u32( in InterleaveLow32()
365 return vreinterpret_u8_u32( in InterleaveHigh32()
369 return vreinterpret_u8_u32( in InterleaveHigh32()
431 const uint8x8x2_t d = {vreinterpret_u8_u32(c.val[0]), in Interleave32()
432 vreinterpret_u8_u32(c.val[1])}; in Interleave32()
439 return vreinterpret_u8_u32(b); in Transpose32()
510 vtrn_u8(vreinterpret_u8_u32(d.val[0]), vreinterpret_u8_u32(d.val[1])); in Transpose4x4()
[all …]
Dintrapred_directional_neon.cc375 StoreLo4(dst, vreinterpret_u8_u32(vget_low_u32(d0.val[0]))); in DirectionalZone3_WxH()
377 StoreLo4(dst, vreinterpret_u8_u32(vget_high_u32(d0.val[0]))); in DirectionalZone3_WxH()
379 StoreLo4(dst, vreinterpret_u8_u32(vget_low_u32(d1.val[0]))); in DirectionalZone3_WxH()
381 StoreLo4(dst, vreinterpret_u8_u32(vget_high_u32(d1.val[0]))); in DirectionalZone3_WxH()
384 StoreLo4(dst, vreinterpret_u8_u32(vget_low_u32(d0.val[1]))); in DirectionalZone3_WxH()
386 StoreLo4(dst, vreinterpret_u8_u32(vget_high_u32(d0.val[1]))); in DirectionalZone3_WxH()
388 StoreLo4(dst, vreinterpret_u8_u32(vget_low_u32(d1.val[1]))); in DirectionalZone3_WxH()
390 StoreLo4(dst, vreinterpret_u8_u32(vget_high_u32(d1.val[1]))); in DirectionalZone3_WxH()
392 vst1_u8(dst, vreinterpret_u8_u32(vget_low_u32(d0.val[0]))); in DirectionalZone3_WxH()
394 vst1_u8(dst, vreinterpret_u8_u32(vget_high_u32(d0.val[0]))); in DirectionalZone3_WxH()
[all …]
/external/skqp/src/core/
DSkBlitRow_D32.cpp163 vsrc = vreinterpret_u8_u32(vld1_u32(src)); in blit_row_s32_blend()
164 vdst = vreinterpret_u8_u32(vld1_u32(dst)); in blit_row_s32_blend()
185 vsrc = vreinterpret_u8_u32(vld1_lane_u32(src, vreinterpret_u32_u8(vsrc), 0)); in blit_row_s32_blend()
186 vdst = vreinterpret_u8_u32(vld1_lane_u32(dst, vreinterpret_u32_u8(vdst), 0)); in blit_row_s32_blend()
208 vsrc = vreinterpret_u8_u32(vld1_lane_u32(src, vreinterpret_u32_u8(vsrc), 0)); in blit_row_s32a_blend()
209 vdst = vreinterpret_u8_u32(vld1_lane_u32(dst, vreinterpret_u32_u8(vdst), 0)); in blit_row_s32a_blend()
241 vsrc = vreinterpret_u8_u32(vld1_u32(src)); in blit_row_s32a_blend()
242 vdst = vreinterpret_u8_u32(vld1_u32(dst)); in blit_row_s32a_blend()
/external/skia/src/core/
DSkBlitRow_D32.cpp162 vsrc = vreinterpret_u8_u32(vld1_u32(src)); in blit_row_s32_blend()
163 vdst = vreinterpret_u8_u32(vld1_u32(dst)); in blit_row_s32_blend()
184 vsrc = vreinterpret_u8_u32(vld1_lane_u32(src, vreinterpret_u32_u8(vsrc), 0)); in blit_row_s32_blend()
185 vdst = vreinterpret_u8_u32(vld1_lane_u32(dst, vreinterpret_u32_u8(vdst), 0)); in blit_row_s32_blend()
207 vsrc = vreinterpret_u8_u32(vld1_lane_u32(src, vreinterpret_u32_u8(vsrc), 0)); in blit_row_s32a_blend()
208 vdst = vreinterpret_u8_u32(vld1_lane_u32(dst, vreinterpret_u32_u8(vdst), 0)); in blit_row_s32a_blend()
240 vsrc = vreinterpret_u8_u32(vld1_u32(src)); in blit_row_s32a_blend()
241 vdst = vreinterpret_u8_u32(vld1_u32(dst)); in blit_row_s32a_blend()
/external/libaom/libaom/av1/common/arm/
Dtranspose_neon.h60 *a0 = vreinterpret_u8_u32(vget_low_u32(d0.val[0])); in transpose_u8_8x8()
61 *a1 = vreinterpret_u8_u32(vget_high_u32(d0.val[0])); in transpose_u8_8x8()
62 *a2 = vreinterpret_u8_u32(vget_low_u32(d1.val[0])); in transpose_u8_8x8()
63 *a3 = vreinterpret_u8_u32(vget_high_u32(d1.val[0])); in transpose_u8_8x8()
64 *a4 = vreinterpret_u8_u32(vget_low_u32(d0.val[1])); in transpose_u8_8x8()
65 *a5 = vreinterpret_u8_u32(vget_high_u32(d0.val[1])); in transpose_u8_8x8()
66 *a6 = vreinterpret_u8_u32(vget_low_u32(d1.val[1])); in transpose_u8_8x8()
67 *a7 = vreinterpret_u8_u32(vget_high_u32(d1.val[1])); in transpose_u8_8x8()
126 vtrn_u8(vreinterpret_u8_u32(c0.val[0]), vreinterpret_u8_u32(c0.val[1])); in transpose_u8_4x4()
Dblend_a64_hmask_neon.c86 const uint8x8_t m = vreinterpret_u8_u32(vld1_dup_u32((uint32_t *)mask)); in aom_blend_a64_hmask_neon()
94 tmp0 = vreinterpret_u8_u32(tmp0_32); in aom_blend_a64_hmask_neon()
96 tmp1 = vreinterpret_u8_u32(tmp1_32); in aom_blend_a64_hmask_neon()
Dblend_a64_vmask_neon.c93 tmp0 = vreinterpret_u8_u32(tmp0_32); in aom_blend_a64_vmask_neon()
95 tmp1 = vreinterpret_u8_u32(tmp1_32); in aom_blend_a64_vmask_neon()
/external/libhevc/encoder/arm/
Dihevce_coarse_layer_sad_neon.c164 src2[0] = vreinterpret_u8_u32(a[0]); in hme_store_4x4_sads_high_speed_neon()
165 src2[1] = vreinterpret_u8_u32(a[1]); in hme_store_4x4_sads_high_speed_neon()
166 src2[2] = vreinterpret_u8_u32(a[2]); in hme_store_4x4_sads_high_speed_neon()
167 src2[3] = vreinterpret_u8_u32(a[3]); in hme_store_4x4_sads_high_speed_neon()
174 src = vcombine_u8(vreinterpret_u8_u32(l.val[0]), vreinterpret_u8_u32(m.val[0])); in hme_store_4x4_sads_high_speed_neon()
315 src2[0] = vreinterpret_u8_u32(a[0]); in hme_store_4x4_sads_high_quality_neon()
316 src2[1] = vreinterpret_u8_u32(a[1]); in hme_store_4x4_sads_high_quality_neon()
317 src2[2] = vreinterpret_u8_u32(a[2]); in hme_store_4x4_sads_high_quality_neon()
318 src2[3] = vreinterpret_u8_u32(a[3]); in hme_store_4x4_sads_high_quality_neon()
325 src = vcombine_u8(vreinterpret_u8_u32(l.val[0]), vreinterpret_u8_u32(m.val[0])); in hme_store_4x4_sads_high_quality_neon()
/external/libvpx/libvpx/vpx_dsp/arm/
Dvpx_convolve_avg_neon.c36 dd0 = vreinterpret_u8_u32( in vpx_convolve_avg_neon()
38 dd0 = vreinterpret_u8_u32(vld1_lane_u32( in vpx_convolve_avg_neon()
40 dd0 = vrhadd_u8(vreinterpret_u8_u32(s01.val[0]), dd0); in vpx_convolve_avg_neon()
Dtranspose_neon.h51 b0.val[0] = vcombine_u8(vreinterpret_u8_u32(vget_low_u32(a0)), in vpx_vtrnq_u64_to_u8()
52 vreinterpret_u8_u32(vget_low_u32(a1))); in vpx_vtrnq_u64_to_u8()
53 b0.val[1] = vcombine_u8(vreinterpret_u8_u32(vget_high_u32(a0)), in vpx_vtrnq_u64_to_u8()
54 vreinterpret_u8_u32(vget_high_u32(a1))); in vpx_vtrnq_u64_to_u8()
90 vtrn_u8(vreinterpret_u8_u32(c0.val[0]), vreinterpret_u8_u32(c0.val[1])); in transpose_u8_4x4()
561 *a0 = vreinterpret_u8_u32(vget_low_u32(d0.val[0])); in transpose_u8_8x8()
562 *a1 = vreinterpret_u8_u32(vget_high_u32(d0.val[0])); in transpose_u8_8x8()
563 *a2 = vreinterpret_u8_u32(vget_low_u32(d1.val[0])); in transpose_u8_8x8()
564 *a3 = vreinterpret_u8_u32(vget_high_u32(d1.val[0])); in transpose_u8_8x8()
565 *a4 = vreinterpret_u8_u32(vget_low_u32(d0.val[1])); in transpose_u8_8x8()
[all …]
Dvpx_convolve8_avg_vert_filter_type1_neon.asm399 ; vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1);
406 ; vreinterpret_u8_u32(src_tmp1), coeffabs_0);
412 ; vreinterpret_u8_u32(src_tmp3), coeffabs_2);
418 ; vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3);
423 ; vreinterpret_u8_u32(src_tmp1), coeffabs_4);
430 ; vreinterpret_u8_u32(src_tmp2), coeffabs_5);
437 ; vreinterpret_u8_u32(src_tmp3), coeffabs_6);
444 ; vreinterpret_u8_u32(src_tmp4), coeffabs_7);
Dvpx_convolve8_avg_vert_filter_type2_neon.asm400 ; vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1);
407 ; vreinterpret_u8_u32(src_tmp1), coeffabs_0);
413 ; vreinterpret_u8_u32(src_tmp3), coeffabs_2);
419 ; vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3);
424 ; vreinterpret_u8_u32(src_tmp1), coeffabs_4);
431 ; vreinterpret_u8_u32(src_tmp2), coeffabs_5);
438 ; vreinterpret_u8_u32(src_tmp3), coeffabs_6);
445 ; vreinterpret_u8_u32(src_tmp4), coeffabs_7);
Dvpx_convolve8_vert_filter_type1_neon.asm376 ; vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1);
384 ; vreinterpret_u8_u32(src_tmp1), coeffabs_0);
390 ; vreinterpret_u8_u32(src_tmp3), coeffabs_2);
396 ; vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3);
401 ; vreinterpret_u8_u32(src_tmp1), coeffabs_4);
408 ; vreinterpret_u8_u32(src_tmp2), coeffabs_5);
415 ; vreinterpret_u8_u32(src_tmp3), coeffabs_6);
422 ; vreinterpret_u8_u32(src_tmp4), coeffabs_7);
Dvpx_convolve8_vert_filter_type2_neon.asm375 ; vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1);
382 ; vreinterpret_u8_u32(src_tmp1), coeffabs_0);
388 ; vreinterpret_u8_u32(src_tmp3), coeffabs_2);
394 ; vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3);
399 ; vreinterpret_u8_u32(src_tmp1), coeffabs_4);
406 ; vreinterpret_u8_u32(src_tmp2), coeffabs_5);
413 ; vreinterpret_u8_u32(src_tmp3), coeffabs_6);
420 ; vreinterpret_u8_u32(src_tmp4), coeffabs_7);
Dmem_neon.h107 return vreinterpret_u8_u32(a_u32); in load_unaligned_u8()
170 return vreinterpret_u8_u32(a); in load_u8()
Didct4x4_1_add_neon.c25 a = vaddw_u8(vreinterpretq_u16_s16(res), vreinterpret_u8_u32(*d)); in idct4x4_1_add_kernel()
/external/libhevc/common/arm/
Dihevc_intra_pred_filters_neon_intr.c1250 sub_val = vsubl_u8(vreinterpret_u8_u32(pu1_ref_val1), dup_sub); in ihevc_intra_pred_luma_horz_neonintr()
1259 dup_val = vdup_lane_u8(vreinterpret_u8_u32(pu1_ref_val2), 2); in ihevc_intra_pred_luma_horz_neonintr()
1263 dup_val = vdup_lane_u8(vreinterpret_u8_u32(pu1_ref_val2), 1); in ihevc_intra_pred_luma_horz_neonintr()
1267 dup_val = vdup_lane_u8(vreinterpret_u8_u32(pu1_ref_val2), 0); in ihevc_intra_pred_luma_horz_neonintr()
1488 sub_val = vsubl_u8(vreinterpret_u8_u32(src_val1), dup_2_sub); in ihevc_intra_pred_luma_ver_neonintr()
1495 vext_val = vext_u8(round_val, vreinterpret_u8_u32(src_val2), 7); in ihevc_intra_pred_luma_ver_neonintr()
1501 … vext_val = vext_u8(vreinterpret_u8_u64(shift_val2), vreinterpret_u8_u32(src_val2), 7); in ihevc_intra_pred_luma_ver_neonintr()
1507 … vext_val = vext_u8(vreinterpret_u8_u64(shift_val2), vreinterpret_u8_u32(src_val2), 7); in ihevc_intra_pred_luma_ver_neonintr()
1513 … vext_val = vext_u8(vreinterpret_u8_u64(shift_val2), vreinterpret_u8_u32(src_val2), 7); in ihevc_intra_pred_luma_ver_neonintr()
1526 vst1_u8(pu1_dst_val2, vreinterpret_u8_u32(src_val3)); in ihevc_intra_pred_luma_ver_neonintr()
[all …]
Dihevc_inter_pred_filters_luma_vert.s440 …vmull.u8 q0,d5,d23 @mul_res1 = vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffa…
445 …vmlsl.u8 q0,d4,d22 @mul_res1 = vmlsl_u8(mul_res1, vreinterpret_u8_u32(src_tmp…
449 …vmlsl.u8 q0,d6,d24 @mul_res1 = vmlsl_u8(mul_res1, vreinterpret_u8_u32(src_tmp…
453 …vmull.u8 q1,d7,d25 @mul_res2 = vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffa…
456 …vmlal.u8 q0,d4,d26 @mul_res1 = vmlal_u8(mul_res1, vreinterpret_u8_u32(src_tmp…
461 …vmlsl.u8 q1,d5,d27 @mul_res2 = vmlsl_u8(mul_res2, vreinterpret_u8_u32(src_tmp…
466 …vmlal.u8 q0,d6,d28 @mul_res1 = vmlal_u8(mul_res1, vreinterpret_u8_u32(src_tmp…
471 …vmlsl.u8 q1,d7,d29 @mul_res2 = vmlsl_u8(mul_res2, vreinterpret_u8_u32(src_tmp…
888 …vmull.u8 q0,d5,d23 @mul_res1 = vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffa…
893 …vmlsl.u8 q0,d4,d22 @mul_res1 = vmlsl_u8(mul_res1, vreinterpret_u8_u32(src_tmp…
[all …]
/external/libvpx/config/arm-neon/vpx_dsp/arm/
Dvpx_convolve8_avg_vert_filter_type2_neon.asm.S406 @ vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1);
413 @ vreinterpret_u8_u32(src_tmp1), coeffabs_0);
419 @ vreinterpret_u8_u32(src_tmp3), coeffabs_2);
425 @ vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3);
430 @ vreinterpret_u8_u32(src_tmp1), coeffabs_4);
437 @ vreinterpret_u8_u32(src_tmp2), coeffabs_5);
444 @ vreinterpret_u8_u32(src_tmp3), coeffabs_6);
451 @ vreinterpret_u8_u32(src_tmp4), coeffabs_7);
Dvpx_convolve8_avg_vert_filter_type1_neon.asm.S405 @ vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1);
412 @ vreinterpret_u8_u32(src_tmp1), coeffabs_0);
418 @ vreinterpret_u8_u32(src_tmp3), coeffabs_2);
424 @ vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3);
429 @ vreinterpret_u8_u32(src_tmp1), coeffabs_4);
436 @ vreinterpret_u8_u32(src_tmp2), coeffabs_5);
443 @ vreinterpret_u8_u32(src_tmp3), coeffabs_6);
450 @ vreinterpret_u8_u32(src_tmp4), coeffabs_7);
Dvpx_convolve8_vert_filter_type1_neon.asm.S382 @ vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1);
390 @ vreinterpret_u8_u32(src_tmp1), coeffabs_0);
396 @ vreinterpret_u8_u32(src_tmp3), coeffabs_2);
402 @ vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3);
407 @ vreinterpret_u8_u32(src_tmp1), coeffabs_4);
414 @ vreinterpret_u8_u32(src_tmp2), coeffabs_5);
421 @ vreinterpret_u8_u32(src_tmp3), coeffabs_6);
428 @ vreinterpret_u8_u32(src_tmp4), coeffabs_7);
Dvpx_convolve8_vert_filter_type2_neon.asm.S381 @ vmull_u8(vreinterpret_u8_u32(src_tmp2), coeffabs_1);
388 @ vreinterpret_u8_u32(src_tmp1), coeffabs_0);
394 @ vreinterpret_u8_u32(src_tmp3), coeffabs_2);
400 @ vmull_u8(vreinterpret_u8_u32(src_tmp4), coeffabs_3);
405 @ vreinterpret_u8_u32(src_tmp1), coeffabs_4);
412 @ vreinterpret_u8_u32(src_tmp2), coeffabs_5);
419 @ vreinterpret_u8_u32(src_tmp3), coeffabs_6);
426 @ vreinterpret_u8_u32(src_tmp4), coeffabs_7);
/external/libvpx/libvpx/vp8/common/arm/neon/
Ddc_only_idct_add_neon.c33 q1u16 = vaddw_u8(qAdd, vreinterpret_u8_u32(d2u32)); in vp8_dc_only_idct_add_neon()

123