/external/libvpx/libvpx/vpx_dsp/arm/ |
D | transpose_neon.h | 168 const uint32x4x2_t b0 = in transpose_u16_4x4q() 451 const uint32x4x2_t c0 = vtrnq_u32(vreinterpretq_u32_u16(b0.val[0]), in transpose_u16_8x4() 453 const uint32x4x2_t c1 = vtrnq_u32(vreinterpretq_u32_u16(b0.val[1]), in transpose_u16_8x4() 556 const uint32x4x2_t d0 = vuzpq_u32(vreinterpretq_u32_u16(c0.val[0]), in transpose_u8_8x8() 558 const uint32x4x2_t d1 = vuzpq_u32(vreinterpretq_u32_u16(c0.val[1]), in transpose_u8_8x8() 680 const uint32x4x2_t c0 = vtrnq_u32(vreinterpretq_u32_u16(b0.val[0]), in transpose_u16_8x8() 682 const uint32x4x2_t c1 = vtrnq_u32(vreinterpretq_u32_u16(b0.val[1]), in transpose_u16_8x8() 684 const uint32x4x2_t c2 = vtrnq_u32(vreinterpretq_u32_u16(b2.val[0]), in transpose_u16_8x8() 686 const uint32x4x2_t c3 = vtrnq_u32(vreinterpretq_u32_u16(b2.val[1]), in transpose_u16_8x8() 847 const uint32x4x2_t d0 = vtrnq_u32(vreinterpretq_u32_u16(c0.val[0]), in transpose_u8_16x8() [all …]
|
/external/libgav1/libgav1/src/dsp/arm/ |
D | loop_restoration_neon.cc | 633 inline uint32x4x2_t Sum3_32(const uint32x4x2_t src[3]) { in Sum3_32() 634 uint32x4x2_t d; in Sum3_32() 666 inline uint32x4x2_t Sum5_32(const uint32x4x2_t src[5]) { in Sum5_32() 667 uint32x4x2_t d; in Sum5_32() 688 inline uint32x4x2_t Sum3WHorizontal(const uint16x8x2_t src) { in Sum3WHorizontal() 690 uint32x4x2_t sum; in Sum3WHorizontal() 706 inline uint32x4x2_t Sum5WHorizontal(const uint16x8x2_t src) { in Sum5WHorizontal() 709 uint32x4x2_t sum; in Sum5WHorizontal() 725 uint32x4x2_t* const row_sq3, uint32x4x2_t* const row_sq5) { in SumHorizontal() 752 inline uint32x4x2_t Sum343W(const uint16x8x2_t src) { in Sum343W() [all …]
|
D | common_neon.h | 575 const uint32x4x2_t d0 = vuzpq_u32(vreinterpretq_u32_u16(c0.val[0]), in Transpose8x8() 577 const uint32x4x2_t d1 = vuzpq_u32(vreinterpretq_u32_u16(c0.val[1]), in Transpose8x8() 646 const uint32x4x2_t c0 = vtrnq_u32(vreinterpretq_u32_u16(b0.val[0]), in Transpose8x8() 648 const uint32x4x2_t c1 = vtrnq_u32(vreinterpretq_u32_u16(b0.val[1]), in Transpose8x8() 650 const uint32x4x2_t c2 = vtrnq_u32(vreinterpretq_u32_u16(b2.val[0]), in Transpose8x8() 652 const uint32x4x2_t c3 = vtrnq_u32(vreinterpretq_u32_u16(b2.val[1]), in Transpose8x8()
|
D | intrapred_directional_neon.cc | 369 const uint32x4x2_t d0 = vuzpq_u32(vreinterpretq_u32_u16(c0.val[0]), in DirectionalZone3_WxH() 371 const uint32x4x2_t d1 = vuzpq_u32(vreinterpretq_u32_u16(c0.val[1]), in DirectionalZone3_WxH()
|
/external/libaom/libaom/av1/common/arm/ |
D | transpose_neon.h | 55 const uint32x4x2_t d0 = vuzpq_u32(vreinterpretq_u32_u16(c0.val[0]), in transpose_u8_8x8() 57 const uint32x4x2_t d1 = vuzpq_u32(vreinterpretq_u32_u16(c0.val[1]), in transpose_u8_8x8() 356 const uint32x4x2_t c0 = vtrnq_u32(vreinterpretq_u32_u16(b0.val[0]), in transpose_u16_8x8() 358 const uint32x4x2_t c1 = vtrnq_u32(vreinterpretq_u32_u16(b0.val[1]), in transpose_u16_8x8() 360 const uint32x4x2_t c2 = vtrnq_u32(vreinterpretq_u32_u16(b2.val[0]), in transpose_u16_8x8() 362 const uint32x4x2_t c3 = vtrnq_u32(vreinterpretq_u32_u16(b2.val[1]), in transpose_u16_8x8()
|
/external/XNNPACK/src/x32-zip/ |
D | xm-neon.c | 46 const uint32x4x2_t vxy = vzipq_u32(vx, vy); in xnn_x32_zip_xm_ukernel__neon() 47 const uint32x4x2_t vzw = vzipq_u32(vz, vw); in xnn_x32_zip_xm_ukernel__neon()
|
D | x2-neon.c | 26 uint32x4x2_t vxy; in xnn_x32_zip_x2_ukernel__neon()
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/ARM/ |
D | arm-vlddup.ll | 24 %struct.uint32x4x2_t = type { <4 x i32>, <4 x i32> } 49 declare %struct.uint32x4x2_t @llvm.arm.neon.vld2dup.v4i32.p0i8(i8*, i32) 167 define %struct.uint32x4x2_t @test_vld2q_dup_u32(i8* %src) { 169 %tmp = tail call %struct.uint32x4x2_t @llvm.arm.neon.vld2dup.v4i32.p0i8(i8* %src, i32 4) 170 ret %struct.uint32x4x2_t %tmp
|
D | arm-vld1.ll | 24 %struct.uint32x4x2_t = type { <4 x i32>, <4 x i32> } 56 declare %struct.uint32x4x2_t @llvm.arm.neon.vld1x2.v4i32.p0i32(i32*) nounwind readonly 177 define %struct.uint32x4x2_t @test_vld1q_u32_x2(i32* %a) nounwind { 178 %tmp = tail call %struct.uint32x4x2_t @llvm.arm.neon.vld1x2.v4i32.p0i32(i32* %a) 179 ret %struct.uint32x4x2_t %tmp
|
D | arm-vst1.ll | 24 ; %struct.uint32x4x2_t = type { <4 x i32>, <4 x i32> } 51 %struct.uint32x4x2_t = type { [2 x <4 x i32>] } 262 define void @test_vst1q_u32_x2(i32* %a, %struct.uint32x4x2_t %b) nounwind { 264 %b0 = extractvalue %struct.uint32x4x2_t %b, 0, 0 265 %b1 = extractvalue %struct.uint32x4x2_t %b, 0, 1
|
/external/clang/test/CodeGen/ |
D | aarch64-neon-perm.c | 1272 uint32x4x2_t test_vuzpq_u32(uint32x4_t a, uint32x4_t b) { in test_vuzpq_u32() 1736 uint32x4x2_t test_vzipq_u32(uint32x4_t a, uint32x4_t b) { in test_vzipq_u32() 2200 uint32x4x2_t test_vtrnq_u32(uint32x4_t a, uint32x4_t b) { in test_vtrnq_u32()
|
D | aarch64-neon-ldst-one.c | 353 uint32x4x2_t test_vld2q_dup_u32(uint32_t *a) { in test_vld2q_dup_u32() 2338 uint32x4x2_t test_vld2q_lane_u32(uint32_t *a, uint32x4x2_t b) { in test_vld2q_lane_u32() 5651 void test_vst2q_lane_u32(uint32_t *a, uint32x4x2_t b) { in test_vst2q_lane_u32()
|
D | arm_neon_intrinsics.c | 4873 uint32x4x2_t test_vld2q_u32(uint32_t const * a) { in test_vld2q_u32() 5458 uint32x4x2_t test_vld2q_lane_u32(uint32_t const * a, uint32x4x2_t b) { in test_vld2q_lane_u32() 18504 void test_vst2q_u32(uint32_t * a, uint32x4x2_t b) { in test_vst2q_u32() 19026 void test_vst2q_lane_u32(uint32_t * a, uint32x4x2_t b) { in test_vst2q_lane_u32() 23040 uint32x4x2_t test_vtrnq_u32(uint32x4_t a, uint32x4_t b) { in test_vtrnq_u32() 23580 uint32x4x2_t test_vuzpq_u32(uint32x4_t a, uint32x4_t b) { in test_vuzpq_u32() 23935 uint32x4x2_t test_vzipq_u32(uint32x4_t a, uint32x4_t b) { in test_vzipq_u32()
|
D | aarch64-neon-intrinsics.c | 10166 uint32x4x2_t test_vld2q_u32(uint32_t const *a) { in test_vld2q_u32() 11837 void test_vst2q_u32(uint32_t *a, uint32x4x2_t b) { in test_vst2q_u32() 14033 uint32x4x2_t test_vld1q_u32_x2(uint32_t const *a) { in test_vld1q_u32_x2() 15529 void test_vst1q_u32_x2(uint32_t *a, uint32x4x2_t b) { in test_vst1q_u32_x2()
|
/external/libvpx/libvpx/vp8/common/arm/neon/ |
D | vp8_loopfilter_neon.c | 326 uint32x4x2_t q2tmp0, q2tmp1, q2tmp2, q2tmp3; in vp8_loop_filter_vertical_edge_y_neon() 438 uint32x4x2_t q2tmp0, q2tmp1, q2tmp2, q2tmp3; in vp8_loop_filter_vertical_edge_uv_neon()
|
D | mbloopfilter_neon.c | 298 uint32x4x2_t q2tmp0, q2tmp1, q2tmp2, q2tmp3; in vp8_mbloop_filter_vertical_edge_y_neon() 462 uint32x4x2_t q2tmp0, q2tmp1, q2tmp2, q2tmp3; in vp8_mbloop_filter_vertical_edge_uv_neon()
|
/external/webp/src/dsp/ |
D | dec_neon.c | 202 const uint32x4x2_t row04 = vtrnq_u32(vreinterpretq_u32_u16(row02.val[0]), in Load8x8x2T_NEON() 204 const uint32x4x2_t row26 = vtrnq_u32(vreinterpretq_u32_u16(row02.val[1]), in Load8x8x2T_NEON() 206 const uint32x4x2_t row15 = vtrnq_u32(vreinterpretq_u32_u16(row13.val[0]), in Load8x8x2T_NEON() 208 const uint32x4x2_t row37 = vtrnq_u32(vreinterpretq_u32_u16(row13.val[1]), in Load8x8x2T_NEON()
|
/external/llvm/test/CodeGen/AArch64/ |
D | neon-perm.ll | 17 %struct.uint32x4x2_t = type { [2 x <4 x i32>] } 2646 define %struct.uint32x4x2_t @test_vuzpq_u32(<4 x i32> %a, <4 x i32> %b) { 2653 %.fca.0.0.insert = insertvalue %struct.uint32x4x2_t undef, <4 x i32> %vuzp.i, 0, 0 2654 %.fca.0.1.insert = insertvalue %struct.uint32x4x2_t %.fca.0.0.insert, <4 x i32> %vuzp1.i, 0, 1 2655 ret %struct.uint32x4x2_t %.fca.0.1.insert 2862 define %struct.uint32x4x2_t @test_vzipq_u32(<4 x i32> %a, <4 x i32> %b) { 2869 %.fca.0.0.insert = insertvalue %struct.uint32x4x2_t undef, <4 x i32> %vzip.i, 0, 0 2870 %.fca.0.1.insert = insertvalue %struct.uint32x4x2_t %.fca.0.0.insert, <4 x i32> %vzip1.i, 0, 1 2871 ret %struct.uint32x4x2_t %.fca.0.1.insert 3078 define %struct.uint32x4x2_t @test_vtrnq_u32(<4 x i32> %a, <4 x i32> %b) { [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/AArch64/ |
D | neon-perm.ll | 17 %struct.uint32x4x2_t = type { [2 x <4 x i32>] } 2646 define %struct.uint32x4x2_t @test_vuzpq_u32(<4 x i32> %a, <4 x i32> %b) { 2653 %.fca.0.0.insert = insertvalue %struct.uint32x4x2_t undef, <4 x i32> %vuzp.i, 0, 0 2654 %.fca.0.1.insert = insertvalue %struct.uint32x4x2_t %.fca.0.0.insert, <4 x i32> %vuzp1.i, 0, 1 2655 ret %struct.uint32x4x2_t %.fca.0.1.insert 2862 define %struct.uint32x4x2_t @test_vzipq_u32(<4 x i32> %a, <4 x i32> %b) { 2869 %.fca.0.0.insert = insertvalue %struct.uint32x4x2_t undef, <4 x i32> %vzip.i, 0, 0 2870 %.fca.0.1.insert = insertvalue %struct.uint32x4x2_t %.fca.0.0.insert, <4 x i32> %vzip1.i, 0, 1 2871 ret %struct.uint32x4x2_t %.fca.0.1.insert 3078 define %struct.uint32x4x2_t @test_vtrnq_u32(<4 x i32> %a, <4 x i32> %b) { [all …]
|
/external/libaom/libaom/aom_dsp/simd/ |
D | v128_intrinsics_arm.h | 546 uint32x4x2_t r = in v128_unziplo_32() 557 uint32x4x2_t r = in v128_unziphi_32()
|
/external/neon_2_sse/ |
D | NEON_2_SSE.h | 235 typedef struct int32x4x2_t uint32x4x2_t; typedef 1330 _NEON2SSESTORAGE uint32x4x2_t vld2q_u32(__transfersize(8) uint32_t const * ptr); // VLD2.32 {d0, d2… 1434 _NEON2SSESTORAGE uint32x4x2_t vld2q_lane_u32_ptr(__transfersize(2) uint32_t const * ptr, uint32x4x2… 1487 _NEON2SSESTORAGE void vst2q_u32_ptr(__transfersize(8) uint32_t * ptr, uint32x4x2_t * val); // VST2.… 1553 _NEON2SSESTORAGE void vst2q_lane_u32_ptr(__transfersize(2) uint32_t * ptr, uint32x4x2_t * val, __co… 2232 _NEON2SSESTORAGE uint32x4x2_t vtrnq_u32(uint32x4_t a, uint32x4_t b); // VTRN.32 q0,q0 2251 _NEON2SSESTORAGE uint32x4x2_t vzipq_u32(uint32x4_t a, uint32x4_t b); // VZIP.32 q0,q0 2270 _NEON2SSESTORAGE uint32x4x2_t vuzpq_u32(uint32x4_t a, uint32x4_t b); // VUZP.32 q0,q0 9805 _NEON2SSESTORAGE uint32x4x2_t vld2q_u32(__transfersize(8) uint32_t const * ptr); // VLD2.32 {d0, d2… 9806 _NEON2SSE_INLINE uint32x4x2_t vld2q_u32(__transfersize(8) uint32_t const * ptr) // VLD2.32 {d0, d2}… [all …]
|