/external/llvm-project/clang/test/CodeGen/ |
D | aarch64-neon-3v.c | 108 uint64x1_t test_vand_u64(uint64x1_t a, uint64x1_t b) { in test_vand_u64() 220 uint64x1_t test_vorr_u64(uint64x1_t a, uint64x1_t b) { in test_vorr_u64() 332 uint64x1_t test_veor_u64(uint64x1_t a, uint64x1_t b) { in test_veor_u64() 459 uint64x1_t test_vbic_u64(uint64x1_t a, uint64x1_t b) { in test_vbic_u64() 587 uint64x1_t test_vorn_u64(uint64x1_t a, uint64x1_t b) { in test_vorn_u64()
|
D | aarch64-neon-intrinsics.c | 70 uint64x1_t test_vadd_u64(uint64x1_t v1, uint64x1_t v2) { in test_vadd_u64() 203 uint64x1_t test_vsub_u64(uint64x1_t v1, uint64x1_t v2) { in test_vsub_u64() 1007 int64x1_t test_vbsl_s64(uint64x1_t v1, int64x1_t v2, int64x1_t v3) { in test_vbsl_s64() 1056 uint64x1_t test_vbsl_u64(uint64x1_t v1, uint64x1_t v2, uint64x1_t v3) { in test_vbsl_u64() 1088 float64x1_t test_vbsl_f64(uint64x1_t v1, float64x1_t v2, float64x1_t v3) { in test_vbsl_f64() 1341 uint64x1_t test_vcage_f64(float64x1_t a, float64x1_t b) { in test_vcage_f64() 1377 uint64x1_t test_vcagt_f64(float64x1_t a, float64x1_t b) { in test_vcagt_f64() 1414 uint64x1_t test_vcale_f64(float64x1_t a, float64x1_t b) { in test_vcale_f64() 1453 uint64x1_t test_vcalt_f64(float64x1_t a, float64x1_t b) { in test_vcalt_f64() 1670 uint64x1_t test_vtst_s64(int64x1_t a, int64x1_t b) { in test_vtst_s64() [all …]
|
D | aarch64-neon-misc.c | 40 uint64x1_t test_vceqz_s64(int64x1_t a) { in test_vceqz_s64() 49 uint64x1_t test_vceqz_u64(uint64x1_t a) { in test_vceqz_u64() 58 uint64x1_t test_vceqz_p64(poly64x1_t a) { in test_vceqz_p64() 172 uint64x1_t test_vceqz_f64(float64x1_t a) { in test_vceqz_f64() 268 uint64x1_t test_vcgez_s64(int64x1_t a) { in test_vcgez_s64() 321 uint64x1_t test_vcgez_f64(float64x1_t a) { in test_vcgez_f64() 374 uint64x1_t test_vclez_s64(int64x1_t a) { in test_vclez_s64() 427 uint64x1_t test_vclez_f64(float64x1_t a) { in test_vclez_f64() 480 uint64x1_t test_vcgtz_s64(int64x1_t a) { in test_vcgtz_s64() 533 uint64x1_t test_vcgtz_f64(float64x1_t a) { in test_vcgtz_f64() [all …]
|
D | aarch64-neon-intrinsics-constrained.c | 295 uint64x1_t test_vceq_f64(float64x1_t a, float64x1_t b) { in test_vceq_f64() 336 uint64x1_t test_vcge_f64(float64x1_t a, float64x1_t b) { in test_vcge_f64() 377 uint64x1_t test_vcle_f64(float64x1_t a, float64x1_t b) { in test_vcle_f64() 418 uint64x1_t test_vcgt_f64(float64x1_t a, float64x1_t b) { in test_vcgt_f64() 459 uint64x1_t test_vclt_f64(float64x1_t a, float64x1_t b) { in test_vclt_f64() 864 uint64x1_t test_vcvt_u64_f64(float64x1_t a) { in test_vcvt_u64_f64() 884 float64x1_t test_vcvt_f64_u64(uint64x1_t a) { in test_vcvt_f64_u64()
|
D | arm_neon_intrinsics.c | 513 uint64x1_t test_vadd_u64(uint64x1_t a, uint64x1_t b) { in test_vadd_u64() 812 uint64x1_t test_vand_u64(uint64x1_t a, uint64x1_t b) { in test_vand_u64() 932 uint64x1_t test_vbic_u64(uint64x1_t a, uint64x1_t b) { in test_vbic_u64() 1036 int64x1_t test_vbsl_s64(uint64x1_t a, int64x1_t b, int64x1_t c) { in test_vbsl_s64() 1076 uint64x1_t test_vbsl_u64(uint64x1_t a, uint64x1_t b, uint64x1_t c) { in test_vbsl_u64() 2124 uint64x2_t test_vcombine_u64(uint64x1_t a, uint64x1_t b) { in test_vcombine_u64() 2223 uint64x1_t test_vcreate_u64(uint64_t a) { in test_vcreate_u64() 2224 uint64x1_t tmp = vcreate_u64(a); in test_vcreate_u64() 2578 uint64x1_t test_vdup_lane_u64(uint64x1_t a) { in test_vdup_lane_u64() 2596 uint64x2_t test_vdupq_lane_u64(uint64x1_t a) { in test_vdupq_lane_u64() [all …]
|
D | aarch64-neon-extract.c | 125 uint64x1_t test_vext_u64(uint64x1_t a, uint64x1_t b) { in test_vext_u64()
|
D | aarch64-neon-vcombine.c | 59 uint64x2_t test_vcombine_u64(uint64x1_t low, uint64x1_t high) { in test_vcombine_u64()
|
D | arm-neon-vget.c | 48 uint64x1_t low_u64(uint64x2_t a) { in low_u64() 104 uint64x1_t high_u64(uint64x2_t a) { in high_u64()
|
D | aarch64-neon-vget.c | 175 uint64_t test_vget_lane_u64(uint64x1_t a) { in test_vget_lane_u64() 370 uint64x1_t test_vset_lane_u64(uint64_t a, uint64x1_t b) { in test_vset_lane_u64()
|
D | aarch64-neon-vget-hilo.c | 60 uint64x1_t test_vget_high_u64(uint64x2_t a) { in test_vget_high_u64() 158 uint64x1_t test_vget_low_u64(uint64x2_t a) { in test_vget_low_u64()
|
/external/clang/test/CodeGen/ |
D | aarch64-neon-3v.c | 108 uint64x1_t test_vand_u64(uint64x1_t a, uint64x1_t b) { in test_vand_u64() 220 uint64x1_t test_vorr_u64(uint64x1_t a, uint64x1_t b) { in test_vorr_u64() 332 uint64x1_t test_veor_u64(uint64x1_t a, uint64x1_t b) { in test_veor_u64() 459 uint64x1_t test_vbic_u64(uint64x1_t a, uint64x1_t b) { in test_vbic_u64() 587 uint64x1_t test_vorn_u64(uint64x1_t a, uint64x1_t b) { in test_vorn_u64()
|
D | aarch64-neon-intrinsics.c | 69 uint64x1_t test_vadd_u64(uint64x1_t v1, uint64x1_t v2) { in test_vadd_u64() 200 uint64x1_t test_vsub_u64(uint64x1_t v1, uint64x1_t v2) { in test_vsub_u64() 1072 uint64x1_t test_vbsl_s64(uint64x1_t v1, uint64x1_t v2, uint64x1_t v3) { in test_vbsl_s64() 1130 uint64x1_t test_vbsl_u64(uint64x1_t v1, uint64x1_t v2, uint64x1_t v3) { in test_vbsl_u64() 1165 float64x1_t test_vbsl_f64(uint64x1_t v1, float64x1_t v2, float64x1_t v3) { in test_vbsl_f64() 1467 uint64x1_t test_vcage_f64(float64x1_t a, float64x1_t b) { in test_vcage_f64() 1511 uint64x1_t test_vcagt_f64(float64x1_t a, float64x1_t b) { in test_vcagt_f64() 1556 uint64x1_t test_vcale_f64(float64x1_t a, float64x1_t b) { in test_vcale_f64() 1603 uint64x1_t test_vcalt_f64(float64x1_t a, float64x1_t b) { in test_vcalt_f64() 1850 uint64x1_t test_vtst_s64(int64x1_t a, int64x1_t b) { in test_vtst_s64() [all …]
|
D | arm_neon_intrinsics.c | 609 uint64x1_t test_vadd_u64(uint64x1_t a, uint64x1_t b) { in test_vadd_u64() 936 uint64x1_t test_vand_u64(uint64x1_t a, uint64x1_t b) { in test_vand_u64() 1057 uint64x1_t test_vbic_u64(uint64x1_t a, uint64x1_t b) { in test_vbic_u64() 1162 int64x1_t test_vbsl_s64(uint64x1_t a, int64x1_t b, int64x1_t c) { in test_vbsl_s64() 1202 uint64x1_t test_vbsl_u64(uint64x1_t a, uint64x1_t b, uint64x1_t c) { in test_vbsl_u64() 2303 uint64x2_t test_vcombine_u64(uint64x1_t a, uint64x1_t b) { in test_vcombine_u64() 2408 uint64x1_t test_vcreate_u64(uint64_t a) { in test_vcreate_u64() 2409 uint64x1_t tmp = vcreate_u64(a); in test_vcreate_u64() 2755 uint64x1_t test_vdup_lane_u64(uint64x1_t a) { in test_vdup_lane_u64() 2769 uint64x2_t test_vdupq_lane_u64(uint64x1_t a) { in test_vdupq_lane_u64() [all …]
|
D | aarch64-neon-misc.c | 43 uint64x1_t test_vceqz_s64(int64x1_t a) { in test_vceqz_s64() 53 uint64x1_t test_vceqz_u64(uint64x1_t a) { in test_vceqz_u64() 63 uint64x1_t test_vceqz_p64(poly64x1_t a) { in test_vceqz_p64() 187 uint64x1_t test_vceqz_f64(float64x1_t a) { in test_vceqz_f64() 291 uint64x1_t test_vcgez_s64(int64x1_t a) { in test_vcgez_s64() 349 uint64x1_t test_vcgez_f64(float64x1_t a) { in test_vcgez_f64() 407 uint64x1_t test_vclez_s64(int64x1_t a) { in test_vclez_s64() 465 uint64x1_t test_vclez_f64(float64x1_t a) { in test_vclez_f64() 523 uint64x1_t test_vcgtz_s64(int64x1_t a) { in test_vcgtz_s64() 581 uint64x1_t test_vcgtz_f64(float64x1_t a) { in test_vcgtz_f64() [all …]
|
D | aarch64-neon-extract.c | 125 uint64x1_t test_vext_u64(uint64x1_t a, uint64x1_t b) { in test_vext_u64()
|
D | aarch64-neon-vcombine.c | 59 uint64x2_t test_vcombine_u64(uint64x1_t low, uint64x1_t high) { in test_vcombine_u64()
|
D | arm-neon-vget.c | 48 uint64x1_t low_u64(uint64x2_t a) { in low_u64() 104 uint64x1_t high_u64(uint64x2_t a) { in high_u64()
|
D | aarch64-neon-vget.c | 207 uint64_t test_vget_lane_u64(uint64x1_t a) { in test_vget_lane_u64() 438 uint64x1_t test_vset_lane_u64(uint64_t a, uint64x1_t b) { in test_vset_lane_u64()
|
D | aarch64-neon-vget-hilo.c | 60 uint64x1_t test_vget_high_u64(uint64x2_t a) { in test_vget_high_u64() 158 uint64x1_t test_vget_low_u64(uint64x2_t a) { in test_vget_low_u64()
|
/external/arm-optimized-routines/networking/arm/ |
D | chksum_simd.c | 21 uint64x1_t vsum = { 0 }; in __chksum_arm_simd() 36 uint64x1_t vword64 = vld1_u64(ptr64); in __chksum_arm_simd() 38 uint64x1_t vmask = vdup_n_u64(ALL_ONES); in __chksum_arm_simd() 116 uint64x1_t vword64 = vld1_u64((const uint64_t *) ptr32); in __chksum_arm_simd() 118 uint64x1_t vmask = vdup_n_u64(ALL_ONES); in __chksum_arm_simd()
|
/external/llvm-project/libc/AOR_v20.02/networking/arm/ |
D | chksum_simd.c | 22 uint64x1_t vsum = { 0 }; in __chksum_arm_simd() 37 uint64x1_t vword64 = vld1_u64(ptr64); in __chksum_arm_simd() 39 uint64x1_t vmask = vdup_n_u64(ALL_ONES); in __chksum_arm_simd() 117 uint64x1_t vword64 = vld1_u64((const uint64_t *) ptr32); in __chksum_arm_simd() 119 uint64x1_t vmask = vdup_n_u64(ALL_ONES); in __chksum_arm_simd()
|
/external/webp/src/dsp/ |
D | neon.h | 60 const uint64x1_t row0h = vget_high_u64(row01.val[0]); in Transpose4x4_NEON() 61 const uint64x1_t row2l = vget_low_u64(row23.val[0]); in Transpose4x4_NEON() 62 const uint64x1_t row1h = vget_high_u64(row01.val[1]); in Transpose4x4_NEON() 63 const uint64x1_t row3l = vget_low_u64(row23.val[1]); in Transpose4x4_NEON()
|
/external/clang/test/CodeGenCXX/ |
D | mangle-neon-vectors.cpp | 19 typedef __attribute__((neon_vector_type(1))) uint64_t uint64x1_t; typedef 41 void f3(uint64x1_t v) { } in f3()
|
/external/XNNPACK/src/x8-zip/ |
D | xm-neon.c | 81 const uint64x1_t vx = vshl_u64(vreinterpret_u64_u8(vld1_u8(x)), vshift); in xnn_x8_zip_xm_ukernel__neon() 82 const uint64x1_t vy = vshl_u64(vreinterpret_u64_u8(vld1_u8(y)), vshift); in xnn_x8_zip_xm_ukernel__neon() 83 const uint64x1_t vz = vshl_u64(vreinterpret_u64_u8(vld1_u8(z)), vshift); in xnn_x8_zip_xm_ukernel__neon() 84 const uint64x1_t vw = vshl_u64(vreinterpret_u64_u8(vld1_u8(w)), vshift); w += 8; in xnn_x8_zip_xm_ukernel__neon()
|
/external/llvm-project/clang/test/CodeGenCXX/ |
D | mangle-neon-vectors.cpp | 24 typedef __attribute__((neon_vector_type(1))) uint64_t uint64x1_t; typedef 50 void f3(uint64x1_t v) { } in f3()
|