/external/clang/test/CodeGen/ |
D | mmx-builtins.c | 9 __m64 test_mm_abs_pi8(__m64 a) { in test_mm_abs_pi8() 15 __m64 test_mm_abs_pi16(__m64 a) { in test_mm_abs_pi16() 21 __m64 test_mm_abs_pi32(__m64 a) { in test_mm_abs_pi32() 27 __m64 test_mm_add_pi8(__m64 a, __m64 b) { in test_mm_add_pi8() 33 __m64 test_mm_add_pi16(__m64 a, __m64 b) { in test_mm_add_pi16() 39 __m64 test_mm_add_pi32(__m64 a, __m64 b) { in test_mm_add_pi32() 45 __m64 test_mm_add_si64(__m64 a, __m64 b) { in test_mm_add_si64() 51 __m64 test_mm_adds_pi8(__m64 a, __m64 b) { in test_mm_adds_pi8() 57 __m64 test_mm_adds_pi16(__m64 a, __m64 b) { in test_mm_adds_pi16() 63 __m64 test_mm_adds_pu8(__m64 a, __m64 b) { in test_mm_adds_pu8() [all …]
|
D | 3dnow-builtins.c | 9 __m64 test_m_pavgusb(__m64 m1, __m64 m2) { in test_m_pavgusb() 16 __m64 test_m_pf2id(__m64 m) { in test_m_pf2id() 23 __m64 test_m_pfacc(__m64 m1, __m64 m2) { in test_m_pfacc() 30 __m64 test_m_pfadd(__m64 m1, __m64 m2) { in test_m_pfadd() 37 __m64 test_m_pfcmpeq(__m64 m1, __m64 m2) { in test_m_pfcmpeq() 44 __m64 test_m_pfcmpge(__m64 m1, __m64 m2) { in test_m_pfcmpge() 51 __m64 test_m_pfcmpgt(__m64 m1, __m64 m2) { in test_m_pfcmpgt() 58 __m64 test_m_pfmax(__m64 m1, __m64 m2) { in test_m_pfmax() 65 __m64 test_m_pfmin(__m64 m1, __m64 m2) { in test_m_pfmin() 72 __m64 test_m_pfmul(__m64 m1, __m64 m2) { in test_m_pfmul() [all …]
|
D | asm-inout.c | 43 typedef long long __m64 __attribute__((__vector_size__(8))); typedef 44 __m64 test5(__m64 __A, __m64 __B) { in test5()
|
D | vector.c | 30 __m64 *p = (__m64 *)array; in test4() 32 __m64 accum = _mm_setzero_si64(); in test4() 37 __m64 accum2 = _mm_unpackhi_pi32(accum, accum); in test4()
|
/external/clang/lib/Headers/ |
D | mmintrin.h | 27 typedef long long __m64 __attribute__((__vector_size__(8))); typedef 61 static __inline__ __m64 __DEFAULT_FN_ATTRS 64 return (__m64)__builtin_ia32_vec_init_v2si(__i, 0); in _mm_cvtsi32_si64() 79 _mm_cvtsi64_si32(__m64 __m) in _mm_cvtsi64_si32() 94 static __inline__ __m64 __DEFAULT_FN_ATTRS 97 return (__m64)__i; in _mm_cvtsi64_m64() 111 _mm_cvtm64_si64(__m64 __m) in _mm_cvtm64_si64() 140 static __inline__ __m64 __DEFAULT_FN_ATTRS 141 _mm_packs_pi16(__m64 __m1, __m64 __m2) in _mm_packs_pi16() 143 return (__m64)__builtin_ia32_packsswb((__v4hi)__m1, (__v4hi)__m2); in _mm_packs_pi16() [all …]
|
D | mm3dnow.h | 40 static __inline__ __m64 __DEFAULT_FN_ATTRS 41 _m_pavgusb(__m64 __m1, __m64 __m2) { in _m_pavgusb() 42 return (__m64)__builtin_ia32_pavgusb((__v8qi)__m1, (__v8qi)__m2); in _m_pavgusb() 45 static __inline__ __m64 __DEFAULT_FN_ATTRS 46 _m_pf2id(__m64 __m) { in _m_pf2id() 47 return (__m64)__builtin_ia32_pf2id((__v2sf)__m); in _m_pf2id() 50 static __inline__ __m64 __DEFAULT_FN_ATTRS 51 _m_pfacc(__m64 __m1, __m64 __m2) { in _m_pfacc() 52 return (__m64)__builtin_ia32_pfacc((__v2sf)__m1, (__v2sf)__m2); in _m_pfacc() 55 static __inline__ __m64 __DEFAULT_FN_ATTRS [all …]
|
D | tmmintrin.h | 44 static __inline__ __m64 __DEFAULT_FN_ATTRS 45 _mm_abs_pi8(__m64 __a) in _mm_abs_pi8() 47 return (__m64)__builtin_ia32_pabsb((__v8qi)__a); in _mm_abs_pi8() 80 static __inline__ __m64 __DEFAULT_FN_ATTRS 81 _mm_abs_pi16(__m64 __a) in _mm_abs_pi16() 83 return (__m64)__builtin_ia32_pabsw((__v4hi)__a); in _mm_abs_pi16() 116 static __inline__ __m64 __DEFAULT_FN_ATTRS 117 _mm_abs_pi32(__m64 __a) in _mm_abs_pi32() 119 return (__m64)__builtin_ia32_pabsd((__v2si)__a); in _mm_abs_pi32() 184 (__m64)__builtin_ia32_palignr((__v8qi)(__m64)(a), (__v8qi)(__m64)(b), (n)); }) [all …]
|
D | xmmintrin.h | 1316 static __inline__ __m64 __DEFAULT_FN_ATTRS 1319 return (__m64)__builtin_ia32_cvtps2pi((__v4sf)__a); in _mm_cvtps_pi32() 1332 static __inline__ __m64 __DEFAULT_FN_ATTRS 1403 static __inline__ __m64 __DEFAULT_FN_ATTRS 1406 return (__m64)__builtin_ia32_cvttps2pi((__v4sf)__a); in _mm_cvttps_pi32() 1420 static __inline__ __m64 __DEFAULT_FN_ATTRS 1516 _mm_cvtpi32_ps(__m128 __a, __m64 __b) in _mm_cvtpi32_ps() 1539 _mm_cvt_pi2ps(__m128 __a, __m64 __b) in _mm_cvt_pi2ps() 1577 _mm_loadh_pi(__m128 __a, const __m64 *__p) in _mm_loadh_pi() 1604 _mm_loadl_pi(__m128 __a, const __m64 *__p) in _mm_loadl_pi() [all …]
|
/external/XNNPACK/src/f32-ibilinear-chw/gen/ |
D | sse-p8.c | 59 const __m128 vtltr0 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl0); in xnn_f32_ibilinear_chw_ukernel__sse_p8() 60 const __m128 vblbr0 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl0); in xnn_f32_ibilinear_chw_ukernel__sse_p8() 61 const __m128 vtltr2 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl2); in xnn_f32_ibilinear_chw_ukernel__sse_p8() 62 const __m128 vblbr2 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl2); in xnn_f32_ibilinear_chw_ukernel__sse_p8() 63 const __m128 vtltr4 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl4); in xnn_f32_ibilinear_chw_ukernel__sse_p8() 64 const __m128 vblbr4 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl4); in xnn_f32_ibilinear_chw_ukernel__sse_p8() 65 const __m128 vtltr6 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl6); in xnn_f32_ibilinear_chw_ukernel__sse_p8() 66 const __m128 vblbr6 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl6); in xnn_f32_ibilinear_chw_ukernel__sse_p8() 73 const __m128 vtltr01 = _mm_loadh_pi(vtltr0, (const __m64*) itl1); in xnn_f32_ibilinear_chw_ukernel__sse_p8() 74 const __m128 vblbr01 = _mm_loadh_pi(vblbr0, (const __m64*) ibl1); in xnn_f32_ibilinear_chw_ukernel__sse_p8() [all …]
|
D | sse-p4.c | 50 const __m128 vtltr0 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl0); in xnn_f32_ibilinear_chw_ukernel__sse_p4() 51 const __m128 vblbr0 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl0); in xnn_f32_ibilinear_chw_ukernel__sse_p4() 52 const __m128 vtltr2 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl2); in xnn_f32_ibilinear_chw_ukernel__sse_p4() 53 const __m128 vblbr2 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl2); in xnn_f32_ibilinear_chw_ukernel__sse_p4() 58 const __m128 vtltr01 = _mm_loadh_pi(vtltr0, (const __m64*) itl1); in xnn_f32_ibilinear_chw_ukernel__sse_p4() 59 const __m128 vblbr01 = _mm_loadh_pi(vblbr0, (const __m64*) ibl1); in xnn_f32_ibilinear_chw_ukernel__sse_p4() 60 const __m128 vtltr23 = _mm_loadh_pi(vtltr2, (const __m64*) itl3); in xnn_f32_ibilinear_chw_ukernel__sse_p4() 61 const __m128 vblbr23 = _mm_loadh_pi(vblbr2, (const __m64*) ibl3); in xnn_f32_ibilinear_chw_ukernel__sse_p4() 96 …8 vtltr = _mm_loadh_pi(_mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl0), (const __m64*) itl1); in xnn_f32_ibilinear_chw_ukernel__sse_p4() 97 …8 vblbr = _mm_loadh_pi(_mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl0), (const __m64*) ibl1); in xnn_f32_ibilinear_chw_ukernel__sse_p4() [all …]
|
/external/XNNPACK/src/f32-ibilinear-chw/ |
D | sse.c.in | 46 … const __m128 vtltr${ABC[P]} = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl${ABC[P]}); 47 … const __m128 vblbr${ABC[P]} = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl${ABC[P]}); 54 … const __m128 vtltr${ABC[P:P+2]} = _mm_loadh_pi(vtltr${ABC[P]}, (const __m64*) itl${ABC[P+1]}); 55 … const __m128 vblbr${ABC[P:P+2]} = _mm_loadh_pi(vblbr${ABC[P]}, (const __m64*) ibl${ABC[P+1]}); 94 const __m128 vtltr${ABC[P]} = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl${P}); 95 const __m128 vblbr${ABC[P]} = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl${P}); 101 const __m128 vtltr${ABC[P:P+2]} = _mm_loadh_pi(vtltr${ABC[P]}, (const __m64*) itl${P+1}); 102 const __m128 vblbr${ABC[P:P+2]} = _mm_loadh_pi(vblbr${ABC[P]}, (const __m64*) ibl${P+1}); 136 …8 vtltr = _mm_loadh_pi(_mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl0), (const __m64*) itl1); 137 …8 vblbr = _mm_loadh_pi(_mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl0), (const __m64*) ibl1); [all …]
|
/external/clang/test/Sema/ |
D | x86-builtin-palignr.c | 6 __m64 test1(__m64 a, __m64 b, int c) { in test1()
|
/external/libmpeg2/common/x86/ |
D | icv_sad_ssse3.c | 114 src_r0 = _mm_loadh_pi (src_r0, (__m64 *) (pu1_src)); in icv_sad_8x4_ssse3() 117 src_r1 = _mm_loadh_pi (src_r1, (__m64 *) (pu1_src)); in icv_sad_8x4_ssse3() 128 ref_r0 = _mm_loadh_pi (ref_r0, (__m64 *) (pu1_ref)); in icv_sad_8x4_ssse3() 131 ref_r1 = _mm_loadh_pi (ref_r1, (__m64 *) (pu1_ref)); in icv_sad_8x4_ssse3()
|
/external/fec/ |
D | viterbi39_sse.c | 13 typedef union { signed short s[256]; __m64 v[64];} metric_t; 15 static union branchtab39 { unsigned short s[128]; __m64 v[32];} Branchtab39[3]; 128 __m64 sym0v,sym1v,sym2v; in update_viterbi39_blk_sse() 139 __m64 decision0,decision1,metric,m_metric,m0,m1,m2,m3,survivor0,survivor1; in update_viterbi39_blk_sse() 174 __m64 adjustv; in update_viterbi39_blk_sse() 175 union { __m64 v; signed short w[4]; } t; in update_viterbi39_blk_sse()
|
D | viterbi615_sse.c | 13 typedef union { signed short s[16384]; __m64 v[4096];} metric_t; 15 static union branchtab615 { unsigned short s[8192]; __m64 v[2048];} Branchtab615[6]; 124 __m64 sym0v,sym1v,sym2v,sym3v,sym4v,sym5v; in update_viterbi615_blk_sse() 138 __m64 decision0,decision1,metric,m_metric,m0,m1,m2,m3,survivor0,survivor1; in update_viterbi615_blk_sse() 175 __m64 adjustv; in update_viterbi615_blk_sse() 176 union { __m64 v; signed short w[4]; } t; in update_viterbi615_blk_sse()
|
/external/XNNPACK/src/f32-gemm/gen/ |
D | 8x8-minmax-fma3-broadcast.c | 214 _mm_storel_pi((__m64*) c7, vacc7x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast() 215 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast() 216 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast() 217 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast() 218 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast() 219 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast() 220 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast() 221 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast()
|
D | 7x8-minmax-fma3-broadcast.c | 195 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_gemm_minmax_ukernel_7x8__fma3_broadcast() 196 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_gemm_minmax_ukernel_7x8__fma3_broadcast() 197 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_gemm_minmax_ukernel_7x8__fma3_broadcast() 198 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_gemm_minmax_ukernel_7x8__fma3_broadcast() 199 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_gemm_minmax_ukernel_7x8__fma3_broadcast() 200 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_gemm_minmax_ukernel_7x8__fma3_broadcast() 201 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_gemm_minmax_ukernel_7x8__fma3_broadcast()
|
D | 7x8-minmax-avx-broadcast.c | 195 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_gemm_minmax_ukernel_7x8__avx_broadcast() 196 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_gemm_minmax_ukernel_7x8__avx_broadcast() 197 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_gemm_minmax_ukernel_7x8__avx_broadcast() 198 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_gemm_minmax_ukernel_7x8__avx_broadcast() 199 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_gemm_minmax_ukernel_7x8__avx_broadcast() 200 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_gemm_minmax_ukernel_7x8__avx_broadcast() 201 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_gemm_minmax_ukernel_7x8__avx_broadcast()
|
/external/XNNPACK/src/f32-igemm/gen/ |
D | 8x8-minmax-fma3-broadcast.c | 241 _mm_storel_pi((__m64*) c7, vacc7x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast() 242 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast() 243 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast() 244 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast() 245 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast() 246 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast() 247 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast() 248 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast()
|
D | 7x8-minmax-avx-broadcast.c | 220 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_igemm_minmax_ukernel_7x8__avx_broadcast() 221 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_igemm_minmax_ukernel_7x8__avx_broadcast() 222 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_igemm_minmax_ukernel_7x8__avx_broadcast() 223 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_igemm_minmax_ukernel_7x8__avx_broadcast() 224 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_igemm_minmax_ukernel_7x8__avx_broadcast() 225 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_igemm_minmax_ukernel_7x8__avx_broadcast() 226 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_igemm_minmax_ukernel_7x8__avx_broadcast()
|
D | 7x8-minmax-fma3-broadcast.c | 220 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_igemm_minmax_ukernel_7x8__fma3_broadcast() 221 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_igemm_minmax_ukernel_7x8__fma3_broadcast() 222 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_igemm_minmax_ukernel_7x8__fma3_broadcast() 223 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_igemm_minmax_ukernel_7x8__fma3_broadcast() 224 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_igemm_minmax_ukernel_7x8__fma3_broadcast() 225 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_igemm_minmax_ukernel_7x8__fma3_broadcast() 226 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_igemm_minmax_ukernel_7x8__fma3_broadcast()
|
/external/XNNPACK/src/f32-gemm/gen-inc/ |
D | 8x8inc-minmax-fma3-broadcast.c | 216 _mm_storel_pi((__m64*) c7, vacc7x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast() 217 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast() 218 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast() 219 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast() 220 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast() 221 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast() 222 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast() 223 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast()
|
D | 7x8inc-minmax-avx-broadcast.c | 197 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__avx_broadcast() 198 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__avx_broadcast() 199 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__avx_broadcast() 200 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__avx_broadcast() 201 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__avx_broadcast() 202 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__avx_broadcast() 203 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__avx_broadcast()
|
D | 7x8inc-minmax-fma3-broadcast.c | 197 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__fma3_broadcast() 198 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__fma3_broadcast() 199 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__fma3_broadcast() 200 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__fma3_broadcast() 201 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__fma3_broadcast() 202 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__fma3_broadcast() 203 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__fma3_broadcast()
|
/external/pffft/ |
D | sse2neon.h | 148 typedef int64x1_t __m64; typedef 509 FORCE_INLINE __m128i _mm_setr_epi64(__m64 e1, __m64 e0) in _mm_setr_epi64() 624 FORCE_INLINE __m128i _mm_set1_epi64(__m64 _i) in _mm_set1_epi64() 656 FORCE_INLINE __m128i _mm_set_epi64(__m64 i1, __m64 i2) in _mm_set_epi64() 747 FORCE_INLINE void _mm_storel_pi(__m64 *p, __m128 a) in _mm_storel_pi() 759 FORCE_INLINE void _mm_storeh_pi(__m64 *p, __m128 a) in _mm_storeh_pi() 794 FORCE_INLINE __m128 _mm_loadl_pi(__m128 a, __m64 const *p) in _mm_loadl_pi() 826 FORCE_INLINE __m128 _mm_loadh_pi(__m128 a, __m64 const *p) in _mm_loadh_pi() 1251 FORCE_INLINE __m64 _mm_abs_pi32(__m64 a) in _mm_abs_pi32() 1265 FORCE_INLINE __m64 _mm_abs_pi16(__m64 a) in _mm_abs_pi16() [all …]
|