Home
last modified time | relevance | path

Searched refs:__m64 (Results 1 – 25 of 768) sorted by relevance

12345678910>>...31

/external/clang/test/CodeGen/
Dmmx-builtins.c9 __m64 test_mm_abs_pi8(__m64 a) { in test_mm_abs_pi8()
15 __m64 test_mm_abs_pi16(__m64 a) { in test_mm_abs_pi16()
21 __m64 test_mm_abs_pi32(__m64 a) { in test_mm_abs_pi32()
27 __m64 test_mm_add_pi8(__m64 a, __m64 b) { in test_mm_add_pi8()
33 __m64 test_mm_add_pi16(__m64 a, __m64 b) { in test_mm_add_pi16()
39 __m64 test_mm_add_pi32(__m64 a, __m64 b) { in test_mm_add_pi32()
45 __m64 test_mm_add_si64(__m64 a, __m64 b) { in test_mm_add_si64()
51 __m64 test_mm_adds_pi8(__m64 a, __m64 b) { in test_mm_adds_pi8()
57 __m64 test_mm_adds_pi16(__m64 a, __m64 b) { in test_mm_adds_pi16()
63 __m64 test_mm_adds_pu8(__m64 a, __m64 b) { in test_mm_adds_pu8()
[all …]
D3dnow-builtins.c9 __m64 test_m_pavgusb(__m64 m1, __m64 m2) { in test_m_pavgusb()
16 __m64 test_m_pf2id(__m64 m) { in test_m_pf2id()
23 __m64 test_m_pfacc(__m64 m1, __m64 m2) { in test_m_pfacc()
30 __m64 test_m_pfadd(__m64 m1, __m64 m2) { in test_m_pfadd()
37 __m64 test_m_pfcmpeq(__m64 m1, __m64 m2) { in test_m_pfcmpeq()
44 __m64 test_m_pfcmpge(__m64 m1, __m64 m2) { in test_m_pfcmpge()
51 __m64 test_m_pfcmpgt(__m64 m1, __m64 m2) { in test_m_pfcmpgt()
58 __m64 test_m_pfmax(__m64 m1, __m64 m2) { in test_m_pfmax()
65 __m64 test_m_pfmin(__m64 m1, __m64 m2) { in test_m_pfmin()
72 __m64 test_m_pfmul(__m64 m1, __m64 m2) { in test_m_pfmul()
[all …]
Dasm-inout.c43 typedef long long __m64 __attribute__((__vector_size__(8))); typedef
44 __m64 test5(__m64 __A, __m64 __B) { in test5()
Dvector.c30 __m64 *p = (__m64 *)array; in test4()
32 __m64 accum = _mm_setzero_si64(); in test4()
37 __m64 accum2 = _mm_unpackhi_pi32(accum, accum); in test4()
/external/clang/lib/Headers/
Dmmintrin.h27 typedef long long __m64 __attribute__((__vector_size__(8))); typedef
61 static __inline__ __m64 __DEFAULT_FN_ATTRS
64 return (__m64)__builtin_ia32_vec_init_v2si(__i, 0); in _mm_cvtsi32_si64()
79 _mm_cvtsi64_si32(__m64 __m) in _mm_cvtsi64_si32()
94 static __inline__ __m64 __DEFAULT_FN_ATTRS
97 return (__m64)__i; in _mm_cvtsi64_m64()
111 _mm_cvtm64_si64(__m64 __m) in _mm_cvtm64_si64()
140 static __inline__ __m64 __DEFAULT_FN_ATTRS
141 _mm_packs_pi16(__m64 __m1, __m64 __m2) in _mm_packs_pi16()
143 return (__m64)__builtin_ia32_packsswb((__v4hi)__m1, (__v4hi)__m2); in _mm_packs_pi16()
[all …]
Dmm3dnow.h40 static __inline__ __m64 __DEFAULT_FN_ATTRS
41 _m_pavgusb(__m64 __m1, __m64 __m2) { in _m_pavgusb()
42 return (__m64)__builtin_ia32_pavgusb((__v8qi)__m1, (__v8qi)__m2); in _m_pavgusb()
45 static __inline__ __m64 __DEFAULT_FN_ATTRS
46 _m_pf2id(__m64 __m) { in _m_pf2id()
47 return (__m64)__builtin_ia32_pf2id((__v2sf)__m); in _m_pf2id()
50 static __inline__ __m64 __DEFAULT_FN_ATTRS
51 _m_pfacc(__m64 __m1, __m64 __m2) { in _m_pfacc()
52 return (__m64)__builtin_ia32_pfacc((__v2sf)__m1, (__v2sf)__m2); in _m_pfacc()
55 static __inline__ __m64 __DEFAULT_FN_ATTRS
[all …]
Dtmmintrin.h44 static __inline__ __m64 __DEFAULT_FN_ATTRS
45 _mm_abs_pi8(__m64 __a) in _mm_abs_pi8()
47 return (__m64)__builtin_ia32_pabsb((__v8qi)__a); in _mm_abs_pi8()
80 static __inline__ __m64 __DEFAULT_FN_ATTRS
81 _mm_abs_pi16(__m64 __a) in _mm_abs_pi16()
83 return (__m64)__builtin_ia32_pabsw((__v4hi)__a); in _mm_abs_pi16()
116 static __inline__ __m64 __DEFAULT_FN_ATTRS
117 _mm_abs_pi32(__m64 __a) in _mm_abs_pi32()
119 return (__m64)__builtin_ia32_pabsd((__v2si)__a); in _mm_abs_pi32()
184 (__m64)__builtin_ia32_palignr((__v8qi)(__m64)(a), (__v8qi)(__m64)(b), (n)); })
[all …]
Dxmmintrin.h1316 static __inline__ __m64 __DEFAULT_FN_ATTRS
1319 return (__m64)__builtin_ia32_cvtps2pi((__v4sf)__a); in _mm_cvtps_pi32()
1332 static __inline__ __m64 __DEFAULT_FN_ATTRS
1403 static __inline__ __m64 __DEFAULT_FN_ATTRS
1406 return (__m64)__builtin_ia32_cvttps2pi((__v4sf)__a); in _mm_cvttps_pi32()
1420 static __inline__ __m64 __DEFAULT_FN_ATTRS
1516 _mm_cvtpi32_ps(__m128 __a, __m64 __b) in _mm_cvtpi32_ps()
1539 _mm_cvt_pi2ps(__m128 __a, __m64 __b) in _mm_cvt_pi2ps()
1577 _mm_loadh_pi(__m128 __a, const __m64 *__p) in _mm_loadh_pi()
1604 _mm_loadl_pi(__m128 __a, const __m64 *__p) in _mm_loadl_pi()
[all …]
/external/XNNPACK/src/f32-ibilinear-chw/gen/
Dsse-p8.c59 const __m128 vtltr0 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl0); in xnn_f32_ibilinear_chw_ukernel__sse_p8()
60 const __m128 vblbr0 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl0); in xnn_f32_ibilinear_chw_ukernel__sse_p8()
61 const __m128 vtltr2 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl2); in xnn_f32_ibilinear_chw_ukernel__sse_p8()
62 const __m128 vblbr2 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl2); in xnn_f32_ibilinear_chw_ukernel__sse_p8()
63 const __m128 vtltr4 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl4); in xnn_f32_ibilinear_chw_ukernel__sse_p8()
64 const __m128 vblbr4 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl4); in xnn_f32_ibilinear_chw_ukernel__sse_p8()
65 const __m128 vtltr6 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl6); in xnn_f32_ibilinear_chw_ukernel__sse_p8()
66 const __m128 vblbr6 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl6); in xnn_f32_ibilinear_chw_ukernel__sse_p8()
73 const __m128 vtltr01 = _mm_loadh_pi(vtltr0, (const __m64*) itl1); in xnn_f32_ibilinear_chw_ukernel__sse_p8()
74 const __m128 vblbr01 = _mm_loadh_pi(vblbr0, (const __m64*) ibl1); in xnn_f32_ibilinear_chw_ukernel__sse_p8()
[all …]
Dsse-p4.c50 const __m128 vtltr0 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl0); in xnn_f32_ibilinear_chw_ukernel__sse_p4()
51 const __m128 vblbr0 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl0); in xnn_f32_ibilinear_chw_ukernel__sse_p4()
52 const __m128 vtltr2 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl2); in xnn_f32_ibilinear_chw_ukernel__sse_p4()
53 const __m128 vblbr2 = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl2); in xnn_f32_ibilinear_chw_ukernel__sse_p4()
58 const __m128 vtltr01 = _mm_loadh_pi(vtltr0, (const __m64*) itl1); in xnn_f32_ibilinear_chw_ukernel__sse_p4()
59 const __m128 vblbr01 = _mm_loadh_pi(vblbr0, (const __m64*) ibl1); in xnn_f32_ibilinear_chw_ukernel__sse_p4()
60 const __m128 vtltr23 = _mm_loadh_pi(vtltr2, (const __m64*) itl3); in xnn_f32_ibilinear_chw_ukernel__sse_p4()
61 const __m128 vblbr23 = _mm_loadh_pi(vblbr2, (const __m64*) ibl3); in xnn_f32_ibilinear_chw_ukernel__sse_p4()
96 …8 vtltr = _mm_loadh_pi(_mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl0), (const __m64*) itl1); in xnn_f32_ibilinear_chw_ukernel__sse_p4()
97 …8 vblbr = _mm_loadh_pi(_mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl0), (const __m64*) ibl1); in xnn_f32_ibilinear_chw_ukernel__sse_p4()
[all …]
/external/XNNPACK/src/f32-ibilinear-chw/
Dsse.c.in46 … const __m128 vtltr${ABC[P]} = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl${ABC[P]});
47 … const __m128 vblbr${ABC[P]} = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl${ABC[P]});
54 … const __m128 vtltr${ABC[P:P+2]} = _mm_loadh_pi(vtltr${ABC[P]}, (const __m64*) itl${ABC[P+1]});
55 … const __m128 vblbr${ABC[P:P+2]} = _mm_loadh_pi(vblbr${ABC[P]}, (const __m64*) ibl${ABC[P+1]});
94 const __m128 vtltr${ABC[P]} = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl${P});
95 const __m128 vblbr${ABC[P]} = _mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl${P});
101 const __m128 vtltr${ABC[P:P+2]} = _mm_loadh_pi(vtltr${ABC[P]}, (const __m64*) itl${P+1});
102 const __m128 vblbr${ABC[P:P+2]} = _mm_loadh_pi(vblbr${ABC[P]}, (const __m64*) ibl${P+1});
136 …8 vtltr = _mm_loadh_pi(_mm_loadl_pi(_mm_undefined_ps(), (const __m64*) itl0), (const __m64*) itl1);
137 …8 vblbr = _mm_loadh_pi(_mm_loadl_pi(_mm_undefined_ps(), (const __m64*) ibl0), (const __m64*) ibl1);
[all …]
/external/clang/test/Sema/
Dx86-builtin-palignr.c6 __m64 test1(__m64 a, __m64 b, int c) { in test1()
/external/libmpeg2/common/x86/
Dicv_sad_ssse3.c114 src_r0 = _mm_loadh_pi (src_r0, (__m64 *) (pu1_src)); in icv_sad_8x4_ssse3()
117 src_r1 = _mm_loadh_pi (src_r1, (__m64 *) (pu1_src)); in icv_sad_8x4_ssse3()
128 ref_r0 = _mm_loadh_pi (ref_r0, (__m64 *) (pu1_ref)); in icv_sad_8x4_ssse3()
131 ref_r1 = _mm_loadh_pi (ref_r1, (__m64 *) (pu1_ref)); in icv_sad_8x4_ssse3()
/external/fec/
Dviterbi39_sse.c13 typedef union { signed short s[256]; __m64 v[64];} metric_t;
15 static union branchtab39 { unsigned short s[128]; __m64 v[32];} Branchtab39[3];
128 __m64 sym0v,sym1v,sym2v; in update_viterbi39_blk_sse()
139 __m64 decision0,decision1,metric,m_metric,m0,m1,m2,m3,survivor0,survivor1; in update_viterbi39_blk_sse()
174 __m64 adjustv; in update_viterbi39_blk_sse()
175 union { __m64 v; signed short w[4]; } t; in update_viterbi39_blk_sse()
Dviterbi615_sse.c13 typedef union { signed short s[16384]; __m64 v[4096];} metric_t;
15 static union branchtab615 { unsigned short s[8192]; __m64 v[2048];} Branchtab615[6];
124 __m64 sym0v,sym1v,sym2v,sym3v,sym4v,sym5v; in update_viterbi615_blk_sse()
138 __m64 decision0,decision1,metric,m_metric,m0,m1,m2,m3,survivor0,survivor1; in update_viterbi615_blk_sse()
175 __m64 adjustv; in update_viterbi615_blk_sse()
176 union { __m64 v; signed short w[4]; } t; in update_viterbi615_blk_sse()
/external/XNNPACK/src/f32-gemm/gen/
D8x8-minmax-fma3-broadcast.c214 _mm_storel_pi((__m64*) c7, vacc7x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast()
215 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast()
216 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast()
217 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast()
218 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast()
219 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast()
220 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast()
221 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_gemm_minmax_ukernel_8x8__fma3_broadcast()
D7x8-minmax-fma3-broadcast.c195 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_gemm_minmax_ukernel_7x8__fma3_broadcast()
196 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_gemm_minmax_ukernel_7x8__fma3_broadcast()
197 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_gemm_minmax_ukernel_7x8__fma3_broadcast()
198 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_gemm_minmax_ukernel_7x8__fma3_broadcast()
199 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_gemm_minmax_ukernel_7x8__fma3_broadcast()
200 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_gemm_minmax_ukernel_7x8__fma3_broadcast()
201 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_gemm_minmax_ukernel_7x8__fma3_broadcast()
D7x8-minmax-avx-broadcast.c195 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_gemm_minmax_ukernel_7x8__avx_broadcast()
196 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_gemm_minmax_ukernel_7x8__avx_broadcast()
197 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_gemm_minmax_ukernel_7x8__avx_broadcast()
198 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_gemm_minmax_ukernel_7x8__avx_broadcast()
199 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_gemm_minmax_ukernel_7x8__avx_broadcast()
200 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_gemm_minmax_ukernel_7x8__avx_broadcast()
201 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_gemm_minmax_ukernel_7x8__avx_broadcast()
/external/XNNPACK/src/f32-igemm/gen/
D8x8-minmax-fma3-broadcast.c241 _mm_storel_pi((__m64*) c7, vacc7x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast()
242 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast()
243 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast()
244 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast()
245 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast()
246 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast()
247 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast()
248 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_igemm_minmax_ukernel_8x8__fma3_broadcast()
D7x8-minmax-avx-broadcast.c220 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_igemm_minmax_ukernel_7x8__avx_broadcast()
221 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_igemm_minmax_ukernel_7x8__avx_broadcast()
222 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_igemm_minmax_ukernel_7x8__avx_broadcast()
223 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_igemm_minmax_ukernel_7x8__avx_broadcast()
224 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_igemm_minmax_ukernel_7x8__avx_broadcast()
225 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_igemm_minmax_ukernel_7x8__avx_broadcast()
226 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_igemm_minmax_ukernel_7x8__avx_broadcast()
D7x8-minmax-fma3-broadcast.c220 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_igemm_minmax_ukernel_7x8__fma3_broadcast()
221 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_igemm_minmax_ukernel_7x8__fma3_broadcast()
222 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_igemm_minmax_ukernel_7x8__fma3_broadcast()
223 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_igemm_minmax_ukernel_7x8__fma3_broadcast()
224 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_igemm_minmax_ukernel_7x8__fma3_broadcast()
225 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_igemm_minmax_ukernel_7x8__fma3_broadcast()
226 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_igemm_minmax_ukernel_7x8__fma3_broadcast()
/external/XNNPACK/src/f32-gemm/gen-inc/
D8x8inc-minmax-fma3-broadcast.c216 _mm_storel_pi((__m64*) c7, vacc7x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast()
217 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast()
218 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast()
219 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast()
220 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast()
221 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast()
222 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast()
223 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_gemminc_minmax_ukernel_8x8__fma3_broadcast()
D7x8inc-minmax-avx-broadcast.c197 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__avx_broadcast()
198 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__avx_broadcast()
199 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__avx_broadcast()
200 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__avx_broadcast()
201 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__avx_broadcast()
202 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__avx_broadcast()
203 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__avx_broadcast()
D7x8inc-minmax-fma3-broadcast.c197 _mm_storel_pi((__m64*) c6, vacc6x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__fma3_broadcast()
198 _mm_storel_pi((__m64*) c5, vacc5x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__fma3_broadcast()
199 _mm_storel_pi((__m64*) c4, vacc4x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__fma3_broadcast()
200 _mm_storel_pi((__m64*) c3, vacc3x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__fma3_broadcast()
201 _mm_storel_pi((__m64*) c2, vacc2x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__fma3_broadcast()
202 _mm_storel_pi((__m64*) c1, vacc1x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__fma3_broadcast()
203 _mm_storel_pi((__m64*) c0, vacc0x0123); in xnn_f32_gemminc_minmax_ukernel_7x8__fma3_broadcast()
/external/pffft/
Dsse2neon.h148 typedef int64x1_t __m64; typedef
509 FORCE_INLINE __m128i _mm_setr_epi64(__m64 e1, __m64 e0) in _mm_setr_epi64()
624 FORCE_INLINE __m128i _mm_set1_epi64(__m64 _i) in _mm_set1_epi64()
656 FORCE_INLINE __m128i _mm_set_epi64(__m64 i1, __m64 i2) in _mm_set_epi64()
747 FORCE_INLINE void _mm_storel_pi(__m64 *p, __m128 a) in _mm_storel_pi()
759 FORCE_INLINE void _mm_storeh_pi(__m64 *p, __m128 a) in _mm_storeh_pi()
794 FORCE_INLINE __m128 _mm_loadl_pi(__m128 a, __m64 const *p) in _mm_loadl_pi()
826 FORCE_INLINE __m128 _mm_loadh_pi(__m128 a, __m64 const *p) in _mm_loadh_pi()
1251 FORCE_INLINE __m64 _mm_abs_pi32(__m64 a) in _mm_abs_pi32()
1265 FORCE_INLINE __m64 _mm_abs_pi16(__m64 a) in _mm_abs_pi16()
[all …]

12345678910>>...31