/external/clang/lib/Headers/ |
D | xmmintrin.h | 35 typedef float __m128 __attribute__((__vector_size__(16))); typedef 43 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 44 _mm_add_ss(__m128 __a, __m128 __b) in _mm_add_ss() 50 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 51 _mm_add_ps(__m128 __a, __m128 __b) in _mm_add_ps() 56 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 57 _mm_sub_ss(__m128 __a, __m128 __b) in _mm_sub_ss() 63 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 64 _mm_sub_ps(__m128 __a, __m128 __b) in _mm_sub_ps() 69 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) [all …]
|
D | fma4intrin.h | 37 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 38 _mm_macc_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_macc_ps() 40 return (__m128)__builtin_ia32_vfmaddps(__A, __B, __C); in _mm_macc_ps() 49 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 50 _mm_macc_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_macc_ss() 52 return (__m128)__builtin_ia32_vfmaddss(__A, __B, __C); in _mm_macc_ss() 61 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 62 _mm_msub_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_msub_ps() 64 return (__m128)__builtin_ia32_vfmsubps(__A, __B, __C); in _mm_msub_ps() 73 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) [all …]
|
D | fmaintrin.h | 35 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 36 _mm_fmadd_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_fmadd_ps() 38 return (__m128)__builtin_ia32_vfmaddps(__A, __B, __C); in _mm_fmadd_ps() 47 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 48 _mm_fmadd_ss(__m128 __A, __m128 __B, __m128 __C) in _mm_fmadd_ss() 50 return (__m128)__builtin_ia32_vfmaddss(__A, __B, __C); in _mm_fmadd_ss() 59 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 60 _mm_fmsub_ps(__m128 __A, __m128 __B, __m128 __C) in _mm_fmsub_ps() 62 return (__m128)__builtin_ia32_vfmsubps(__A, __B, __C); in _mm_fmsub_ps() 71 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) [all …]
|
D | pmmintrin.h | 39 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 40 _mm_addsub_ps(__m128 __a, __m128 __b) in _mm_addsub_ps() 45 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 46 _mm_hadd_ps(__m128 __a, __m128 __b) in _mm_hadd_ps() 51 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 52 _mm_hsub_ps(__m128 __a, __m128 __b) in _mm_hsub_ps() 57 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 58 _mm_movehdup_ps(__m128 __a) in _mm_movehdup_ps() 63 static __inline__ __m128 __attribute__((__always_inline__, __nodebug__)) 64 _mm_moveldup_ps(__m128 __a) in _mm_moveldup_ps()
|
D | avx512erintrin.h | 130 (__m128)__builtin_ia32_rsqrt28ss_mask((__v4sf)(__m128)(A), \ 131 (__v4sf)(__m128)(B), \ 136 (__m128)__builtin_ia32_rsqrt28ss_mask((__v4sf)(__m128)(A), \ 137 (__v4sf)(__m128)(B), \ 138 (__v4sf)(__m128)(S), \ 142 (__m128)__builtin_ia32_rsqrt28ss_mask((__v4sf)(__m128)(A), \ 143 (__v4sf)(__m128)(B), \ 233 (__m128)__builtin_ia32_rcp28ss_mask((__v4sf)(__m128)(A), \ 234 (__v4sf)(__m128)(B), \ 239 (__m128)__builtin_ia32_rcp28ss_mask((__v4sf)(__m128)(A), \ [all …]
|
/external/clang/test/CodeGen/ |
D | sse-builtins.c | 7 __m128 test_rsqrt_ss(__m128 x) { in test_rsqrt_ss() 17 __m128 test_rcp_ss(__m128 x) { in test_rcp_ss() 27 __m128 test_sqrt_ss(__m128 x) { in test_sqrt_ss() 37 __m128 test_loadl_pi(__m128 x, void* y) { in test_loadl_pi() 45 __m128 test_loadh_pi(__m128 x, void* y) { in test_loadh_pi() 53 __m128 test_load_ss(void* y) { in test_load_ss() 59 __m128 test_load1_ps(void* y) { in test_load1_ps() 65 void test_store_ss(__m128 x, void* y) { in test_store_ss() 71 __m128d test_load1_pd(__m128 x, void* y) { in test_load1_pd() 77 __m128d test_loadr_pd(__m128 x, void* y) { in test_loadr_pd() [all …]
|
D | fma4-builtins.c | 8 __m128 test_mm_macc_ps(__m128 a, __m128 b, __m128 c) { in test_mm_macc_ps() 18 __m128 test_mm_macc_ss(__m128 a, __m128 b, __m128 c) { in test_mm_macc_ss() 28 __m128 test_mm_msub_ps(__m128 a, __m128 b, __m128 c) { in test_mm_msub_ps() 38 __m128 test_mm_msub_ss(__m128 a, __m128 b, __m128 c) { in test_mm_msub_ss() 48 __m128 test_mm_nmacc_ps(__m128 a, __m128 b, __m128 c) { in test_mm_nmacc_ps() 58 __m128 test_mm_nmacc_ss(__m128 a, __m128 b, __m128 c) { in test_mm_nmacc_ss() 68 __m128 test_mm_nmsub_ps(__m128 a, __m128 b, __m128 c) { in test_mm_nmsub_ps() 78 __m128 test_mm_nmsub_ss(__m128 a, __m128 b, __m128 c) { in test_mm_nmsub_ss() 88 __m128 test_mm_maddsub_ps(__m128 a, __m128 b, __m128 c) { in test_mm_maddsub_ps() 98 __m128 test_mm_msubadd_ps(__m128 a, __m128 b, __m128 c) { in test_mm_msubadd_ps()
|
D | fma-builtins.c | 8 __m128 test_mm_fmadd_ps(__m128 a, __m128 b, __m128 c) { in test_mm_fmadd_ps() 18 __m128 test_mm_fmadd_ss(__m128 a, __m128 b, __m128 c) { in test_mm_fmadd_ss() 28 __m128 test_mm_fmsub_ps(__m128 a, __m128 b, __m128 c) { in test_mm_fmsub_ps() 38 __m128 test_mm_fmsub_ss(__m128 a, __m128 b, __m128 c) { in test_mm_fmsub_ss() 48 __m128 test_mm_fnmadd_ps(__m128 a, __m128 b, __m128 c) { in test_mm_fnmadd_ps() 58 __m128 test_mm_fnmadd_ss(__m128 a, __m128 b, __m128 c) { in test_mm_fnmadd_ss() 68 __m128 test_mm_fnmsub_ps(__m128 a, __m128 b, __m128 c) { in test_mm_fnmsub_ps() 78 __m128 test_mm_fnmsub_ss(__m128 a, __m128 b, __m128 c) { in test_mm_fnmsub_ss() 88 __m128 test_mm_fmaddsub_ps(__m128 a, __m128 b, __m128 c) { in test_mm_fmaddsub_ps() 98 __m128 test_mm_fmsubadd_ps(__m128 a, __m128 b, __m128 c) { in test_mm_fmsubadd_ps()
|
D | avx512er-builtins.c | 77 __m128 test_mm_rsqrt28_round_ss(__m128 a, __m128 b) { in test_mm_rsqrt28_round_ss() 83 __m128 test_mm_mask_rsqrt28_round_ss(__m128 s, __mmask16 m, __m128 a, __m128 b) { in test_mm_mask_rsqrt28_round_ss() 89 __m128 test_mm_maskz_rsqrt28_round_ss(__mmask16 m, __m128 a, __m128 b) { in test_mm_maskz_rsqrt28_round_ss() 95 __m128 test_mm_rsqrt28_ss(__m128 a, __m128 b) { in test_mm_rsqrt28_ss() 101 __m128 test_mm_mask_rsqrt28_ss(__m128 s, __mmask16 m, __m128 a, __m128 b) { in test_mm_mask_rsqrt28_ss() 107 __m128 test_mm_maskz_rsqrt28_ss(__mmask16 m, __m128 a, __m128 b) { in test_mm_maskz_rsqrt28_ss() 203 __m128 test_mm_rcp28_round_ss(__m128 a, __m128 b) { in test_mm_rcp28_round_ss() 209 __m128 test_mm_mask_rcp28_round_ss(__m128 s, __mmask16 m, __m128 a, __m128 b) { in test_mm_mask_rcp28_round_ss() 215 __m128 test_mm_maskz_rcp28_round_ss(__mmask16 m, __m128 a, __m128 b) { in test_mm_maskz_rcp28_round_ss() 221 __m128 test_mm_rcp28_ss(__m128 a, __m128 b) { in test_mm_rcp28_ss() [all …]
|
D | sse.c | 6 __m128 test_mm_slli_si128(__m128 a) { in test_mm_slli_si128() 12 __m128 test_mm_slli_si128_0(__m128 a) { in test_mm_slli_si128_0() 18 __m128 test_mm_slli_si128_16(__m128 a) { in test_mm_slli_si128_16() 24 __m128 test_mm_srli_si128(__m128 a) { in test_mm_srli_si128() 30 __m128 test_mm_srli_si128_0(__m128 a) { in test_mm_srli_si128_0() 36 __m128 test_mm_srli_si128_16(__m128 a) { in test_mm_srli_si128_16()
|
D | avx-cmp-builtins.c | 18 __m128d test_cmp_ps(__m128 a, __m128 b) { in test_cmp_ps() 42 __m128d test_cmp_ss(__m128 a, __m128 b) { in test_cmp_ss() 48 __m128 test_cmpgt_ss(__m128 a, __m128 b) { in test_cmpgt_ss() 54 __m128 test_cmpge_ss(__m128 a, __m128 b) { in test_cmpge_ss() 60 __m128 test_cmpngt_ss(__m128 a, __m128 b) { in test_cmpngt_ss() 66 __m128 test_cmpnge_ss(__m128 a, __m128 b) { in test_cmpnge_ss()
|
D | avx-shuffle-builtins.c | 30 __m128 test_mm_permute_ps(__m128 a) { in test_mm_permute_ps() 37 __m128 test_mm_permute_ps2(__m128 a) { in test_mm_permute_ps2() 67 __m128 103 __m256 test_mm256_insertf128_ps_0(__m256 a, __m128 b) { in test_mm256_insertf128_ps_0() 121 __m256 test_mm256_insertf128_ps_1(__m256 a, __m128 b) { in test_mm256_insertf128_ps_1() 141 __m128 test_mm256_extractf128_ps_0(__m256 a) { in test_mm256_extractf128_ps_0() 159 __m128 test_mm256_extractf128_ps_1(__m256 a) { in test_mm256_extractf128_ps_1()
|
/external/webrtc/src/modules/audio_processing/aec/ |
D | aec_rdft_sse2.c | 22 const __m128 mm_swap_sign = _mm_load_ps(k_swap_sign); in cft1st_128_SSE2() 26 __m128 a00v = _mm_loadu_ps(&a[j + 0]); in cft1st_128_SSE2() 27 __m128 a04v = _mm_loadu_ps(&a[j + 4]); in cft1st_128_SSE2() 28 __m128 a08v = _mm_loadu_ps(&a[j + 8]); in cft1st_128_SSE2() 29 __m128 a12v = _mm_loadu_ps(&a[j + 12]); in cft1st_128_SSE2() 30 __m128 a01v = _mm_shuffle_ps(a00v, a08v, _MM_SHUFFLE(1, 0, 1 ,0)); in cft1st_128_SSE2() 31 __m128 a23v = _mm_shuffle_ps(a00v, a08v, _MM_SHUFFLE(3, 2, 3 ,2)); in cft1st_128_SSE2() 32 __m128 a45v = _mm_shuffle_ps(a04v, a12v, _MM_SHUFFLE(1, 0, 1 ,0)); in cft1st_128_SSE2() 33 __m128 a67v = _mm_shuffle_ps(a04v, a12v, _MM_SHUFFLE(3, 2, 3 ,2)); in cft1st_128_SSE2() 35 const __m128 wk1rv = _mm_load_ps(&rdft_wk1r[k2]); in cft1st_128_SSE2() [all …]
|
D | aec_core_sse2.c | 48 const __m128 xfBuf_re = _mm_loadu_ps(&aec->xfBuf[0][xPos + j]); in FilterFarSSE2() 49 const __m128 xfBuf_im = _mm_loadu_ps(&aec->xfBuf[1][xPos + j]); in FilterFarSSE2() 50 const __m128 wfBuf_re = _mm_loadu_ps(&aec->wfBuf[0][pos + j]); in FilterFarSSE2() 51 const __m128 wfBuf_im = _mm_loadu_ps(&aec->wfBuf[1][pos + j]); in FilterFarSSE2() 52 const __m128 yf_re = _mm_loadu_ps(&yf[0][j]); in FilterFarSSE2() 53 const __m128 yf_im = _mm_loadu_ps(&yf[1][j]); in FilterFarSSE2() 54 const __m128 a = _mm_mul_ps(xfBuf_re, wfBuf_re); in FilterFarSSE2() 55 const __m128 b = _mm_mul_ps(xfBuf_im, wfBuf_im); in FilterFarSSE2() 56 const __m128 c = _mm_mul_ps(xfBuf_re, wfBuf_im); in FilterFarSSE2() 57 const __m128 d = _mm_mul_ps(xfBuf_im, wfBuf_re); in FilterFarSSE2() [all …]
|
/external/speex/libspeex/ |
D | vq_sse.h | 36 void vq_nbest(spx_word16_t *_in, const __m128 *codebook, int len, int entries, __m128 *E, int N, in… in vq_nbest() 40 VARDECL(__m128 *in); in vq_nbest() 41 __m128 half; in vq_nbest() 45 ALLOC(in, len, __m128); in vq_nbest() 50 __m128 d = _mm_mul_ps(E[i], half); in vq_nbest() 75 void vq_nbest_sign(spx_word16_t *_in, const __m128 *codebook, int len, int entries, __m128 *E, int … in vq_nbest_sign() 79 VARDECL(__m128 *in); in vq_nbest_sign() 80 __m128 half; in vq_nbest_sign() 84 ALLOC(in, len, __m128); in vq_nbest_sign() 89 __m128 d = _mm_setzero_ps(); in vq_nbest_sign()
|
D | cb_search_sse.h | 37 static inline void _spx_mm_getr_ps (__m128 U, float *__Z, float *__Y, float *__X, float *__W) in _spx_mm_getr_ps() 41 __m128 __v; in _spx_mm_getr_ps() 54 …ok(const signed char *shape_cb, const spx_sig_t *_r, float *resp, __m128 *resp2, __m128 *E, int sh… in compute_weighted_codebook() 57 __m128 resj, EE; in compute_weighted_codebook() 58 VARDECL(__m128 *r); in compute_weighted_codebook() 59 VARDECL(__m128 *shape); in compute_weighted_codebook() 60 ALLOC(r, subvect_size, __m128); in compute_weighted_codebook() 61 ALLOC(shape, subvect_size, __m128); in compute_weighted_codebook()
|
D | filters_sse.h | 39 __m128 num[3], den[3], mem[3]; in filter_mem16_10() 56 __m128 xx; in filter_mem16_10() 57 __m128 yy; in filter_mem16_10() 92 __m128 num[2], den[2], mem[2]; in filter_mem16_8() 106 __m128 xx; in filter_mem16_8() 107 __m128 yy; in filter_mem16_8() 146 __m128 den[3], mem[3]; in iir_mem16_10() 161 __m128 xx; in iir_mem16_10() 162 __m128 yy; in iir_mem16_10() 195 __m128 den[2], mem[2]; in iir_mem16_8() [all …]
|
D | resample_sse.h | 44 __m128 sum = _mm_setzero_ps(); in inner_product_single() 60 __m128 sum = _mm_setzero_ps(); in interpolate_product_single() 61 __m128 f = _mm_loadu_ps(frac); in interpolate_product_single() 83 __m128 t; in inner_product_double() 94 sum = _mm_add_sd(sum, (__m128d) _mm_movehl_ps((__m128) sum, (__m128) sum)); in inner_product_double() 106 __m128 f = _mm_loadu_ps(frac); in interpolate_product_double() 109 __m128 t; in interpolate_product_double() 123 sum = _mm_add_sd(sum, (__m128d) _mm_movehl_ps((__m128) sum, (__m128) sum)); in interpolate_product_double()
|
D | ltp_sse.h | 42 __m128 sum = _mm_setzero_ps(); in inner_prod() 60 VARDECL(__m128 *x); in pitch_xcorr() 61 VARDECL(__m128 *y); in pitch_xcorr() 65 ALLOC(x, N, __m128); in pitch_xcorr() 66 ALLOC(y, N+L, __m128); in pitch_xcorr() 76 __m128 sum, *xx, *yy; in pitch_xcorr()
|
/external/clang/test/Sema/ |
D | uninit-variables-vectors.c | 4 typedef float __m128 __attribute__((__vector_size__(16))); typedef 5 __m128 _mm_xor_ps(__m128 a, __m128 b); 6 __m128 _mm_loadu_ps(const float *p); 9 …__m128 x, y, z, w, X; // expected-note {{variable 'x' is declared here}} expected-note {{variable … in test1()
|
D | builtins-x86.c | 4 typedef float __m128 __attribute__((__vector_size__(16))); typedef 13 __m128 test__builtin_ia32_cmpps(__m128 __a, __m128 __b) { in test__builtin_ia32_cmpps() 21 __m128 test__builtin_ia32_cmpss(__m128 __a, __m128 __b) { in test__builtin_ia32_cmpss()
|
/external/libvpx/libvpx/vpx_ports/ |
D | emmintrin_compat.h | 18 extern __inline __m128 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 21 return (__m128) __A; in _mm_castpd_ps() 31 _mm_castps_pd(__m128 __A) in _mm_castps_pd() 37 _mm_castps_si128(__m128 __A) in _mm_castps_si128() 42 extern __inline __m128 __attribute__((__gnu_inline__, __always_inline__, __artificial__)) 45 return (__m128) __A; in _mm_castsi128_ps()
|
/external/libopus/celt/x86/ |
D | pitch_sse.h | 42 __m128 xsum1, xsum2; in xcorr_kernel() 48 __m128 x0 = _mm_loadu_ps(x+j); in xcorr_kernel() 49 __m128 yj = _mm_loadu_ps(y+j); in xcorr_kernel() 50 __m128 y3 = _mm_loadu_ps(y+j+3); in xcorr_kernel() 79 __m128 xsum1, xsum2; in dual_inner_prod() 84 __m128 xi = _mm_loadu_ps(x+i); in dual_inner_prod() 85 __m128 y1i = _mm_loadu_ps(y01+i); in dual_inner_prod() 86 __m128 y2i = _mm_loadu_ps(y02+i); in dual_inner_prod() 109 __m128 x0v; in comb_filter_const() 110 __m128 g10v, g11v, g12v; in comb_filter_const() [all …]
|
/external/eigen/Eigen/src/Geometry/arch/ |
D | Geometry_SSE.h | 23 const __m128 mask = _mm_castsi128_ps(_mm_setr_epi32(0,0,0,0x80000000)); 25 __m128 a = _a.coeffs().template packet<Aligned>(0); 26 __m128 b = _b.coeffs().template packet<Aligned>(0); 27 __m128 flip1 = _mm_xor_ps(_mm_mul_ps(vec4f_swizzle1(a,1,2,0,2), 29 __m128 flip2 = _mm_xor_ps(_mm_mul_ps(vec4f_swizzle1(a,3,3,3,1), 46 __m128 a = lhs.template packet<VectorLhs::Flags&AlignedBit ? Aligned : Unaligned>(0); 47 __m128 b = rhs.template packet<VectorRhs::Flags&AlignedBit ? Aligned : Unaligned>(0); 48 __m128 mul1=_mm_mul_ps(vec4f_swizzle1(a,1,2,0,3),vec4f_swizzle1(b,2,0,1,3)); 49 __m128 mul2=_mm_mul_ps(vec4f_swizzle1(a,2,0,1,3),vec4f_swizzle1(b,1,2,0,3));
|
/external/libmpeg2/common/x86/ |
D | icv_sad_ssse3.c | 98 __m128 src_r0, src_r1; in icv_sad_8x4_ssse3() 99 __m128 ref_r0, ref_r1; in icv_sad_8x4_ssse3() 108 src_r0 = (__m128)_mm_loadl_epi64((__m128i *) (pu1_src)); in icv_sad_8x4_ssse3() 111 src_r1 = (__m128)_mm_loadl_epi64((__m128i *) (pu1_src)); in icv_sad_8x4_ssse3() 122 ref_r0 = (__m128)_mm_loadl_epi64((__m128i *) (pu1_ref)); in icv_sad_8x4_ssse3() 125 ref_r1 = (__m128)_mm_loadl_epi64((__m128i *) (pu1_ref)); in icv_sad_8x4_ssse3()
|