Home
last modified time | relevance | path

Searched refs:xmm6 (Results 1 – 25 of 265) sorted by relevance

1234567891011

/external/flac/libFLAC/
Dlpc_intrin_sse2.c422 __m128i xmm0, xmm1, xmm2, xmm3, xmm4, xmm5, xmm6, xmm7; in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() local
447 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-10)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
448 xmm6 = _mm_shuffle_epi32(xmm6, _MM_SHUFFLE(2,0,3,1)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
449 xmm6 = _mm_mul_epu32(xmm6, xmm4); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
450 xmm7 = _mm_add_epi32(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
454 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-8)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
455 xmm6 = _mm_shuffle_epi32(xmm6, _MM_SHUFFLE(2,0,3,1)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
456 xmm6 = _mm_mul_epu32(xmm6, xmm3); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
457 xmm7 = _mm_add_epi32(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
461 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-6)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
[all …]
Dlpc_intrin_sse41.c69 __m128i xmm0, xmm1, xmm2, xmm3, xmm4, xmm5, xmm6, xmm7; in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() local
94 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-10)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
95 xmm6 = _mm_shuffle_epi32(xmm6, _MM_SHUFFLE(2,0,3,1)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
96 xmm6 = _mm_mul_epi32(xmm6, xmm4); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
97 xmm7 = _mm_add_epi64(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
101 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-8)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
102 xmm6 = _mm_shuffle_epi32(xmm6, _MM_SHUFFLE(2,0,3,1)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
103 xmm6 = _mm_mul_epi32(xmm6, xmm3); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
104 xmm7 = _mm_add_epi64(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
108 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-6)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
[all …]
/external/boringssl/mac-x86_64/crypto/fipsmodule/
Dbsaes-x86_64.S26 pxor %xmm8,%xmm6
38 pxor %xmm6,%xmm5
42 pxor %xmm5,%xmm6
67 pxor %xmm6,%xmm4
71 pxor %xmm4,%xmm6
95 pxor %xmm6,%xmm2
99 pxor %xmm2,%xmm6
134 pxor 112(%rax),%xmm6
148 pxor %xmm6,%xmm2
149 pxor %xmm4,%xmm6
[all …]
Daesni-gcm-x86_64.S35 vpclmulqdq $0x01,%xmm3,%xmm7,%xmm6
68 vpxor %xmm5,%xmm6,%xmm6
96 vpxor %xmm1,%xmm6,%xmm6
99 vpxor %xmm2,%xmm6,%xmm6
114 vpxor %xmm2,%xmm6,%xmm6
117 vpxor %xmm3,%xmm6,%xmm6
136 vpxor %xmm3,%xmm6,%xmm6
139 vpxor %xmm5,%xmm6,%xmm6
158 vpxor %xmm5,%xmm6,%xmm6
161 vpxor %xmm1,%xmm6,%xmm6
[all …]
/external/boringssl/linux-x86_64/crypto/fipsmodule/
Dbsaes-x86_64.S28 pxor %xmm8,%xmm6
40 pxor %xmm6,%xmm5
44 pxor %xmm5,%xmm6
69 pxor %xmm6,%xmm4
73 pxor %xmm4,%xmm6
97 pxor %xmm6,%xmm2
101 pxor %xmm2,%xmm6
136 pxor 112(%rax),%xmm6
150 pxor %xmm6,%xmm2
151 pxor %xmm4,%xmm6
[all …]
Daesni-gcm-x86_64.S35 vpclmulqdq $0x01,%xmm3,%xmm7,%xmm6
68 vpxor %xmm5,%xmm6,%xmm6
96 vpxor %xmm1,%xmm6,%xmm6
99 vpxor %xmm2,%xmm6,%xmm6
114 vpxor %xmm2,%xmm6,%xmm6
117 vpxor %xmm3,%xmm6,%xmm6
136 vpxor %xmm3,%xmm6,%xmm6
139 vpxor %xmm5,%xmm6,%xmm6
158 vpxor %xmm5,%xmm6,%xmm6
161 vpxor %xmm1,%xmm6,%xmm6
[all …]
/external/boringssl/win-x86_64/crypto/fipsmodule/
Dbsaes-x86_64.asm30 pxor xmm6,xmm8
42 pxor xmm5,xmm6
46 pxor xmm6,xmm5
71 pxor xmm4,xmm6
75 pxor xmm6,xmm4
99 pxor xmm2,xmm6
103 pxor xmm6,xmm2
138 pxor xmm6,XMMWORD[112+rax]
152 pxor xmm2,xmm6
153 pxor xmm6,xmm4
[all …]
Daesni-gcm-x86_64.asm39 vpclmulqdq xmm6,xmm7,xmm3,0x01
72 vpxor xmm6,xmm6,xmm5
100 vpxor xmm6,xmm6,xmm1
103 vpxor xmm6,xmm6,xmm2
118 vpxor xmm6,xmm6,xmm2
121 vpxor xmm6,xmm6,xmm3
140 vpxor xmm6,xmm6,xmm3
143 vpxor xmm6,xmm6,xmm5
162 vpxor xmm6,xmm6,xmm5
165 vpxor xmm6,xmm6,xmm1
[all …]
/external/boringssl/mac-x86_64/crypto/cipher_extra/
Daes128gcmsiv-x86_64.S144 vpclmulqdq $0x10,(%rdi,%r11,1),%xmm0,%xmm6
145 vpxor %xmm6,%xmm5,%xmm5
159 vpclmulqdq $0x00,(%rdi,%r11,1),%xmm0,%xmm6
160 vpxor %xmm6,%xmm3,%xmm3
161 vpclmulqdq $0x11,(%rdi,%r11,1),%xmm0,%xmm6
162 vpxor %xmm6,%xmm4,%xmm4
163 vpclmulqdq $0x01,(%rdi,%r11,1),%xmm0,%xmm6
164 vpxor %xmm6,%xmm5,%xmm5
165 vpclmulqdq $0x10,(%rdi,%r11,1),%xmm0,%xmm6
166 vpxor %xmm6,%xmm5,%xmm5
[all …]
/external/boringssl/linux-x86_64/crypto/cipher_extra/
Daes128gcmsiv-x86_64.S144 vpclmulqdq $0x10,(%rdi,%r11,1),%xmm0,%xmm6
145 vpxor %xmm6,%xmm5,%xmm5
159 vpclmulqdq $0x00,(%rdi,%r11,1),%xmm0,%xmm6
160 vpxor %xmm6,%xmm3,%xmm3
161 vpclmulqdq $0x11,(%rdi,%r11,1),%xmm0,%xmm6
162 vpxor %xmm6,%xmm4,%xmm4
163 vpclmulqdq $0x01,(%rdi,%r11,1),%xmm0,%xmm6
164 vpxor %xmm6,%xmm5,%xmm5
165 vpclmulqdq $0x10,(%rdi,%r11,1),%xmm0,%xmm6
166 vpxor %xmm6,%xmm5,%xmm5
[all …]
/external/boringssl/win-x86_64/crypto/cipher_extra/
Daes128gcmsiv-x86_64.asm178 vpclmulqdq xmm6,xmm0,XMMWORD[r11*1+rdi],0x10
179 vpxor xmm5,xmm5,xmm6
193 vpclmulqdq xmm6,xmm0,XMMWORD[r11*1+rdi],0x00
194 vpxor xmm3,xmm3,xmm6
195 vpclmulqdq xmm6,xmm0,XMMWORD[r11*1+rdi],0x11
196 vpxor xmm4,xmm4,xmm6
197 vpclmulqdq xmm6,xmm0,XMMWORD[r11*1+rdi],0x01
198 vpxor xmm5,xmm5,xmm6
199 vpclmulqdq xmm6,xmm0,XMMWORD[r11*1+rdi],0x10
200 vpxor xmm5,xmm5,xmm6
[all …]
/external/libvpx/libvpx/third_party/libyuv/source/
Drotate_win.cc57 movq xmm6, qword ptr [eax] in TransposeWx8_SSSE3()
60 punpcklbw xmm6, xmm7 in TransposeWx8_SSSE3()
62 movdqa xmm7, xmm6 in TransposeWx8_SSSE3()
71 punpcklwd xmm4, xmm6 in TransposeWx8_SSSE3()
73 movdqa xmm6, xmm4 in TransposeWx8_SSSE3()
75 palignr xmm6, xmm6, 8 in TransposeWx8_SSSE3()
85 punpckldq xmm2, xmm6 in TransposeWx8_SSSE3()
86 movdqa xmm6, xmm2 in TransposeWx8_SSSE3()
87 palignr xmm6, xmm6, 8 in TransposeWx8_SSSE3()
90 movq qword ptr [edx + esi], xmm6 in TransposeWx8_SSSE3()
[all …]
/external/libyuv/files/source/
Drotate_win.cc59 movq xmm6, qword ptr [eax] in TransposeWx8_SSSE3()
62 punpcklbw xmm6, xmm7 in TransposeWx8_SSSE3()
64 movdqa xmm7, xmm6 in TransposeWx8_SSSE3()
73 punpcklwd xmm4, xmm6 in TransposeWx8_SSSE3()
75 movdqa xmm6, xmm4 in TransposeWx8_SSSE3()
77 palignr xmm6, xmm6, 8 in TransposeWx8_SSSE3()
87 punpckldq xmm2, xmm6 in TransposeWx8_SSSE3()
88 movdqa xmm6, xmm2 in TransposeWx8_SSSE3()
89 palignr xmm6, xmm6, 8 in TransposeWx8_SSSE3()
92 movq qword ptr [edx + esi], xmm6 in TransposeWx8_SSSE3()
[all …]
/external/boringssl/src/crypto/fipsmodule/aes/asm/
Dvpaes-x86_64.pl81 ## Preserves %xmm6 - %xmm8 so you get some local vectors
345 ## the high bits of %xmm6.
356 movdqa %xmm0, %xmm6 # save short part
358 movhlps %xmm4, %xmm6 # clobber low side with zeros
363 palignr \$8,%xmm6,%xmm0
381 ## %xmm6. The low side's rounds are the same as the
392 movdqa %xmm0, %xmm6 # save cur_lo in xmm6
403 movdqa %xmm6, %xmm7
446 pxor %xmm6, %xmm6
458 ## %xmm6: low side, d c 0 0
[all …]
/external/libvpx/libvpx/vpx_dsp/x86/
Dhighbd_variance_impl_sse2.asm68 pxor xmm6, xmm6 ; clear xmm6 for accumulating sse
93 paddd xmm6, xmm1
100 paddd xmm6, xmm3
107 paddd xmm6, xmm1
112 paddd xmm6, xmm3
131 movdqa xmm4, xmm6
132 punpckldq xmm6, xmm0
137 paddd xmm6, xmm4
143 movdqa xmm4, xmm6
149 paddd xmm6, xmm4
[all …]
/external/capstone/suite/MC/X86/
Dx86_64-xop-encoding.s.cs9 0x8f,0xe9,0x78,0xc7,0xd6 = vphaddwq %xmm6, %xmm2
12 0x8f,0xe9,0x78,0xd7,0x34,0x01 = vphadduwq (%rcx, %rax), %xmm6
17 0x8f,0xe9,0x78,0xdb,0xd6 = vphaddudq %xmm6, %xmm2
27 0x8f,0xe9,0x78,0xc1,0xf5 = vphaddbw %xmm5, %xmm6
37 0x8f,0xe9,0x78,0x80,0xee = vfrczps %xmm6, %xmm5
47 0x8f,0xe9,0x68,0x97,0xf4 = vpshlq %xmm2, %xmm4, %xmm6
49 0x8f,0xe9,0x50,0x97,0x34,0x0a = vpshlq %xmm5, (%rdx, %rcx), %xmm6
61 0x8f,0xe9,0x48,0x9b,0x2c,0x08 = vpshaq %xmm6, (%rax, %rcx), %xmm5
68 0x8f,0xe9,0xe0,0x91,0x30 = vprotw (%rax), %xmm3, %xmm6
83 0x8f,0xe8,0x78,0xc2,0x31,0x2b = vprotd $43, (%rcx), %xmm6
[all …]
/external/boringssl/mac-x86/crypto/chacha/
Dchacha-x86.S411 pshufd $170,%xmm7,%xmm6
420 movdqa %xmm6,-32(%ebp)
430 pshufd $170,%xmm7,%xmm6
438 movdqa %xmm6,-96(%ebp)
449 movdqa -32(%ebp),%xmm6
455 movdqa %xmm6,-32(%ebx)
461 movdqa 96(%ebp),%xmm6
468 movdqa %xmm6,96(%ebx)
472 movdqa %xmm4,%xmm6
482 pxor %xmm0,%xmm6
[all …]
/external/boringssl/linux-x86/crypto/chacha/
Dchacha-x86.S414 pshufd $170,%xmm7,%xmm6
423 movdqa %xmm6,-32(%ebp)
433 pshufd $170,%xmm7,%xmm6
441 movdqa %xmm6,-96(%ebp)
452 movdqa -32(%ebp),%xmm6
458 movdqa %xmm6,-32(%ebx)
464 movdqa 96(%ebp),%xmm6
471 movdqa %xmm6,96(%ebx)
475 movdqa %xmm4,%xmm6
485 pxor %xmm0,%xmm6
[all …]
/external/llvm/test/MC/X86/
Dx86_64-xop-encoding.s37 vphaddwq %xmm6, %xmm2
50 vphadduwq (%rcx,%rax), %xmm6
69 vphaddudq %xmm6, %xmm2
109 vphaddbw %xmm5, %xmm6
149 vfrczps %xmm6, %xmm5
191 vpshlq %xmm2, %xmm4, %xmm6
197 vpshlq %xmm5, (%rdx,%rcx), %xmm6
241 vpshaq %xmm6, (%rax,%rcx), %xmm5
268 vprotw (%rax), %xmm3, %xmm6
317 vprotd $43, (%rcx), %xmm6
[all …]
/external/libvpx/libvpx/vp8/common/x86/
Dloopfilter_sse2.asm54 movdqa xmm6, xmm1 ; q2
58 psubusb xmm2, xmm6 ; q3-=q2
60 psubusb xmm4, xmm6 ; q1-=q2
61 psubusb xmm6, xmm3 ; q2-=q1
63 por xmm4, xmm6 ; abs(q2-q1)
80 movdqa xmm6, [rsi+2*rax] ; p1
84 movlps xmm6, [rsi + rcx] ; p1
88 movhps xmm6, [rdi + rcx]
91 movdqa [rsp+_p1], xmm6 ; store p1
95 movdqa xmm3, xmm6 ; p1
[all …]
/external/boringssl/win-x86/crypto/chacha/
Dchacha-x86.asm423 pshufd xmm6,xmm7,170
432 movdqa [ebp-32],xmm6
442 pshufd xmm6,xmm7,170
450 movdqa [ebp-96],xmm6
461 movdqa xmm6,[ebp-32]
467 movdqa [ebx-32],xmm6
473 movdqa xmm6,[96+ebp]
480 movdqa [96+ebx],xmm6
484 movdqa xmm6,xmm4
494 pxor xmm6,xmm0
[all …]
/external/libjpeg-turbo/simd/
Djccolext-sse2-64.asm299 movdqa xmm6,xmm1
301 punpckhwd xmm6,xmm3
303 movdqa xmm4,xmm6
305 pmaddwd xmm6,[rel PW_F0299_F0337] ; xmm6=ROH*FIX(0.299)+GOH*FIX(0.337)
310 movdqa XMMWORD [wk(5)], xmm6 ; wk(5)=ROH*FIX(0.299)+GOH*FIX(0.337)
313 pxor xmm6,xmm6
315 punpckhwd xmm6,xmm5 ; xmm6=BOH
317 psrld xmm6,1 ; xmm6=BOH*FIX(0.500)
322 paddd xmm4,xmm6
331 movdqa xmm6,xmm0
[all …]
Djccolext-sse2.asm312 movdqa xmm6,xmm1
314 punpckhwd xmm6,xmm3
316 movdqa xmm4,xmm6
318 pmaddwd xmm6,[GOTOFF(eax,PW_F0299_F0337)] ; xmm6=ROH*FIX(0.299)+GOH*FIX(0.337)
323 movdqa XMMWORD [wk(5)], xmm6 ; wk(5)=ROH*FIX(0.299)+GOH*FIX(0.337)
326 pxor xmm6,xmm6
328 punpckhwd xmm6,xmm5 ; xmm6=BOH
330 psrld xmm6,1 ; xmm6=BOH*FIX(0.500)
335 paddd xmm4,xmm6
344 movdqa xmm6,xmm0
[all …]
/external/llvm/test/CodeGen/X86/
Dvec_minmax_uint.ll84 ; SSE2-NEXT: movdqa %xmm1, %xmm6
85 ; SSE2-NEXT: pxor %xmm4, %xmm6
86 ; SSE2-NEXT: movdqa %xmm6, %xmm7
89 ; SSE2-NEXT: pcmpeqd %xmm5, %xmm6
90 ; SSE2-NEXT: pshufd {{.*#+}} xmm5 = xmm6[1,1,3,3]
92 ; SSE2-NEXT: pshufd {{.*#+}} xmm6 = xmm7[1,1,3,3]
93 ; SSE2-NEXT: por %xmm5, %xmm6
108 ; SSE2-NEXT: pand %xmm6, %xmm1
109 ; SSE2-NEXT: pandn %xmm3, %xmm6
110 ; SSE2-NEXT: por %xmm6, %xmm1
[all …]
/external/boringssl/mac-x86/crypto/fipsmodule/
Daesni-x86.S254 pxor %xmm0,%xmm6
307 pxor %xmm0,%xmm6
376 movdqu 64(%esi),%xmm6
391 movups %xmm6,64(%edi)
392 movdqu 64(%esi),%xmm6
407 movups %xmm6,64(%edi)
423 movups 64(%esi),%xmm6
430 movups %xmm6,64(%edi)
478 movdqu 64(%esi),%xmm6
493 movups %xmm6,64(%edi)
[all …]

1234567891011