/external/flac/libFLAC/ |
D | lpc_intrin_sse2.c | 422 __m128i xmm0, xmm1, xmm2, xmm3, xmm4, xmm5, xmm6, xmm7; in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() local 447 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-10)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 448 xmm6 = _mm_shuffle_epi32(xmm6, _MM_SHUFFLE(2,0,3,1)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 449 xmm6 = _mm_mul_epu32(xmm6, xmm4); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 450 xmm7 = _mm_add_epi32(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 454 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-8)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 455 xmm6 = _mm_shuffle_epi32(xmm6, _MM_SHUFFLE(2,0,3,1)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 456 xmm6 = _mm_mul_epu32(xmm6, xmm3); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 457 xmm7 = _mm_add_epi32(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 461 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-6)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() [all …]
|
D | lpc_intrin_sse41.c | 69 __m128i xmm0, xmm1, xmm2, xmm3, xmm4, xmm5, xmm6, xmm7; in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() local 94 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-10)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 95 xmm6 = _mm_shuffle_epi32(xmm6, _MM_SHUFFLE(2,0,3,1)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 96 xmm6 = _mm_mul_epi32(xmm6, xmm4); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 97 xmm7 = _mm_add_epi64(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 101 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-8)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 102 xmm6 = _mm_shuffle_epi32(xmm6, _MM_SHUFFLE(2,0,3,1)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 103 xmm6 = _mm_mul_epi32(xmm6, xmm3); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 104 xmm7 = _mm_add_epi64(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 108 xmm6 = _mm_loadl_epi64((const __m128i*)(data+i-6)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() [all …]
|
/external/boringssl/mac-x86_64/crypto/fipsmodule/ |
D | bsaes-x86_64.S | 26 pxor %xmm8,%xmm6 38 pxor %xmm6,%xmm5 42 pxor %xmm5,%xmm6 67 pxor %xmm6,%xmm4 71 pxor %xmm4,%xmm6 95 pxor %xmm6,%xmm2 99 pxor %xmm2,%xmm6 134 pxor 112(%rax),%xmm6 148 pxor %xmm6,%xmm2 149 pxor %xmm4,%xmm6 [all …]
|
D | aesni-gcm-x86_64.S | 35 vpclmulqdq $0x01,%xmm3,%xmm7,%xmm6 68 vpxor %xmm5,%xmm6,%xmm6 96 vpxor %xmm1,%xmm6,%xmm6 99 vpxor %xmm2,%xmm6,%xmm6 114 vpxor %xmm2,%xmm6,%xmm6 117 vpxor %xmm3,%xmm6,%xmm6 136 vpxor %xmm3,%xmm6,%xmm6 139 vpxor %xmm5,%xmm6,%xmm6 158 vpxor %xmm5,%xmm6,%xmm6 161 vpxor %xmm1,%xmm6,%xmm6 [all …]
|
/external/boringssl/linux-x86_64/crypto/fipsmodule/ |
D | bsaes-x86_64.S | 28 pxor %xmm8,%xmm6 40 pxor %xmm6,%xmm5 44 pxor %xmm5,%xmm6 69 pxor %xmm6,%xmm4 73 pxor %xmm4,%xmm6 97 pxor %xmm6,%xmm2 101 pxor %xmm2,%xmm6 136 pxor 112(%rax),%xmm6 150 pxor %xmm6,%xmm2 151 pxor %xmm4,%xmm6 [all …]
|
D | aesni-gcm-x86_64.S | 35 vpclmulqdq $0x01,%xmm3,%xmm7,%xmm6 68 vpxor %xmm5,%xmm6,%xmm6 96 vpxor %xmm1,%xmm6,%xmm6 99 vpxor %xmm2,%xmm6,%xmm6 114 vpxor %xmm2,%xmm6,%xmm6 117 vpxor %xmm3,%xmm6,%xmm6 136 vpxor %xmm3,%xmm6,%xmm6 139 vpxor %xmm5,%xmm6,%xmm6 158 vpxor %xmm5,%xmm6,%xmm6 161 vpxor %xmm1,%xmm6,%xmm6 [all …]
|
/external/boringssl/win-x86_64/crypto/fipsmodule/ |
D | bsaes-x86_64.asm | 30 pxor xmm6,xmm8 42 pxor xmm5,xmm6 46 pxor xmm6,xmm5 71 pxor xmm4,xmm6 75 pxor xmm6,xmm4 99 pxor xmm2,xmm6 103 pxor xmm6,xmm2 138 pxor xmm6,XMMWORD[112+rax] 152 pxor xmm2,xmm6 153 pxor xmm6,xmm4 [all …]
|
D | aesni-gcm-x86_64.asm | 39 vpclmulqdq xmm6,xmm7,xmm3,0x01 72 vpxor xmm6,xmm6,xmm5 100 vpxor xmm6,xmm6,xmm1 103 vpxor xmm6,xmm6,xmm2 118 vpxor xmm6,xmm6,xmm2 121 vpxor xmm6,xmm6,xmm3 140 vpxor xmm6,xmm6,xmm3 143 vpxor xmm6,xmm6,xmm5 162 vpxor xmm6,xmm6,xmm5 165 vpxor xmm6,xmm6,xmm1 [all …]
|
/external/boringssl/mac-x86_64/crypto/cipher_extra/ |
D | aes128gcmsiv-x86_64.S | 144 vpclmulqdq $0x10,(%rdi,%r11,1),%xmm0,%xmm6 145 vpxor %xmm6,%xmm5,%xmm5 159 vpclmulqdq $0x00,(%rdi,%r11,1),%xmm0,%xmm6 160 vpxor %xmm6,%xmm3,%xmm3 161 vpclmulqdq $0x11,(%rdi,%r11,1),%xmm0,%xmm6 162 vpxor %xmm6,%xmm4,%xmm4 163 vpclmulqdq $0x01,(%rdi,%r11,1),%xmm0,%xmm6 164 vpxor %xmm6,%xmm5,%xmm5 165 vpclmulqdq $0x10,(%rdi,%r11,1),%xmm0,%xmm6 166 vpxor %xmm6,%xmm5,%xmm5 [all …]
|
/external/boringssl/linux-x86_64/crypto/cipher_extra/ |
D | aes128gcmsiv-x86_64.S | 144 vpclmulqdq $0x10,(%rdi,%r11,1),%xmm0,%xmm6 145 vpxor %xmm6,%xmm5,%xmm5 159 vpclmulqdq $0x00,(%rdi,%r11,1),%xmm0,%xmm6 160 vpxor %xmm6,%xmm3,%xmm3 161 vpclmulqdq $0x11,(%rdi,%r11,1),%xmm0,%xmm6 162 vpxor %xmm6,%xmm4,%xmm4 163 vpclmulqdq $0x01,(%rdi,%r11,1),%xmm0,%xmm6 164 vpxor %xmm6,%xmm5,%xmm5 165 vpclmulqdq $0x10,(%rdi,%r11,1),%xmm0,%xmm6 166 vpxor %xmm6,%xmm5,%xmm5 [all …]
|
/external/boringssl/win-x86_64/crypto/cipher_extra/ |
D | aes128gcmsiv-x86_64.asm | 178 vpclmulqdq xmm6,xmm0,XMMWORD[r11*1+rdi],0x10 179 vpxor xmm5,xmm5,xmm6 193 vpclmulqdq xmm6,xmm0,XMMWORD[r11*1+rdi],0x00 194 vpxor xmm3,xmm3,xmm6 195 vpclmulqdq xmm6,xmm0,XMMWORD[r11*1+rdi],0x11 196 vpxor xmm4,xmm4,xmm6 197 vpclmulqdq xmm6,xmm0,XMMWORD[r11*1+rdi],0x01 198 vpxor xmm5,xmm5,xmm6 199 vpclmulqdq xmm6,xmm0,XMMWORD[r11*1+rdi],0x10 200 vpxor xmm5,xmm5,xmm6 [all …]
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | rotate_win.cc | 57 movq xmm6, qword ptr [eax] in TransposeWx8_SSSE3() 60 punpcklbw xmm6, xmm7 in TransposeWx8_SSSE3() 62 movdqa xmm7, xmm6 in TransposeWx8_SSSE3() 71 punpcklwd xmm4, xmm6 in TransposeWx8_SSSE3() 73 movdqa xmm6, xmm4 in TransposeWx8_SSSE3() 75 palignr xmm6, xmm6, 8 in TransposeWx8_SSSE3() 85 punpckldq xmm2, xmm6 in TransposeWx8_SSSE3() 86 movdqa xmm6, xmm2 in TransposeWx8_SSSE3() 87 palignr xmm6, xmm6, 8 in TransposeWx8_SSSE3() 90 movq qword ptr [edx + esi], xmm6 in TransposeWx8_SSSE3() [all …]
|
/external/libyuv/files/source/ |
D | rotate_win.cc | 59 movq xmm6, qword ptr [eax] in TransposeWx8_SSSE3() 62 punpcklbw xmm6, xmm7 in TransposeWx8_SSSE3() 64 movdqa xmm7, xmm6 in TransposeWx8_SSSE3() 73 punpcklwd xmm4, xmm6 in TransposeWx8_SSSE3() 75 movdqa xmm6, xmm4 in TransposeWx8_SSSE3() 77 palignr xmm6, xmm6, 8 in TransposeWx8_SSSE3() 87 punpckldq xmm2, xmm6 in TransposeWx8_SSSE3() 88 movdqa xmm6, xmm2 in TransposeWx8_SSSE3() 89 palignr xmm6, xmm6, 8 in TransposeWx8_SSSE3() 92 movq qword ptr [edx + esi], xmm6 in TransposeWx8_SSSE3() [all …]
|
/external/boringssl/src/crypto/fipsmodule/aes/asm/ |
D | vpaes-x86_64.pl | 81 ## Preserves %xmm6 - %xmm8 so you get some local vectors 345 ## the high bits of %xmm6. 356 movdqa %xmm0, %xmm6 # save short part 358 movhlps %xmm4, %xmm6 # clobber low side with zeros 363 palignr \$8,%xmm6,%xmm0 381 ## %xmm6. The low side's rounds are the same as the 392 movdqa %xmm0, %xmm6 # save cur_lo in xmm6 403 movdqa %xmm6, %xmm7 446 pxor %xmm6, %xmm6 458 ## %xmm6: low side, d c 0 0 [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | highbd_variance_impl_sse2.asm | 68 pxor xmm6, xmm6 ; clear xmm6 for accumulating sse 93 paddd xmm6, xmm1 100 paddd xmm6, xmm3 107 paddd xmm6, xmm1 112 paddd xmm6, xmm3 131 movdqa xmm4, xmm6 132 punpckldq xmm6, xmm0 137 paddd xmm6, xmm4 143 movdqa xmm4, xmm6 149 paddd xmm6, xmm4 [all …]
|
/external/capstone/suite/MC/X86/ |
D | x86_64-xop-encoding.s.cs | 9 0x8f,0xe9,0x78,0xc7,0xd6 = vphaddwq %xmm6, %xmm2 12 0x8f,0xe9,0x78,0xd7,0x34,0x01 = vphadduwq (%rcx, %rax), %xmm6 17 0x8f,0xe9,0x78,0xdb,0xd6 = vphaddudq %xmm6, %xmm2 27 0x8f,0xe9,0x78,0xc1,0xf5 = vphaddbw %xmm5, %xmm6 37 0x8f,0xe9,0x78,0x80,0xee = vfrczps %xmm6, %xmm5 47 0x8f,0xe9,0x68,0x97,0xf4 = vpshlq %xmm2, %xmm4, %xmm6 49 0x8f,0xe9,0x50,0x97,0x34,0x0a = vpshlq %xmm5, (%rdx, %rcx), %xmm6 61 0x8f,0xe9,0x48,0x9b,0x2c,0x08 = vpshaq %xmm6, (%rax, %rcx), %xmm5 68 0x8f,0xe9,0xe0,0x91,0x30 = vprotw (%rax), %xmm3, %xmm6 83 0x8f,0xe8,0x78,0xc2,0x31,0x2b = vprotd $43, (%rcx), %xmm6 [all …]
|
/external/boringssl/mac-x86/crypto/chacha/ |
D | chacha-x86.S | 411 pshufd $170,%xmm7,%xmm6 420 movdqa %xmm6,-32(%ebp) 430 pshufd $170,%xmm7,%xmm6 438 movdqa %xmm6,-96(%ebp) 449 movdqa -32(%ebp),%xmm6 455 movdqa %xmm6,-32(%ebx) 461 movdqa 96(%ebp),%xmm6 468 movdqa %xmm6,96(%ebx) 472 movdqa %xmm4,%xmm6 482 pxor %xmm0,%xmm6 [all …]
|
/external/boringssl/linux-x86/crypto/chacha/ |
D | chacha-x86.S | 414 pshufd $170,%xmm7,%xmm6 423 movdqa %xmm6,-32(%ebp) 433 pshufd $170,%xmm7,%xmm6 441 movdqa %xmm6,-96(%ebp) 452 movdqa -32(%ebp),%xmm6 458 movdqa %xmm6,-32(%ebx) 464 movdqa 96(%ebp),%xmm6 471 movdqa %xmm6,96(%ebx) 475 movdqa %xmm4,%xmm6 485 pxor %xmm0,%xmm6 [all …]
|
/external/llvm/test/MC/X86/ |
D | x86_64-xop-encoding.s | 37 vphaddwq %xmm6, %xmm2 50 vphadduwq (%rcx,%rax), %xmm6 69 vphaddudq %xmm6, %xmm2 109 vphaddbw %xmm5, %xmm6 149 vfrczps %xmm6, %xmm5 191 vpshlq %xmm2, %xmm4, %xmm6 197 vpshlq %xmm5, (%rdx,%rcx), %xmm6 241 vpshaq %xmm6, (%rax,%rcx), %xmm5 268 vprotw (%rax), %xmm3, %xmm6 317 vprotd $43, (%rcx), %xmm6 [all …]
|
/external/libvpx/libvpx/vp8/common/x86/ |
D | loopfilter_sse2.asm | 54 movdqa xmm6, xmm1 ; q2 58 psubusb xmm2, xmm6 ; q3-=q2 60 psubusb xmm4, xmm6 ; q1-=q2 61 psubusb xmm6, xmm3 ; q2-=q1 63 por xmm4, xmm6 ; abs(q2-q1) 80 movdqa xmm6, [rsi+2*rax] ; p1 84 movlps xmm6, [rsi + rcx] ; p1 88 movhps xmm6, [rdi + rcx] 91 movdqa [rsp+_p1], xmm6 ; store p1 95 movdqa xmm3, xmm6 ; p1 [all …]
|
/external/boringssl/win-x86/crypto/chacha/ |
D | chacha-x86.asm | 423 pshufd xmm6,xmm7,170 432 movdqa [ebp-32],xmm6 442 pshufd xmm6,xmm7,170 450 movdqa [ebp-96],xmm6 461 movdqa xmm6,[ebp-32] 467 movdqa [ebx-32],xmm6 473 movdqa xmm6,[96+ebp] 480 movdqa [96+ebx],xmm6 484 movdqa xmm6,xmm4 494 pxor xmm6,xmm0 [all …]
|
/external/libjpeg-turbo/simd/ |
D | jccolext-sse2-64.asm | 299 movdqa xmm6,xmm1 301 punpckhwd xmm6,xmm3 303 movdqa xmm4,xmm6 305 pmaddwd xmm6,[rel PW_F0299_F0337] ; xmm6=ROH*FIX(0.299)+GOH*FIX(0.337) 310 movdqa XMMWORD [wk(5)], xmm6 ; wk(5)=ROH*FIX(0.299)+GOH*FIX(0.337) 313 pxor xmm6,xmm6 315 punpckhwd xmm6,xmm5 ; xmm6=BOH 317 psrld xmm6,1 ; xmm6=BOH*FIX(0.500) 322 paddd xmm4,xmm6 331 movdqa xmm6,xmm0 [all …]
|
D | jccolext-sse2.asm | 312 movdqa xmm6,xmm1 314 punpckhwd xmm6,xmm3 316 movdqa xmm4,xmm6 318 pmaddwd xmm6,[GOTOFF(eax,PW_F0299_F0337)] ; xmm6=ROH*FIX(0.299)+GOH*FIX(0.337) 323 movdqa XMMWORD [wk(5)], xmm6 ; wk(5)=ROH*FIX(0.299)+GOH*FIX(0.337) 326 pxor xmm6,xmm6 328 punpckhwd xmm6,xmm5 ; xmm6=BOH 330 psrld xmm6,1 ; xmm6=BOH*FIX(0.500) 335 paddd xmm4,xmm6 344 movdqa xmm6,xmm0 [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | vec_minmax_uint.ll | 84 ; SSE2-NEXT: movdqa %xmm1, %xmm6 85 ; SSE2-NEXT: pxor %xmm4, %xmm6 86 ; SSE2-NEXT: movdqa %xmm6, %xmm7 89 ; SSE2-NEXT: pcmpeqd %xmm5, %xmm6 90 ; SSE2-NEXT: pshufd {{.*#+}} xmm5 = xmm6[1,1,3,3] 92 ; SSE2-NEXT: pshufd {{.*#+}} xmm6 = xmm7[1,1,3,3] 93 ; SSE2-NEXT: por %xmm5, %xmm6 108 ; SSE2-NEXT: pand %xmm6, %xmm1 109 ; SSE2-NEXT: pandn %xmm3, %xmm6 110 ; SSE2-NEXT: por %xmm6, %xmm1 [all …]
|
/external/boringssl/mac-x86/crypto/fipsmodule/ |
D | aesni-x86.S | 254 pxor %xmm0,%xmm6 307 pxor %xmm0,%xmm6 376 movdqu 64(%esi),%xmm6 391 movups %xmm6,64(%edi) 392 movdqu 64(%esi),%xmm6 407 movups %xmm6,64(%edi) 423 movups 64(%esi),%xmm6 430 movups %xmm6,64(%edi) 478 movdqu 64(%esi),%xmm6 493 movups %xmm6,64(%edi) [all …]
|