/external/libvpx/libvpx/vp8/common/x86/ |
D | idctllm_sse2.asm | 124 ; Zero out xmm7, for use unpacking 125 pxor xmm7, xmm7 136 movdqa [rax], xmm7 137 movdqa [rax+16], xmm7 138 movdqa [rax+32], xmm7 139 movdqa [rax+48], xmm7 174 movdqa xmm7, xmm3 175 pmulhw xmm7, [GLOBAL(x_c1sqr2less1)] 177 paddw xmm7, xmm3 ; ip3 * cos(pi/8) * sqrt(2) 178 psubw xmm7, xmm5 ; c1 [all …]
|
D | loopfilter_sse2.asm | 52 movdqa xmm7, [rdx] ;limit 135 psubusb xmm1, xmm7 138 movdqa xmm7, [rdx] ; blimit 160 … psubusb xmm5, xmm7 ; abs (p0 - q0) *2 + abs(p1-q1)/2 > blimit 166 pxor xmm7, xmm7 172 pcmpeqb xmm1, xmm7 ; mask xmm1 180 movdqa xmm7, [rsp+_q1] ; q1 183 movdqa xmm7, [rdi] ; q1 188 movdqa xmm7, [rsp+_q1] ; q1 192 … pxor xmm7, xmm3 ; q1 offset to convert to signed values [all …]
|
/external/flac/libFLAC/ |
D | lpc_intrin_sse2.c | 422 __m128i xmm0, xmm1, xmm2, xmm3, xmm4, xmm5, xmm6, xmm7; in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() local 441 xmm7 = _mm_loadl_epi64((const __m128i*)(data+i-12)); // 0 0 d[i-11] d[i-12] in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 442 xmm7 = _mm_shuffle_epi32(xmm7, _MM_SHUFFLE(2,0,3,1)); // 0 d[i-12] 0 d[i-11] in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 443 …xmm7 = _mm_mul_epu32(xmm7, xmm5); /* we use _unsigned_ multiplication and discard high dword of th… in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 450 xmm7 = _mm_add_epi32(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 457 xmm7 = _mm_add_epi32(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 464 xmm7 = _mm_add_epi32(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 471 xmm7 = _mm_add_epi32(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 478 xmm7 = _mm_add_epi32(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() 480 xmm7 = _mm_add_epi32(xmm7, _mm_srli_si128(xmm7, 8)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() [all …]
|
D | lpc_intrin_sse41.c | 69 __m128i xmm0, xmm1, xmm2, xmm3, xmm4, xmm5, xmm6, xmm7; in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() local 88 xmm7 = _mm_loadl_epi64((const __m128i*)(data+i-12)); // 0 0 d[i-11] d[i-12] in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 89 xmm7 = _mm_shuffle_epi32(xmm7, _MM_SHUFFLE(2,0,3,1)); // 0 d[i-12] 0 d[i-11] in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 90 xmm7 = _mm_mul_epi32(xmm7, xmm5); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 97 xmm7 = _mm_add_epi64(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 104 xmm7 = _mm_add_epi64(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 111 xmm7 = _mm_add_epi64(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 118 xmm7 = _mm_add_epi64(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 125 xmm7 = _mm_add_epi64(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() 127 xmm7 = _mm_add_epi64(xmm7, _mm_srli_si128(xmm7, 8)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/ |
D | vector-interleave.ll | 28 ; SSE-NEXT: movdqa %xmm7, %xmm5 30 ; SSE-NEXT: punpckhwd {{.*#+}} xmm7 = xmm7[4],xmm6[4],xmm7[5],xmm6[5],xmm7[6],xmm6[6],xmm7[7],xm… 32 ; SSE-NEXT: punpckhdq {{.*#+}} xmm6 = xmm6[2],xmm7[2],xmm6[3],xmm7[3] 33 ; SSE-NEXT: punpckldq {{.*#+}} xmm4 = xmm4[0],xmm7[0],xmm4[1],xmm7[1] 34 ; SSE-NEXT: movdqa %xmm1, %xmm7 35 ; SSE-NEXT: punpckhdq {{.*#+}} xmm7 = xmm7[2],xmm5[2],xmm7[3],xmm5[3] 41 ; SSE-NEXT: punpcklwd {{.*#+}} xmm1 = xmm1[0],xmm7[0],xmm1[1],xmm7[1],xmm1[2],xmm7[2],xmm1[3],xm… 42 ; SSE-NEXT: punpckhwd {{.*#+}} xmm2 = xmm2[4],xmm7[4],xmm2[5],xmm7[5],xmm2[6],xmm7[6],xmm2[7],xm… 43 ; SSE-NEXT: movdqa %xmm0, %xmm7 44 ; SSE-NEXT: punpcklwd {{.*#+}} xmm7 = xmm7[0],xmm4[0],xmm7[1],xmm4[1],xmm7[2],xmm4[2],xmm7[3],xm… [all …]
|
/external/libaom/libaom/third_party/libyuv/source/ |
D | rotate_win.cc | 60 movq xmm7, qword ptr [eax + edi] in TransposeWx8_SSSE3() 61 punpcklbw xmm6, xmm7 in TransposeWx8_SSSE3() 63 movdqa xmm7, xmm6 in TransposeWx8_SSSE3() 64 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3() 73 punpcklwd xmm5, xmm7 in TransposeWx8_SSSE3() 75 movdqa xmm7, xmm5 in TransposeWx8_SSSE3() 77 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3() 96 punpckldq xmm3, xmm7 in TransposeWx8_SSSE3() 100 movdqa xmm7, xmm3 in TransposeWx8_SSSE3() 101 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3() [all …]
|
/external/libyuv/files/source/ |
D | rotate_win.cc | 61 movq xmm7, qword ptr [eax + edi] in TransposeWx8_SSSE3() 62 punpcklbw xmm6, xmm7 in TransposeWx8_SSSE3() 64 movdqa xmm7, xmm6 in TransposeWx8_SSSE3() 65 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3() 74 punpcklwd xmm5, xmm7 in TransposeWx8_SSSE3() 76 movdqa xmm7, xmm5 in TransposeWx8_SSSE3() 78 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3() 97 punpckldq xmm3, xmm7 in TransposeWx8_SSSE3() 101 movdqa xmm7, xmm3 in TransposeWx8_SSSE3() 102 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3() [all …]
|
/external/libvpx/libvpx/third_party/libyuv/source/ |
D | rotate_win.cc | 61 movq xmm7, qword ptr [eax + edi] in TransposeWx8_SSSE3() 62 punpcklbw xmm6, xmm7 in TransposeWx8_SSSE3() 64 movdqa xmm7, xmm6 in TransposeWx8_SSSE3() 65 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3() 74 punpcklwd xmm5, xmm7 in TransposeWx8_SSSE3() 76 movdqa xmm7, xmm5 in TransposeWx8_SSSE3() 78 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3() 97 punpckldq xmm3, xmm7 in TransposeWx8_SSSE3() 101 movdqa xmm7, xmm3 in TransposeWx8_SSSE3() 102 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3() [all …]
|
/external/boringssl/linux-x86_64/crypto/cipher_extra/ |
D | aes128gcmsiv-x86_64.S | 231 vpclmulqdq $0x10,poly(%rip),%xmm1,%xmm7 244 vpxor %xmm7,%xmm1,%xmm1 260 vpclmulqdq $0x10,poly(%rip),%xmm1,%xmm7 273 vpxor %xmm7,%xmm1,%xmm1 756 vmovdqa %xmm2,%xmm7 761 vpxor (%rcx),%xmm7,%xmm7 767 vaesenc %xmm12,%xmm7,%xmm7 774 vaesenc %xmm12,%xmm7,%xmm7 781 vaesenc %xmm12,%xmm7,%xmm7 788 vaesenc %xmm12,%xmm7,%xmm7 [all …]
|
/external/boringssl/mac-x86_64/crypto/cipher_extra/ |
D | aes128gcmsiv-x86_64.S | 231 vpclmulqdq $0x10,poly(%rip),%xmm1,%xmm7 244 vpxor %xmm7,%xmm1,%xmm1 260 vpclmulqdq $0x10,poly(%rip),%xmm1,%xmm7 273 vpxor %xmm7,%xmm1,%xmm1 754 vmovdqa %xmm2,%xmm7 759 vpxor (%rcx),%xmm7,%xmm7 765 vaesenc %xmm12,%xmm7,%xmm7 772 vaesenc %xmm12,%xmm7,%xmm7 779 vaesenc %xmm12,%xmm7,%xmm7 786 vaesenc %xmm12,%xmm7,%xmm7 [all …]
|
/external/boringssl/win-x86_64/crypto/cipher_extra/ |
D | aes128gcmsiv-x86_64.asm | 260 vpclmulqdq xmm7,xmm1,XMMWORD[poly],0x10 273 vpxor xmm1,xmm1,xmm7 289 vpclmulqdq xmm7,xmm1,XMMWORD[poly],0x10 302 vpxor xmm1,xmm1,xmm7 849 vmovdqa xmm7,xmm2 854 vpxor xmm7,xmm7,XMMWORD[rcx] 860 vaesenc xmm7,xmm7,xmm12 867 vaesenc xmm7,xmm7,xmm12 874 vaesenc xmm7,xmm7,xmm12 881 vaesenc xmm7,xmm7,xmm12 [all …]
|
/external/libaom/libaom/aom_dsp/x86/ |
D | aom_subpixel_8t_sse2.asm | 24 movdqa xmm7, [rdx] ;load filters 25 pshuflw xmm0, xmm7, 0b ;k0 26 pshuflw xmm1, xmm7, 01010101b ;k1 27 pshuflw xmm2, xmm7, 10101010b ;k2 28 pshuflw xmm3, xmm7, 11111111b ;k3 29 psrldq xmm7, 8 30 pshuflw xmm4, xmm7, 0b ;k4 31 pshuflw xmm5, xmm7, 01010101b ;k5 32 pshuflw xmm6, xmm7, 10101010b ;k6 33 pshuflw xmm7, xmm7, 11111111b ;k7 [all …]
|
D | aom_high_subpixel_8t_sse2.asm | 24 movdqa xmm7, [rdx] ;load filters 25 pshuflw xmm0, xmm7, 0b ;k0 26 pshuflw xmm1, xmm7, 01010101b ;k1 27 pshuflw xmm2, xmm7, 10101010b ;k2 28 pshuflw xmm3, xmm7, 11111111b ;k3 29 psrldq xmm7, 8 30 pshuflw xmm4, xmm7, 0b ;k4 31 pshuflw xmm5, xmm7, 01010101b ;k5 32 pshuflw xmm6, xmm7, 10101010b ;k6 33 pshuflw xmm7, xmm7, 11111111b ;k7 [all …]
|
/external/boringssl/linux-x86/crypto/chacha/ |
D | chacha-x86.S | 411 movdqu (%edx),%xmm7 417 pshufd $0,%xmm7,%xmm4 418 pshufd $85,%xmm7,%xmm5 420 pshufd $170,%xmm7,%xmm6 421 pshufd $255,%xmm7,%xmm7 430 movdqa %xmm7,-16(%ebp) 431 movdqa 32(%eax),%xmm7 437 pshufd $0,%xmm7,%xmm4 438 pshufd $85,%xmm7,%xmm5 439 pshufd $170,%xmm7,%xmm6 [all …]
|
/external/boringssl/mac-x86/crypto/chacha/ |
D | chacha-x86.S | 408 movdqu (%edx),%xmm7 414 pshufd $0,%xmm7,%xmm4 415 pshufd $85,%xmm7,%xmm5 417 pshufd $170,%xmm7,%xmm6 418 pshufd $255,%xmm7,%xmm7 427 movdqa %xmm7,-16(%ebp) 428 movdqa 32(%eax),%xmm7 434 pshufd $0,%xmm7,%xmm4 435 pshufd $85,%xmm7,%xmm5 436 pshufd $170,%xmm7,%xmm6 [all …]
|
/external/boringssl/src/crypto/fipsmodule/aes/asm/ |
D | vpaes-x86_64.pl | 189 ## Clobbers %xmm1-%xmm5, %xmm7-%xmm11, %r9, %r10, %r11, %rax 215 movdqa %xmm9, %xmm7 219 pandn %xmm6, %xmm7 223 psrld \$4, %xmm7 231 pshufb %xmm7, %xmm6 291 movdqa %xmm9, %xmm7 295 pandn %xmm6, %xmm7 297 psrld \$4, %xmm7 305 pxor %xmm7, %xmm6 307 pshufb %xmm7, %xmm11 [all …]
|
/external/fec/ |
D | peak_sse2_assist.s | 21 pxor %xmm7,%xmm7 # clear peak 30 pmaxsw %xmm0,%xmm7 # store peak 35 2: movaps %xmm7,%xmm0 37 pmaxsw %xmm0,%xmm7 38 movaps %xmm7,%xmm0 40 pmaxsw %xmm0,%xmm7 41 movaps %xmm7,%xmm0 43 pmaxsw %xmm0,%xmm7 # min value in low word of %xmm7 45 movd %xmm7,%eax
|
D | peakval_sse2_assist.s | 21 pxor %xmm7,%xmm7 # clear peak 30 pmaxsw %xmm0,%xmm7 # store peak 35 2: movaps %xmm7,%xmm0 37 pmaxsw %xmm0,%xmm7 38 movaps %xmm7,%xmm0 40 pmaxsw %xmm0,%xmm7 41 movaps %xmm7,%xmm0 43 pmaxsw %xmm0,%xmm7 # min value in low word of %xmm7 45 movd %xmm7,%eax
|
/external/boringssl/win-x86/crypto/chacha/ |
D | chacha-x86.asm | 420 movdqu xmm7,[edx] 426 pshufd xmm4,xmm7,0 427 pshufd xmm5,xmm7,85 429 pshufd xmm6,xmm7,170 430 pshufd xmm7,xmm7,255 439 movdqa [ebp-16],xmm7 440 movdqa xmm7,[32+eax] 446 pshufd xmm4,xmm7,0 447 pshufd xmm5,xmm7,85 448 pshufd xmm6,xmm7,170 [all …]
|
/external/capstone/suite/MC/X86/ |
D | x86_64-xop-encoding.s.cs | 10 0x8f,0xe9,0x78,0xc6,0x3c,0x02 = vphaddwd (%rdx, %rax), %xmm7 13 0x8f,0xe9,0x78,0xd7,0xc7 = vphadduwq %xmm7, %xmm0 23 0x8f,0xe9,0x78,0xd2,0xfd = vphaddubd %xmm5, %xmm7 33 0x8f,0xe9,0x78,0x82,0xfd = vfrczss %xmm5, %xmm7 35 0x8f,0xe9,0x78,0x83,0xc7 = vfrczsd %xmm7, %xmm0 41 0x8f,0xe9,0x78,0x81,0xc7 = vfrczpd %xmm7, %xmm0 50 0x8f,0xe9,0x40,0x96,0xdd = vpshld %xmm7, %xmm5, %xmm3 54 0x8f,0xe9,0xf8,0x94,0x39 = vpshlb (%rcx), %xmm0, %xmm7 56 0x8f,0xe9,0x40,0x99,0xdd = vpshaw %xmm7, %xmm5, %xmm3 84 0x8f,0xe8,0x78,0xc2,0x3c,0x08,0x2c = vprotd $44, (%rax, %rcx), %xmm7 [all …]
|
/external/boringssl/linux-x86_64/crypto/fipsmodule/ |
D | aesni-gcm-x86_64.S | 44 vpclmulqdq $0x10,%xmm3,%xmm7,%xmm5 47 vpclmulqdq $0x01,%xmm3,%xmm7,%xmm6 71 vpclmulqdq $0x00,%xmm3,%xmm7,%xmm1 75 vpclmulqdq $0x11,%xmm3,%xmm7,%xmm7 114 vpxor %xmm3,%xmm7,%xmm7 133 vpxor %xmm5,%xmm7,%xmm7 155 vpxor %xmm1,%xmm7,%xmm7 177 vpxor %xmm2,%xmm7,%xmm7 196 vpxor %xmm8,%xmm7,%xmm7 216 vpxor %xmm6,%xmm7,%xmm7 [all …]
|
/external/boringssl/mac-x86_64/crypto/fipsmodule/ |
D | aesni-gcm-x86_64.S | 44 vpclmulqdq $0x10,%xmm3,%xmm7,%xmm5 47 vpclmulqdq $0x01,%xmm3,%xmm7,%xmm6 71 vpclmulqdq $0x00,%xmm3,%xmm7,%xmm1 75 vpclmulqdq $0x11,%xmm3,%xmm7,%xmm7 114 vpxor %xmm3,%xmm7,%xmm7 133 vpxor %xmm5,%xmm7,%xmm7 155 vpxor %xmm1,%xmm7,%xmm7 177 vpxor %xmm2,%xmm7,%xmm7 196 vpxor %xmm8,%xmm7,%xmm7 216 vpxor %xmm6,%xmm7,%xmm7 [all …]
|
/external/libvpx/libvpx/vpx_dsp/x86/ |
D | vpx_subpixel_8t_sse2.asm | 21 movdqa xmm7, [rdx] ;load filters 22 pshuflw xmm0, xmm7, 0b ;k0 23 pshuflw xmm1, xmm7, 01010101b ;k1 24 pshuflw xmm2, xmm7, 10101010b ;k2 25 pshuflw xmm3, xmm7, 11111111b ;k3 26 psrldq xmm7, 8 27 pshuflw xmm4, xmm7, 0b ;k4 28 pshuflw xmm5, xmm7, 01010101b ;k5 29 pshuflw xmm6, xmm7, 10101010b ;k6 30 pshuflw xmm7, xmm7, 11111111b ;k7 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/X86/ |
D | x86_64-xop-encoding.s | 42 vphaddwd (%rdx,%rax), %xmm7 53 vphadduwq %xmm7, %xmm0 93 vphaddubd %xmm5, %xmm7 133 vfrczss %xmm5, %xmm7 141 vfrczsd %xmm7, %xmm0 163 vfrczpd %xmm7, %xmm0 202 vpshld %xmm7, %xmm5, %xmm3 216 vpshlb (%rcx), %xmm0, %xmm7 224 vpshaw %xmm7, %xmm5, %xmm3 320 vprotd $44, (%rax,%rcx), %xmm7 [all …]
|
/external/llvm/test/MC/X86/ |
D | x86_64-xop-encoding.s | 42 vphaddwd (%rdx,%rax), %xmm7 53 vphadduwq %xmm7, %xmm0 93 vphaddubd %xmm5, %xmm7 133 vfrczss %xmm5, %xmm7 141 vfrczsd %xmm7, %xmm0 163 vfrczpd %xmm7, %xmm0 202 vpshld %xmm7, %xmm5, %xmm3 216 vpshlb (%rcx), %xmm0, %xmm7 224 vpshaw %xmm7, %xmm5, %xmm3 320 vprotd $44, (%rax,%rcx), %xmm7 [all …]
|