Home
last modified time | relevance | path

Searched refs:xmm7 (Results 1 – 25 of 422) sorted by relevance

12345678910>>...17

/external/libvpx/libvpx/vp8/common/x86/
Didctllm_sse2.asm124 ; Zero out xmm7, for use unpacking
125 pxor xmm7, xmm7
136 movdqa [rax], xmm7
137 movdqa [rax+16], xmm7
138 movdqa [rax+32], xmm7
139 movdqa [rax+48], xmm7
174 movdqa xmm7, xmm3
175 pmulhw xmm7, [GLOBAL(x_c1sqr2less1)]
177 paddw xmm7, xmm3 ; ip3 * cos(pi/8) * sqrt(2)
178 psubw xmm7, xmm5 ; c1
[all …]
Dloopfilter_sse2.asm52 movdqa xmm7, [rdx] ;limit
135 psubusb xmm1, xmm7
138 movdqa xmm7, [rdx] ; blimit
160 … psubusb xmm5, xmm7 ; abs (p0 - q0) *2 + abs(p1-q1)/2 > blimit
166 pxor xmm7, xmm7
172 pcmpeqb xmm1, xmm7 ; mask xmm1
180 movdqa xmm7, [rsp+_q1] ; q1
183 movdqa xmm7, [rdi] ; q1
188 movdqa xmm7, [rsp+_q1] ; q1
192 … pxor xmm7, xmm3 ; q1 offset to convert to signed values
[all …]
/external/flac/libFLAC/
Dlpc_intrin_sse2.c422 __m128i xmm0, xmm1, xmm2, xmm3, xmm4, xmm5, xmm6, xmm7; in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2() local
441 xmm7 = _mm_loadl_epi64((const __m128i*)(data+i-12)); // 0 0 d[i-11] d[i-12] in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
442 xmm7 = _mm_shuffle_epi32(xmm7, _MM_SHUFFLE(2,0,3,1)); // 0 d[i-12] 0 d[i-11] in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
443xmm7 = _mm_mul_epu32(xmm7, xmm5); /* we use _unsigned_ multiplication and discard high dword of th… in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
450 xmm7 = _mm_add_epi32(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
457 xmm7 = _mm_add_epi32(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
464 xmm7 = _mm_add_epi32(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
471 xmm7 = _mm_add_epi32(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
478 xmm7 = _mm_add_epi32(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
480 xmm7 = _mm_add_epi32(xmm7, _mm_srli_si128(xmm7, 8)); in FLAC__lpc_compute_residual_from_qlp_coefficients_intrin_sse2()
[all …]
Dlpc_intrin_sse41.c69 __m128i xmm0, xmm1, xmm2, xmm3, xmm4, xmm5, xmm6, xmm7; in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41() local
88 xmm7 = _mm_loadl_epi64((const __m128i*)(data+i-12)); // 0 0 d[i-11] d[i-12] in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
89 xmm7 = _mm_shuffle_epi32(xmm7, _MM_SHUFFLE(2,0,3,1)); // 0 d[i-12] 0 d[i-11] in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
90 xmm7 = _mm_mul_epi32(xmm7, xmm5); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
97 xmm7 = _mm_add_epi64(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
104 xmm7 = _mm_add_epi64(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
111 xmm7 = _mm_add_epi64(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
118 xmm7 = _mm_add_epi64(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
125 xmm7 = _mm_add_epi64(xmm7, xmm6); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
127 xmm7 = _mm_add_epi64(xmm7, _mm_srli_si128(xmm7, 8)); in FLAC__lpc_compute_residual_from_qlp_coefficients_wide_intrin_sse41()
[all …]
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/
Dvector-interleave.ll28 ; SSE-NEXT: movdqa %xmm7, %xmm5
30 ; SSE-NEXT: punpckhwd {{.*#+}} xmm7 = xmm7[4],xmm6[4],xmm7[5],xmm6[5],xmm7[6],xmm6[6],xmm7[7],xm…
32 ; SSE-NEXT: punpckhdq {{.*#+}} xmm6 = xmm6[2],xmm7[2],xmm6[3],xmm7[3]
33 ; SSE-NEXT: punpckldq {{.*#+}} xmm4 = xmm4[0],xmm7[0],xmm4[1],xmm7[1]
34 ; SSE-NEXT: movdqa %xmm1, %xmm7
35 ; SSE-NEXT: punpckhdq {{.*#+}} xmm7 = xmm7[2],xmm5[2],xmm7[3],xmm5[3]
41 ; SSE-NEXT: punpcklwd {{.*#+}} xmm1 = xmm1[0],xmm7[0],xmm1[1],xmm7[1],xmm1[2],xmm7[2],xmm1[3],xm…
42 ; SSE-NEXT: punpckhwd {{.*#+}} xmm2 = xmm2[4],xmm7[4],xmm2[5],xmm7[5],xmm2[6],xmm7[6],xmm2[7],xm…
43 ; SSE-NEXT: movdqa %xmm0, %xmm7
44 ; SSE-NEXT: punpcklwd {{.*#+}} xmm7 = xmm7[0],xmm4[0],xmm7[1],xmm4[1],xmm7[2],xmm4[2],xmm7[3],xm…
[all …]
/external/libaom/libaom/third_party/libyuv/source/
Drotate_win.cc60 movq xmm7, qword ptr [eax + edi] in TransposeWx8_SSSE3()
61 punpcklbw xmm6, xmm7 in TransposeWx8_SSSE3()
63 movdqa xmm7, xmm6 in TransposeWx8_SSSE3()
64 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3()
73 punpcklwd xmm5, xmm7 in TransposeWx8_SSSE3()
75 movdqa xmm7, xmm5 in TransposeWx8_SSSE3()
77 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3()
96 punpckldq xmm3, xmm7 in TransposeWx8_SSSE3()
100 movdqa xmm7, xmm3 in TransposeWx8_SSSE3()
101 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3()
[all …]
/external/libyuv/files/source/
Drotate_win.cc61 movq xmm7, qword ptr [eax + edi] in TransposeWx8_SSSE3()
62 punpcklbw xmm6, xmm7 in TransposeWx8_SSSE3()
64 movdqa xmm7, xmm6 in TransposeWx8_SSSE3()
65 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3()
74 punpcklwd xmm5, xmm7 in TransposeWx8_SSSE3()
76 movdqa xmm7, xmm5 in TransposeWx8_SSSE3()
78 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3()
97 punpckldq xmm3, xmm7 in TransposeWx8_SSSE3()
101 movdqa xmm7, xmm3 in TransposeWx8_SSSE3()
102 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3()
[all …]
/external/libvpx/libvpx/third_party/libyuv/source/
Drotate_win.cc61 movq xmm7, qword ptr [eax + edi] in TransposeWx8_SSSE3()
62 punpcklbw xmm6, xmm7 in TransposeWx8_SSSE3()
64 movdqa xmm7, xmm6 in TransposeWx8_SSSE3()
65 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3()
74 punpcklwd xmm5, xmm7 in TransposeWx8_SSSE3()
76 movdqa xmm7, xmm5 in TransposeWx8_SSSE3()
78 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3()
97 punpckldq xmm3, xmm7 in TransposeWx8_SSSE3()
101 movdqa xmm7, xmm3 in TransposeWx8_SSSE3()
102 palignr xmm7, xmm7, 8 in TransposeWx8_SSSE3()
[all …]
/external/boringssl/linux-x86_64/crypto/cipher_extra/
Daes128gcmsiv-x86_64.S231 vpclmulqdq $0x10,poly(%rip),%xmm1,%xmm7
244 vpxor %xmm7,%xmm1,%xmm1
260 vpclmulqdq $0x10,poly(%rip),%xmm1,%xmm7
273 vpxor %xmm7,%xmm1,%xmm1
756 vmovdqa %xmm2,%xmm7
761 vpxor (%rcx),%xmm7,%xmm7
767 vaesenc %xmm12,%xmm7,%xmm7
774 vaesenc %xmm12,%xmm7,%xmm7
781 vaesenc %xmm12,%xmm7,%xmm7
788 vaesenc %xmm12,%xmm7,%xmm7
[all …]
/external/boringssl/mac-x86_64/crypto/cipher_extra/
Daes128gcmsiv-x86_64.S231 vpclmulqdq $0x10,poly(%rip),%xmm1,%xmm7
244 vpxor %xmm7,%xmm1,%xmm1
260 vpclmulqdq $0x10,poly(%rip),%xmm1,%xmm7
273 vpxor %xmm7,%xmm1,%xmm1
754 vmovdqa %xmm2,%xmm7
759 vpxor (%rcx),%xmm7,%xmm7
765 vaesenc %xmm12,%xmm7,%xmm7
772 vaesenc %xmm12,%xmm7,%xmm7
779 vaesenc %xmm12,%xmm7,%xmm7
786 vaesenc %xmm12,%xmm7,%xmm7
[all …]
/external/boringssl/win-x86_64/crypto/cipher_extra/
Daes128gcmsiv-x86_64.asm260 vpclmulqdq xmm7,xmm1,XMMWORD[poly],0x10
273 vpxor xmm1,xmm1,xmm7
289 vpclmulqdq xmm7,xmm1,XMMWORD[poly],0x10
302 vpxor xmm1,xmm1,xmm7
849 vmovdqa xmm7,xmm2
854 vpxor xmm7,xmm7,XMMWORD[rcx]
860 vaesenc xmm7,xmm7,xmm12
867 vaesenc xmm7,xmm7,xmm12
874 vaesenc xmm7,xmm7,xmm12
881 vaesenc xmm7,xmm7,xmm12
[all …]
/external/libaom/libaom/aom_dsp/x86/
Daom_subpixel_8t_sse2.asm24 movdqa xmm7, [rdx] ;load filters
25 pshuflw xmm0, xmm7, 0b ;k0
26 pshuflw xmm1, xmm7, 01010101b ;k1
27 pshuflw xmm2, xmm7, 10101010b ;k2
28 pshuflw xmm3, xmm7, 11111111b ;k3
29 psrldq xmm7, 8
30 pshuflw xmm4, xmm7, 0b ;k4
31 pshuflw xmm5, xmm7, 01010101b ;k5
32 pshuflw xmm6, xmm7, 10101010b ;k6
33 pshuflw xmm7, xmm7, 11111111b ;k7
[all …]
Daom_high_subpixel_8t_sse2.asm24 movdqa xmm7, [rdx] ;load filters
25 pshuflw xmm0, xmm7, 0b ;k0
26 pshuflw xmm1, xmm7, 01010101b ;k1
27 pshuflw xmm2, xmm7, 10101010b ;k2
28 pshuflw xmm3, xmm7, 11111111b ;k3
29 psrldq xmm7, 8
30 pshuflw xmm4, xmm7, 0b ;k4
31 pshuflw xmm5, xmm7, 01010101b ;k5
32 pshuflw xmm6, xmm7, 10101010b ;k6
33 pshuflw xmm7, xmm7, 11111111b ;k7
[all …]
/external/boringssl/linux-x86/crypto/chacha/
Dchacha-x86.S411 movdqu (%edx),%xmm7
417 pshufd $0,%xmm7,%xmm4
418 pshufd $85,%xmm7,%xmm5
420 pshufd $170,%xmm7,%xmm6
421 pshufd $255,%xmm7,%xmm7
430 movdqa %xmm7,-16(%ebp)
431 movdqa 32(%eax),%xmm7
437 pshufd $0,%xmm7,%xmm4
438 pshufd $85,%xmm7,%xmm5
439 pshufd $170,%xmm7,%xmm6
[all …]
/external/boringssl/mac-x86/crypto/chacha/
Dchacha-x86.S408 movdqu (%edx),%xmm7
414 pshufd $0,%xmm7,%xmm4
415 pshufd $85,%xmm7,%xmm5
417 pshufd $170,%xmm7,%xmm6
418 pshufd $255,%xmm7,%xmm7
427 movdqa %xmm7,-16(%ebp)
428 movdqa 32(%eax),%xmm7
434 pshufd $0,%xmm7,%xmm4
435 pshufd $85,%xmm7,%xmm5
436 pshufd $170,%xmm7,%xmm6
[all …]
/external/boringssl/src/crypto/fipsmodule/aes/asm/
Dvpaes-x86_64.pl189 ## Clobbers %xmm1-%xmm5, %xmm7-%xmm11, %r9, %r10, %r11, %rax
215 movdqa %xmm9, %xmm7
219 pandn %xmm6, %xmm7
223 psrld \$4, %xmm7
231 pshufb %xmm7, %xmm6
291 movdqa %xmm9, %xmm7
295 pandn %xmm6, %xmm7
297 psrld \$4, %xmm7
305 pxor %xmm7, %xmm6
307 pshufb %xmm7, %xmm11
[all …]
/external/fec/
Dpeak_sse2_assist.s21 pxor %xmm7,%xmm7 # clear peak
30 pmaxsw %xmm0,%xmm7 # store peak
35 2: movaps %xmm7,%xmm0
37 pmaxsw %xmm0,%xmm7
38 movaps %xmm7,%xmm0
40 pmaxsw %xmm0,%xmm7
41 movaps %xmm7,%xmm0
43 pmaxsw %xmm0,%xmm7 # min value in low word of %xmm7
45 movd %xmm7,%eax
Dpeakval_sse2_assist.s21 pxor %xmm7,%xmm7 # clear peak
30 pmaxsw %xmm0,%xmm7 # store peak
35 2: movaps %xmm7,%xmm0
37 pmaxsw %xmm0,%xmm7
38 movaps %xmm7,%xmm0
40 pmaxsw %xmm0,%xmm7
41 movaps %xmm7,%xmm0
43 pmaxsw %xmm0,%xmm7 # min value in low word of %xmm7
45 movd %xmm7,%eax
/external/boringssl/win-x86/crypto/chacha/
Dchacha-x86.asm420 movdqu xmm7,[edx]
426 pshufd xmm4,xmm7,0
427 pshufd xmm5,xmm7,85
429 pshufd xmm6,xmm7,170
430 pshufd xmm7,xmm7,255
439 movdqa [ebp-16],xmm7
440 movdqa xmm7,[32+eax]
446 pshufd xmm4,xmm7,0
447 pshufd xmm5,xmm7,85
448 pshufd xmm6,xmm7,170
[all …]
/external/capstone/suite/MC/X86/
Dx86_64-xop-encoding.s.cs10 0x8f,0xe9,0x78,0xc6,0x3c,0x02 = vphaddwd (%rdx, %rax), %xmm7
13 0x8f,0xe9,0x78,0xd7,0xc7 = vphadduwq %xmm7, %xmm0
23 0x8f,0xe9,0x78,0xd2,0xfd = vphaddubd %xmm5, %xmm7
33 0x8f,0xe9,0x78,0x82,0xfd = vfrczss %xmm5, %xmm7
35 0x8f,0xe9,0x78,0x83,0xc7 = vfrczsd %xmm7, %xmm0
41 0x8f,0xe9,0x78,0x81,0xc7 = vfrczpd %xmm7, %xmm0
50 0x8f,0xe9,0x40,0x96,0xdd = vpshld %xmm7, %xmm5, %xmm3
54 0x8f,0xe9,0xf8,0x94,0x39 = vpshlb (%rcx), %xmm0, %xmm7
56 0x8f,0xe9,0x40,0x99,0xdd = vpshaw %xmm7, %xmm5, %xmm3
84 0x8f,0xe8,0x78,0xc2,0x3c,0x08,0x2c = vprotd $44, (%rax, %rcx), %xmm7
[all …]
/external/boringssl/linux-x86_64/crypto/fipsmodule/
Daesni-gcm-x86_64.S44 vpclmulqdq $0x10,%xmm3,%xmm7,%xmm5
47 vpclmulqdq $0x01,%xmm3,%xmm7,%xmm6
71 vpclmulqdq $0x00,%xmm3,%xmm7,%xmm1
75 vpclmulqdq $0x11,%xmm3,%xmm7,%xmm7
114 vpxor %xmm3,%xmm7,%xmm7
133 vpxor %xmm5,%xmm7,%xmm7
155 vpxor %xmm1,%xmm7,%xmm7
177 vpxor %xmm2,%xmm7,%xmm7
196 vpxor %xmm8,%xmm7,%xmm7
216 vpxor %xmm6,%xmm7,%xmm7
[all …]
/external/boringssl/mac-x86_64/crypto/fipsmodule/
Daesni-gcm-x86_64.S44 vpclmulqdq $0x10,%xmm3,%xmm7,%xmm5
47 vpclmulqdq $0x01,%xmm3,%xmm7,%xmm6
71 vpclmulqdq $0x00,%xmm3,%xmm7,%xmm1
75 vpclmulqdq $0x11,%xmm3,%xmm7,%xmm7
114 vpxor %xmm3,%xmm7,%xmm7
133 vpxor %xmm5,%xmm7,%xmm7
155 vpxor %xmm1,%xmm7,%xmm7
177 vpxor %xmm2,%xmm7,%xmm7
196 vpxor %xmm8,%xmm7,%xmm7
216 vpxor %xmm6,%xmm7,%xmm7
[all …]
/external/libvpx/libvpx/vpx_dsp/x86/
Dvpx_subpixel_8t_sse2.asm21 movdqa xmm7, [rdx] ;load filters
22 pshuflw xmm0, xmm7, 0b ;k0
23 pshuflw xmm1, xmm7, 01010101b ;k1
24 pshuflw xmm2, xmm7, 10101010b ;k2
25 pshuflw xmm3, xmm7, 11111111b ;k3
26 psrldq xmm7, 8
27 pshuflw xmm4, xmm7, 0b ;k4
28 pshuflw xmm5, xmm7, 01010101b ;k5
29 pshuflw xmm6, xmm7, 10101010b ;k6
30 pshuflw xmm7, xmm7, 11111111b ;k7
[all …]
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/X86/
Dx86_64-xop-encoding.s42 vphaddwd (%rdx,%rax), %xmm7
53 vphadduwq %xmm7, %xmm0
93 vphaddubd %xmm5, %xmm7
133 vfrczss %xmm5, %xmm7
141 vfrczsd %xmm7, %xmm0
163 vfrczpd %xmm7, %xmm0
202 vpshld %xmm7, %xmm5, %xmm3
216 vpshlb (%rcx), %xmm0, %xmm7
224 vpshaw %xmm7, %xmm5, %xmm3
320 vprotd $44, (%rax,%rcx), %xmm7
[all …]
/external/llvm/test/MC/X86/
Dx86_64-xop-encoding.s42 vphaddwd (%rdx,%rax), %xmm7
53 vphadduwq %xmm7, %xmm0
93 vphaddubd %xmm5, %xmm7
133 vfrczss %xmm5, %xmm7
141 vfrczsd %xmm7, %xmm0
163 vfrczpd %xmm7, %xmm0
202 vpshld %xmm7, %xmm5, %xmm3
216 vpshlb (%rcx), %xmm0, %xmm7
224 vpshaw %xmm7, %xmm5, %xmm3
320 vprotd $44, (%rax,%rcx), %xmm7
[all …]

12345678910>>...17