Home
last modified time | relevance | path

Searched refs:xmm8 (Results 1 – 25 of 241) sorted by relevance

12345678910

/external/llvm/test/CodeGen/X86/
Dsad.ll169 ; SSE2-NEXT: pshufd {{.*#+}} xmm8 = xmm1[2,3,0,1]
170 …w {{.*#+}} xmm8 = xmm8[0],xmm12[0],xmm8[1],xmm12[1],xmm8[2],xmm12[2],xmm8[3],xmm12[3],xmm8[4],xmm1…
204 ; SSE2-NEXT: movdqa %xmm8, %xmm3
205 ; SSE2-NEXT: punpckhwd {{.*#+}} xmm8 = xmm8[4],xmm12[4],xmm8[5],xmm12[5],xmm8[6],xmm12[6],xmm8[7…
212 ; SSE2-NEXT: psubd %xmm11, %xmm8
220 ; SSE2-NEXT: movdqa %xmm8, %xmm4
222 ; SSE2-NEXT: paddd %xmm4, %xmm8
223 ; SSE2-NEXT: pxor %xmm4, %xmm8
261 ; SSE2-NEXT: paddd %xmm8, %xmm1
402 ; SSE2-NEXT: pxor %xmm8, %xmm8
[all …]
Dstack-folding-fp-sse42.ll14 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
22 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
30 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
38 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
47 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
55 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
64 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
73 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
82 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
96 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
[all …]
Dstack-folding-fp-avx1.ll14 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
22 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
30 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
38 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
46 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
54 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
63 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
71 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
80 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
89 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
[all …]
Dstack-folding-int-avx2.ll14 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
25 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
36 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
50 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
57 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
67 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
76 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
85 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
94 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
103 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
[all …]
/external/boringssl/linux-x86_64/crypto/cipher_extra/
Daes128gcmsiv-x86_64.S757 vmovdqa %xmm3,%xmm8
762 vpxor (%rcx),%xmm8,%xmm8
768 vaesenc %xmm12,%xmm8,%xmm8
775 vaesenc %xmm12,%xmm8,%xmm8
782 vaesenc %xmm12,%xmm8,%xmm8
789 vaesenc %xmm12,%xmm8,%xmm8
797 vaesenc %xmm12,%xmm8,%xmm8
803 vaesenc %xmm12,%xmm8,%xmm8
809 vaesenc %xmm12,%xmm8,%xmm8
815 vaesenc %xmm12,%xmm8,%xmm8
[all …]
Dchacha20_poly1305_x86_64.S274 movdqu 16(%r9),%xmm8
279 movdqa %xmm8,64(%rbp)
286 paddd %xmm12,%xmm8
287 pxor %xmm8,%xmm4
295 paddd %xmm12,%xmm8
296 pxor %xmm8,%xmm4
307 paddd %xmm12,%xmm8
308 pxor %xmm8,%xmm4
316 paddd %xmm12,%xmm8
317 pxor %xmm8,%xmm4
[all …]
/external/boringssl/mac-x86_64/crypto/cipher_extra/
Daes128gcmsiv-x86_64.S755 vmovdqa %xmm3,%xmm8
760 vpxor (%rcx),%xmm8,%xmm8
766 vaesenc %xmm12,%xmm8,%xmm8
773 vaesenc %xmm12,%xmm8,%xmm8
780 vaesenc %xmm12,%xmm8,%xmm8
787 vaesenc %xmm12,%xmm8,%xmm8
795 vaesenc %xmm12,%xmm8,%xmm8
801 vaesenc %xmm12,%xmm8,%xmm8
807 vaesenc %xmm12,%xmm8,%xmm8
813 vaesenc %xmm12,%xmm8,%xmm8
[all …]
Dchacha20_poly1305_x86_64.S273 movdqu 16(%r9),%xmm8
278 movdqa %xmm8,64(%rbp)
285 paddd %xmm12,%xmm8
286 pxor %xmm8,%xmm4
294 paddd %xmm12,%xmm8
295 pxor %xmm8,%xmm4
306 paddd %xmm12,%xmm8
307 pxor %xmm8,%xmm4
315 paddd %xmm12,%xmm8
316 pxor %xmm8,%xmm4
[all …]
/external/boringssl/win-x86_64/crypto/cipher_extra/
Daes128gcmsiv-x86_64.asm850 vmovdqa xmm8,xmm3
855 vpxor xmm8,xmm8,XMMWORD[rcx]
861 vaesenc xmm8,xmm8,xmm12
868 vaesenc xmm8,xmm8,xmm12
875 vaesenc xmm8,xmm8,xmm12
882 vaesenc xmm8,xmm8,xmm12
890 vaesenc xmm8,xmm8,xmm12
896 vaesenc xmm8,xmm8,xmm12
902 vaesenc xmm8,xmm8,xmm12
908 vaesenc xmm8,xmm8,xmm12
[all …]
/external/boringssl/linux-x86_64/crypto/fipsmodule/
Daesni-gcm-x86_64.S82 vpxor %xmm4,%xmm8,%xmm8
93 vpxor 16+8(%rsp),%xmm8,%xmm8
157 vpxor 112+8(%rsp),%xmm8,%xmm8
171 vpclmulqdq $0x10,%xmm3,%xmm8,%xmm5
174 vpclmulqdq $0x01,%xmm3,%xmm8,%xmm1
178 vpclmulqdq $0x00,%xmm3,%xmm8,%xmm2
181 vpclmulqdq $0x11,%xmm3,%xmm8,%xmm8
196 vpxor %xmm8,%xmm7,%xmm7
287 vpalignr $8,%xmm4,%xmm4,%xmm8
340 vpxor 16+8(%rsp),%xmm8,%xmm8
[all …]
Dghash-x86_64.S991 movdqu 0(%rdx),%xmm8
996 pxor %xmm8,%xmm0
1000 pshufd $78,%xmm0,%xmm8
1001 pxor %xmm0,%xmm8
1028 xorps %xmm4,%xmm8
1032 pxor %xmm0,%xmm8
1034 pxor %xmm1,%xmm8
1036 movdqa %xmm8,%xmm9
1038 pslldq $8,%xmm8
1040 pxor %xmm8,%xmm0
[all …]
Dsha1-x86_64.S1332 movdqa %xmm3,%xmm8
1340 psrldq $4,%xmm8
1346 pxor %xmm2,%xmm8
1350 pxor %xmm8,%xmm4
1360 movdqa %xmm4,%xmm8
1366 psrld $31,%xmm8
1376 por %xmm8,%xmm4
1417 movdqa %xmm5,%xmm8
1423 pslldq $12,%xmm8
1431 movdqa %xmm8,%xmm10
[all …]
/external/boringssl/mac-x86_64/crypto/fipsmodule/
Daesni-gcm-x86_64.S82 vpxor %xmm4,%xmm8,%xmm8
93 vpxor 16+8(%rsp),%xmm8,%xmm8
157 vpxor 112+8(%rsp),%xmm8,%xmm8
171 vpclmulqdq $0x10,%xmm3,%xmm8,%xmm5
174 vpclmulqdq $0x01,%xmm3,%xmm8,%xmm1
178 vpclmulqdq $0x00,%xmm3,%xmm8,%xmm2
181 vpclmulqdq $0x11,%xmm3,%xmm8,%xmm8
196 vpxor %xmm8,%xmm7,%xmm7
287 vpalignr $8,%xmm4,%xmm4,%xmm8
340 vpxor 16+8(%rsp),%xmm8,%xmm8
[all …]
Dghash-x86_64.S978 movdqu 0(%rdx),%xmm8
983 pxor %xmm8,%xmm0
987 pshufd $78,%xmm0,%xmm8
988 pxor %xmm0,%xmm8
1015 xorps %xmm4,%xmm8
1019 pxor %xmm0,%xmm8
1021 pxor %xmm1,%xmm8
1023 movdqa %xmm8,%xmm9
1025 pslldq $8,%xmm8
1027 pxor %xmm8,%xmm0
[all …]
Dsha1-x86_64.S1331 movdqa %xmm3,%xmm8
1339 psrldq $4,%xmm8
1345 pxor %xmm2,%xmm8
1349 pxor %xmm8,%xmm4
1359 movdqa %xmm4,%xmm8
1365 psrld $31,%xmm8
1375 por %xmm8,%xmm4
1416 movdqa %xmm5,%xmm8
1422 pslldq $12,%xmm8
1430 movdqa %xmm8,%xmm10
[all …]
/external/libvpx/libvpx/vp8/common/x86/
Dloopfilter_block_sse2_x86_64.asm206 movdqa xmm8, i5
209 LF_FILTER_HEV_MASK xmm0, xmm1, xmm2, xmm3, xmm4, xmm8, xmm9, xmm10
214 movdqa xmm8, i5
215 LF_FILTER xmm1, xmm2, xmm3, xmm8, xmm0, xmm4
221 movdqa i5, xmm8
229 LF_FILTER_HEV_MASK xmm3, xmm8, xmm0, xmm1, xmm2, xmm4, xmm10, xmm11, xmm9
234 movdqa xmm8, i9
235 LF_FILTER xmm0, xmm1, xmm4, xmm8, xmm3, xmm2
241 movdqa i9, xmm8
249 LF_FILTER_HEV_MASK xmm4, xmm8, xmm0, xmm1, xmm2, xmm3, xmm9, xmm11, xmm10
[all …]
/external/boringssl/win-x86_64/crypto/fipsmodule/
Daesni-gcm-x86_64.asm81 vpxor xmm8,xmm8,xmm4
92 vpxor xmm8,xmm8,XMMWORD[((16+8))+rsp]
156 vpxor xmm8,xmm8,XMMWORD[((112+8))+rsp]
170 vpclmulqdq xmm5,xmm8,xmm3,0x10
173 vpclmulqdq xmm1,xmm8,xmm3,0x01
177 vpclmulqdq xmm2,xmm8,xmm3,0x00
180 vpclmulqdq xmm8,xmm8,xmm3,0x11
195 vpxor xmm7,xmm7,xmm8
286 vpalignr xmm8,xmm4,xmm4,8
339 vpxor xmm8,xmm8,XMMWORD[((16+8))+rsp]
[all …]
Dghash-x86_64.asm1015 movdqu xmm8,XMMWORD[r8]
1020 pxor xmm0,xmm8
1024 pshufd xmm8,xmm0,78
1025 pxor xmm8,xmm0
1052 xorps xmm8,xmm4
1056 pxor xmm8,xmm0
1058 pxor xmm8,xmm1
1060 movdqa xmm9,xmm8
1062 pslldq xmm8,8
1064 pxor xmm0,xmm8
[all …]
/external/boringssl/src/crypto/fipsmodule/aes/asm/
Dvpaes-x86_64.pl88 ## Preserves %xmm6 - %xmm8 so you get some local vectors
202 ## as before. The second uses %xmm6-%xmm8,%xmm11-%xmm13. (Add 6 to %xmm2 and
217 movdqa %xmm2, %xmm8
227 pshufb %xmm6, %xmm8
233 pxor %xmm5, %xmm8
236 pxor %xmm8, %xmm6
248 pshufb %xmm8, %xmm12
260 pshufb %xmm8, %xmm13
264 movdqa %xmm2, %xmm8
266 pshufb %xmm11, %xmm8
[all …]
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/
Dvector-interleave.ll13 ; SSE-NEXT: movdqa %xmm0, %xmm8
14 ; SSE-NEXT: punpcklwd {{.*#+}} xmm8 = xmm8[0],xmm1[0],xmm8[1],xmm1[1],xmm8[2],xmm1[2],xmm8[3],xm…
22 ; SSE-NEXT: movdqa %xmm8, %xmm2
24 ; SSE-NEXT: punpckldq {{.*#+}} xmm8 = xmm8[0],xmm1[0],xmm8[1],xmm1[1]
37 ; SSE-NEXT: movdqa %xmm8, %xmm5
39 ; SSE-NEXT: punpckhwd {{.*#+}} xmm8 = xmm8[4],xmm1[4],xmm8[5],xmm1[5],xmm8[6],xmm1[6],xmm8[7],xm…
55 ; SSE-NEXT: movdqa %xmm8, 16(%rdi)
62 ; AVX1-NEXT: vpunpcklwd {{.*#+}} xmm8 = xmm0[0],xmm1[0],xmm0[1],xmm1[1],xmm0[2],xmm1[2],xmm0[3],…
68 ; AVX1-NEXT: vpunpckhdq {{.*#+}} xmm3 = xmm8[2],xmm1[2],xmm8[3],xmm1[3]
69 ; AVX1-NEXT: vpunpckldq {{.*#+}} xmm0 = xmm8[0],xmm1[0],xmm8[1],xmm1[1]
[all …]
Dbitcast-and-setcc-512.ll11 ; SSE-NEXT: movdqa {{[0-9]+}}(%rsp), %xmm8
40 ; SSE-NEXT: pcmpgtq {{[0-9]+}}(%rsp), %xmm8
41 ; SSE-NEXT: pshufd {{.*#+}} xmm3 = xmm8[0,2,2,3]
53 ; AVX1-NEXT: vextractf128 $1, %ymm3, %xmm8
55 ; AVX1-NEXT: vpcmpgtq %xmm8, %xmm9, %xmm8
57 ; AVX1-NEXT: vpackssdw %xmm8, %xmm1, %xmm8
63 ; AVX1-NEXT: vpackssdw %xmm8, %xmm0, %xmm0
136 ; SSE-NEXT: movapd {{[0-9]+}}(%rsp), %xmm8
165 ; SSE-NEXT: cmpltpd {{[0-9]+}}(%rsp), %xmm8
166 ; SSE-NEXT: shufps {{.*#+}} xmm8 = xmm8[0,2,2,3]
[all …]
Dstack-folding-fp-sse42.ll14 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
22 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
30 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
38 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
49 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
57 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
68 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
77 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
86 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
100 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
[all …]
Dstack-folding-fp-avx1.ll14 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
22 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
30 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
38 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
46 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
54 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
66 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
74 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
86 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
95 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
[all …]
Dstack-folding-int-avx2.ll14 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
24 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
34 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
47 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
54 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
64 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
73 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
83 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
93 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
103 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~…
[all …]
/external/llvm/test/MC/X86/
Dintel-syntax-x86-64-avx.s21 vgatherdps xmm10, xmmword ptr [r15 + 2*xmm9], xmm8
25 vgatherqps xmm10, qword ptr [r15 + 2*xmm9], xmm8
33 vgatherqps xmm10, xmmword ptr [r15 + 2*ymm9], xmm8
53 vpgatherdd xmm10, xmmword ptr [r15 + 2*xmm9], xmm8
57 vpgatherqd xmm10, qword ptr [r15 + 2*xmm9], xmm8
65 vpgatherqd xmm10, xmmword ptr [r15 + 2*ymm9], xmm8

12345678910