/external/llvm/test/CodeGen/X86/ |
D | sad.ll | 169 ; SSE2-NEXT: pshufd {{.*#+}} xmm8 = xmm1[2,3,0,1] 170 …w {{.*#+}} xmm8 = xmm8[0],xmm12[0],xmm8[1],xmm12[1],xmm8[2],xmm12[2],xmm8[3],xmm12[3],xmm8[4],xmm1… 204 ; SSE2-NEXT: movdqa %xmm8, %xmm3 205 ; SSE2-NEXT: punpckhwd {{.*#+}} xmm8 = xmm8[4],xmm12[4],xmm8[5],xmm12[5],xmm8[6],xmm12[6],xmm8[7… 212 ; SSE2-NEXT: psubd %xmm11, %xmm8 220 ; SSE2-NEXT: movdqa %xmm8, %xmm4 222 ; SSE2-NEXT: paddd %xmm4, %xmm8 223 ; SSE2-NEXT: pxor %xmm4, %xmm8 261 ; SSE2-NEXT: paddd %xmm8, %xmm1 402 ; SSE2-NEXT: pxor %xmm8, %xmm8 [all …]
|
D | stack-folding-fp-sse42.ll | 14 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 22 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 30 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 38 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 47 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 55 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 64 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 73 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 82 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 96 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… [all …]
|
D | stack-folding-fp-avx1.ll | 14 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 22 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 30 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 38 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 46 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 54 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 63 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 71 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 80 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 89 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… [all …]
|
D | stack-folding-int-avx2.ll | 14 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 25 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 36 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 50 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 57 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 67 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 76 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 85 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 94 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 103 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… [all …]
|
/external/boringssl/linux-x86_64/crypto/cipher_extra/ |
D | aes128gcmsiv-x86_64.S | 757 vmovdqa %xmm3,%xmm8 762 vpxor (%rcx),%xmm8,%xmm8 768 vaesenc %xmm12,%xmm8,%xmm8 775 vaesenc %xmm12,%xmm8,%xmm8 782 vaesenc %xmm12,%xmm8,%xmm8 789 vaesenc %xmm12,%xmm8,%xmm8 797 vaesenc %xmm12,%xmm8,%xmm8 803 vaesenc %xmm12,%xmm8,%xmm8 809 vaesenc %xmm12,%xmm8,%xmm8 815 vaesenc %xmm12,%xmm8,%xmm8 [all …]
|
D | chacha20_poly1305_x86_64.S | 274 movdqu 16(%r9),%xmm8 279 movdqa %xmm8,64(%rbp) 286 paddd %xmm12,%xmm8 287 pxor %xmm8,%xmm4 295 paddd %xmm12,%xmm8 296 pxor %xmm8,%xmm4 307 paddd %xmm12,%xmm8 308 pxor %xmm8,%xmm4 316 paddd %xmm12,%xmm8 317 pxor %xmm8,%xmm4 [all …]
|
/external/boringssl/mac-x86_64/crypto/cipher_extra/ |
D | aes128gcmsiv-x86_64.S | 755 vmovdqa %xmm3,%xmm8 760 vpxor (%rcx),%xmm8,%xmm8 766 vaesenc %xmm12,%xmm8,%xmm8 773 vaesenc %xmm12,%xmm8,%xmm8 780 vaesenc %xmm12,%xmm8,%xmm8 787 vaesenc %xmm12,%xmm8,%xmm8 795 vaesenc %xmm12,%xmm8,%xmm8 801 vaesenc %xmm12,%xmm8,%xmm8 807 vaesenc %xmm12,%xmm8,%xmm8 813 vaesenc %xmm12,%xmm8,%xmm8 [all …]
|
D | chacha20_poly1305_x86_64.S | 273 movdqu 16(%r9),%xmm8 278 movdqa %xmm8,64(%rbp) 285 paddd %xmm12,%xmm8 286 pxor %xmm8,%xmm4 294 paddd %xmm12,%xmm8 295 pxor %xmm8,%xmm4 306 paddd %xmm12,%xmm8 307 pxor %xmm8,%xmm4 315 paddd %xmm12,%xmm8 316 pxor %xmm8,%xmm4 [all …]
|
/external/boringssl/win-x86_64/crypto/cipher_extra/ |
D | aes128gcmsiv-x86_64.asm | 850 vmovdqa xmm8,xmm3 855 vpxor xmm8,xmm8,XMMWORD[rcx] 861 vaesenc xmm8,xmm8,xmm12 868 vaesenc xmm8,xmm8,xmm12 875 vaesenc xmm8,xmm8,xmm12 882 vaesenc xmm8,xmm8,xmm12 890 vaesenc xmm8,xmm8,xmm12 896 vaesenc xmm8,xmm8,xmm12 902 vaesenc xmm8,xmm8,xmm12 908 vaesenc xmm8,xmm8,xmm12 [all …]
|
/external/boringssl/linux-x86_64/crypto/fipsmodule/ |
D | aesni-gcm-x86_64.S | 82 vpxor %xmm4,%xmm8,%xmm8 93 vpxor 16+8(%rsp),%xmm8,%xmm8 157 vpxor 112+8(%rsp),%xmm8,%xmm8 171 vpclmulqdq $0x10,%xmm3,%xmm8,%xmm5 174 vpclmulqdq $0x01,%xmm3,%xmm8,%xmm1 178 vpclmulqdq $0x00,%xmm3,%xmm8,%xmm2 181 vpclmulqdq $0x11,%xmm3,%xmm8,%xmm8 196 vpxor %xmm8,%xmm7,%xmm7 287 vpalignr $8,%xmm4,%xmm4,%xmm8 340 vpxor 16+8(%rsp),%xmm8,%xmm8 [all …]
|
D | ghash-x86_64.S | 991 movdqu 0(%rdx),%xmm8 996 pxor %xmm8,%xmm0 1000 pshufd $78,%xmm0,%xmm8 1001 pxor %xmm0,%xmm8 1028 xorps %xmm4,%xmm8 1032 pxor %xmm0,%xmm8 1034 pxor %xmm1,%xmm8 1036 movdqa %xmm8,%xmm9 1038 pslldq $8,%xmm8 1040 pxor %xmm8,%xmm0 [all …]
|
D | sha1-x86_64.S | 1332 movdqa %xmm3,%xmm8 1340 psrldq $4,%xmm8 1346 pxor %xmm2,%xmm8 1350 pxor %xmm8,%xmm4 1360 movdqa %xmm4,%xmm8 1366 psrld $31,%xmm8 1376 por %xmm8,%xmm4 1417 movdqa %xmm5,%xmm8 1423 pslldq $12,%xmm8 1431 movdqa %xmm8,%xmm10 [all …]
|
/external/boringssl/mac-x86_64/crypto/fipsmodule/ |
D | aesni-gcm-x86_64.S | 82 vpxor %xmm4,%xmm8,%xmm8 93 vpxor 16+8(%rsp),%xmm8,%xmm8 157 vpxor 112+8(%rsp),%xmm8,%xmm8 171 vpclmulqdq $0x10,%xmm3,%xmm8,%xmm5 174 vpclmulqdq $0x01,%xmm3,%xmm8,%xmm1 178 vpclmulqdq $0x00,%xmm3,%xmm8,%xmm2 181 vpclmulqdq $0x11,%xmm3,%xmm8,%xmm8 196 vpxor %xmm8,%xmm7,%xmm7 287 vpalignr $8,%xmm4,%xmm4,%xmm8 340 vpxor 16+8(%rsp),%xmm8,%xmm8 [all …]
|
D | ghash-x86_64.S | 978 movdqu 0(%rdx),%xmm8 983 pxor %xmm8,%xmm0 987 pshufd $78,%xmm0,%xmm8 988 pxor %xmm0,%xmm8 1015 xorps %xmm4,%xmm8 1019 pxor %xmm0,%xmm8 1021 pxor %xmm1,%xmm8 1023 movdqa %xmm8,%xmm9 1025 pslldq $8,%xmm8 1027 pxor %xmm8,%xmm0 [all …]
|
D | sha1-x86_64.S | 1331 movdqa %xmm3,%xmm8 1339 psrldq $4,%xmm8 1345 pxor %xmm2,%xmm8 1349 pxor %xmm8,%xmm4 1359 movdqa %xmm4,%xmm8 1365 psrld $31,%xmm8 1375 por %xmm8,%xmm4 1416 movdqa %xmm5,%xmm8 1422 pslldq $12,%xmm8 1430 movdqa %xmm8,%xmm10 [all …]
|
/external/libvpx/libvpx/vp8/common/x86/ |
D | loopfilter_block_sse2_x86_64.asm | 206 movdqa xmm8, i5 209 LF_FILTER_HEV_MASK xmm0, xmm1, xmm2, xmm3, xmm4, xmm8, xmm9, xmm10 214 movdqa xmm8, i5 215 LF_FILTER xmm1, xmm2, xmm3, xmm8, xmm0, xmm4 221 movdqa i5, xmm8 229 LF_FILTER_HEV_MASK xmm3, xmm8, xmm0, xmm1, xmm2, xmm4, xmm10, xmm11, xmm9 234 movdqa xmm8, i9 235 LF_FILTER xmm0, xmm1, xmm4, xmm8, xmm3, xmm2 241 movdqa i9, xmm8 249 LF_FILTER_HEV_MASK xmm4, xmm8, xmm0, xmm1, xmm2, xmm3, xmm9, xmm11, xmm10 [all …]
|
/external/boringssl/win-x86_64/crypto/fipsmodule/ |
D | aesni-gcm-x86_64.asm | 81 vpxor xmm8,xmm8,xmm4 92 vpxor xmm8,xmm8,XMMWORD[((16+8))+rsp] 156 vpxor xmm8,xmm8,XMMWORD[((112+8))+rsp] 170 vpclmulqdq xmm5,xmm8,xmm3,0x10 173 vpclmulqdq xmm1,xmm8,xmm3,0x01 177 vpclmulqdq xmm2,xmm8,xmm3,0x00 180 vpclmulqdq xmm8,xmm8,xmm3,0x11 195 vpxor xmm7,xmm7,xmm8 286 vpalignr xmm8,xmm4,xmm4,8 339 vpxor xmm8,xmm8,XMMWORD[((16+8))+rsp] [all …]
|
D | ghash-x86_64.asm | 1015 movdqu xmm8,XMMWORD[r8] 1020 pxor xmm0,xmm8 1024 pshufd xmm8,xmm0,78 1025 pxor xmm8,xmm0 1052 xorps xmm8,xmm4 1056 pxor xmm8,xmm0 1058 pxor xmm8,xmm1 1060 movdqa xmm9,xmm8 1062 pslldq xmm8,8 1064 pxor xmm0,xmm8 [all …]
|
/external/boringssl/src/crypto/fipsmodule/aes/asm/ |
D | vpaes-x86_64.pl | 88 ## Preserves %xmm6 - %xmm8 so you get some local vectors 202 ## as before. The second uses %xmm6-%xmm8,%xmm11-%xmm13. (Add 6 to %xmm2 and 217 movdqa %xmm2, %xmm8 227 pshufb %xmm6, %xmm8 233 pxor %xmm5, %xmm8 236 pxor %xmm8, %xmm6 248 pshufb %xmm8, %xmm12 260 pshufb %xmm8, %xmm13 264 movdqa %xmm2, %xmm8 266 pshufb %xmm11, %xmm8 [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/ |
D | vector-interleave.ll | 13 ; SSE-NEXT: movdqa %xmm0, %xmm8 14 ; SSE-NEXT: punpcklwd {{.*#+}} xmm8 = xmm8[0],xmm1[0],xmm8[1],xmm1[1],xmm8[2],xmm1[2],xmm8[3],xm… 22 ; SSE-NEXT: movdqa %xmm8, %xmm2 24 ; SSE-NEXT: punpckldq {{.*#+}} xmm8 = xmm8[0],xmm1[0],xmm8[1],xmm1[1] 37 ; SSE-NEXT: movdqa %xmm8, %xmm5 39 ; SSE-NEXT: punpckhwd {{.*#+}} xmm8 = xmm8[4],xmm1[4],xmm8[5],xmm1[5],xmm8[6],xmm1[6],xmm8[7],xm… 55 ; SSE-NEXT: movdqa %xmm8, 16(%rdi) 62 ; AVX1-NEXT: vpunpcklwd {{.*#+}} xmm8 = xmm0[0],xmm1[0],xmm0[1],xmm1[1],xmm0[2],xmm1[2],xmm0[3],… 68 ; AVX1-NEXT: vpunpckhdq {{.*#+}} xmm3 = xmm8[2],xmm1[2],xmm8[3],xmm1[3] 69 ; AVX1-NEXT: vpunpckldq {{.*#+}} xmm0 = xmm8[0],xmm1[0],xmm8[1],xmm1[1] [all …]
|
D | bitcast-and-setcc-512.ll | 11 ; SSE-NEXT: movdqa {{[0-9]+}}(%rsp), %xmm8 40 ; SSE-NEXT: pcmpgtq {{[0-9]+}}(%rsp), %xmm8 41 ; SSE-NEXT: pshufd {{.*#+}} xmm3 = xmm8[0,2,2,3] 53 ; AVX1-NEXT: vextractf128 $1, %ymm3, %xmm8 55 ; AVX1-NEXT: vpcmpgtq %xmm8, %xmm9, %xmm8 57 ; AVX1-NEXT: vpackssdw %xmm8, %xmm1, %xmm8 63 ; AVX1-NEXT: vpackssdw %xmm8, %xmm0, %xmm0 136 ; SSE-NEXT: movapd {{[0-9]+}}(%rsp), %xmm8 165 ; SSE-NEXT: cmpltpd {{[0-9]+}}(%rsp), %xmm8 166 ; SSE-NEXT: shufps {{.*#+}} xmm8 = xmm8[0,2,2,3] [all …]
|
D | stack-folding-fp-sse42.ll | 14 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 22 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 30 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 38 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 49 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 57 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 68 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 77 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 86 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 100 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… [all …]
|
D | stack-folding-fp-avx1.ll | 14 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 22 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 30 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 38 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 46 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 54 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 66 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 74 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 86 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 95 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… [all …]
|
D | stack-folding-int-avx2.ll | 14 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 24 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 34 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 47 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 54 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 64 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 73 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 83 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 93 …fect "nop", "=x,~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… 103 …m sideeffect "nop", "=x,~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~… [all …]
|
/external/llvm/test/MC/X86/ |
D | intel-syntax-x86-64-avx.s | 21 vgatherdps xmm10, xmmword ptr [r15 + 2*xmm9], xmm8 25 vgatherqps xmm10, qword ptr [r15 + 2*xmm9], xmm8 33 vgatherqps xmm10, xmmword ptr [r15 + 2*ymm9], xmm8 53 vpgatherdd xmm10, xmmword ptr [r15 + 2*xmm9], xmm8 57 vpgatherqd xmm10, qword ptr [r15 + 2*xmm9], xmm8 65 vpgatherqd xmm10, xmmword ptr [r15 + 2*ymm9], xmm8
|