/external/llvm/test/MC/X86/ |
D | shuffle-comments.s | 3 palignr $8, %xmm0, %xmm1 4 # CHECK: xmm1 = xmm0[8,9,10,11,12,13,14,15],xmm1[0,1,2,3,4,5,6,7] 5 palignr $8, (%rax), %xmm1 6 # CHECK: xmm1 = mem[8,9,10,11,12,13,14,15],xmm1[0,1,2,3,4,5,6,7] 8 palignr $16, %xmm0, %xmm1 9 # CHECK: xmm1 = xmm1[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15] 10 palignr $16, (%rax), %xmm1 11 # CHECK: xmm1 = xmm1[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15] 13 palignr $0, %xmm0, %xmm1 14 # CHECK: xmm1 = xmm0[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15] [all …]
|
D | x86-32-fma3.s | 5 vfmadd132pd %xmm2, %xmm5, %xmm1 9 vfmadd132pd (%eax), %xmm5, %xmm1 13 vfmadd132ps %xmm2, %xmm5, %xmm1 17 vfmadd132ps (%eax), %xmm5, %xmm1 21 vfmadd213pd %xmm2, %xmm5, %xmm1 25 vfmadd213pd (%eax), %xmm5, %xmm1 29 vfmadd213ps %xmm2, %xmm5, %xmm1 33 vfmadd213ps (%eax), %xmm5, %xmm1 37 vfmadd231pd %xmm2, %xmm5, %xmm1 41 vfmadd231pd (%eax), %xmm5, %xmm1 [all …]
|
D | x86_64-fma4-encoding.s | 6 vfmaddss (%rcx), %xmm1, %xmm0, %xmm0 10 vfmaddss %xmm1, (%rcx),%xmm0, %xmm0 14 vfmaddss %xmm2, %xmm1, %xmm0, %xmm0 18 vfmaddsd (%rcx), %xmm1, %xmm0, %xmm0 22 vfmaddsd %xmm1, (%rcx),%xmm0, %xmm0 26 vfmaddsd %xmm2, %xmm1, %xmm0, %xmm0 30 vfmaddsd %xmm10, %xmm1, %xmm0, %xmm0 34 vfmaddps (%rcx), %xmm1, %xmm0, %xmm0 38 vfmaddps %xmm1, (%rcx),%xmm0, %xmm0 42 vfmaddps %xmm2, %xmm1, %xmm0, %xmm0 [all …]
|
D | x86-32-avx.s | 276 vunpckhps %xmm1, %xmm2, %xmm4 280 vunpckhpd %xmm1, %xmm2, %xmm4 284 vunpcklps %xmm1, %xmm2, %xmm4 288 vunpcklpd %xmm1, %xmm2, %xmm4 308 vcmpps $0, %xmm0, %xmm6, %xmm1 312 vcmpps $0, (%eax), %xmm6, %xmm1 316 vcmpps $7, %xmm0, %xmm6, %xmm1 320 vcmppd $0, %xmm0, %xmm6, %xmm1 324 vcmppd $0, (%eax), %xmm6, %xmm1 328 vcmppd $7, %xmm0, %xmm6, %xmm1 [all …]
|
D | x86_64-xop-encoding.s | 10 vphsubwd (%rcx,%rax), %xmm1 13 vphsubwd %xmm0, %xmm1 18 vphsubdq (%rcx,%rax), %xmm1 21 vphsubdq %xmm0, %xmm1 26 vphsubbw (%rax), %xmm1 29 vphsubbw %xmm2, %xmm1 61 vphadduwd %xmm2, %xmm1 106 vphaddbw (%rcx,%rax), %xmm1 114 vphaddbq (%rcx,%rax), %xmm1 122 vphaddbd (%rcx,%rax), %xmm1 [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | vector-zext.ll | 11 ; SSE2-NEXT: pxor %xmm1, %xmm1 12 …}} xmm0 = xmm0[0],xmm1[0],xmm0[1],xmm1[1],xmm0[2],xmm1[2],xmm0[3],xmm1[3],xmm0[4],xmm1[4],xmm0[5],… 17 ; SSSE3-NEXT: pxor %xmm1, %xmm1 18 …}} xmm0 = xmm0[0],xmm1[0],xmm0[1],xmm1[1],xmm0[2],xmm1[2],xmm0[3],xmm1[3],xmm0[4],xmm1[4],xmm0[5],… 40 ; SSE2-NEXT: movdqa %xmm0, %xmm1 43 … {{.*#+}} xmm1 = xmm1[8],xmm2[8],xmm1[9],xmm2[9],xmm1[10],xmm2[10],xmm1[11],xmm2[11],xmm1[12],xmm2… 48 ; SSSE3-NEXT: movdqa %xmm0, %xmm1 51 … {{.*#+}} xmm1 = xmm1[8],xmm2[8],xmm1[9],xmm2[9],xmm1[10],xmm2[10],xmm1[11],xmm2[11],xmm1[12],xmm2… 56 ; SSE41-NEXT: movdqa %xmm0, %xmm1 58 … pmovzxbw {{.*#+}} xmm0 = xmm1[0],zero,xmm1[1],zero,xmm1[2],zero,xmm1[3],zero,xmm1[4],zero,xmm1… [all …]
|
D | vector-shift-ashr-128.ll | 19 ; SSE2-NEXT: pshufd {{.*#+}} xmm3 = xmm1[2,3,0,1] 23 ; SSE2-NEXT: psrlq %xmm1, %xmm2 27 ; SSE2-NEXT: psrlq %xmm1, %xmm0 38 ; SSE41-NEXT: psrlq %xmm1, %xmm3 39 ; SSE41-NEXT: pshufd {{.*#+}} xmm4 = xmm1[2,3,0,1] 43 ; SSE41-NEXT: psrlq %xmm1, %xmm3 53 ; AVX1-NEXT: vpsrlq %xmm1, %xmm2, %xmm3 54 ; AVX1-NEXT: vpshufd {{.*#+}} xmm4 = xmm1[2,3,0,1] 57 ; AVX1-NEXT: vpsrlq %xmm1, %xmm0, %xmm1 59 ; AVX1-NEXT: vpblendw {{.*#+}} xmm0 = xmm1[0,1,2,3],xmm0[4,5,6,7] [all …]
|
D | vector-tzcnt-128.ll | 16 ; SSE2-NEXT: movd %rax, %xmm1 22 ; SSE2-NEXT: punpcklqdq {{.*#+}} xmm1 = xmm1[0],xmm0[0] 23 ; SSE2-NEXT: movdqa %xmm1, %xmm0 32 ; SSE3-NEXT: movd %rax, %xmm1 38 ; SSE3-NEXT: punpcklqdq {{.*#+}} xmm1 = xmm1[0],xmm0[0] 39 ; SSE3-NEXT: movdqa %xmm1, %xmm0 48 ; SSSE3-NEXT: movd %rax, %xmm1 54 ; SSSE3-NEXT: punpcklqdq {{.*#+}} xmm1 = xmm1[0],xmm0[0] 55 ; SSSE3-NEXT: movdqa %xmm1, %xmm0 64 ; SSE41-NEXT: movd %rax, %xmm1 [all …]
|
D | vector-popcnt-128.ll | 12 ; SSE2-NEXT: movdqa %xmm0, %xmm1 13 ; SSE2-NEXT: psrlq $1, %xmm1 14 ; SSE2-NEXT: pand {{.*}}(%rip), %xmm1 15 ; SSE2-NEXT: psubq %xmm1, %xmm0 16 ; SSE2-NEXT: movdqa {{.*#+}} xmm1 = [3689348814741910323,3689348814741910323] 18 ; SSE2-NEXT: pand %xmm1, %xmm2 20 ; SSE2-NEXT: pand %xmm1, %xmm0 22 ; SSE2-NEXT: movdqa %xmm0, %xmm1 23 ; SSE2-NEXT: psrlq $4, %xmm1 24 ; SSE2-NEXT: paddq %xmm0, %xmm1 [all …]
|
D | vector-shuffle-128-v4.ll | 308 ; SSE2-NEXT: shufps {{.*#+}} xmm1 = xmm1[0,0],xmm0[2,0] 309 ; SSE2-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,0] 314 ; SSE3-NEXT: shufps {{.*#+}} xmm1 = xmm1[0,0],xmm0[2,0] 315 ; SSE3-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,0] 320 ; SSSE3-NEXT: shufps {{.*#+}} xmm1 = xmm1[0,0],xmm0[2,0] 321 ; SSSE3-NEXT: shufps {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,0] 326 ; SSE41-NEXT: pshufd {{.*#+}} xmm1 = xmm1[0,1,2,0] 327 ; SSE41-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1,2,3,4,5],xmm1[6,7] 332 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[0,1,2,0] 333 ; AVX1-NEXT: vpblendw {{.*#+}} xmm0 = xmm0[0,1,2,3,4,5],xmm1[6,7] [all …]
|
D | vec_cmp_sint-128.ll | 19 ; SSE2-NEXT: pcmpeqd %xmm1, %xmm0 20 ; SSE2-NEXT: pshufd {{.*#+}} xmm1 = xmm0[1,0,3,2] 21 ; SSE2-NEXT: pand %xmm1, %xmm0 26 ; SSE41-NEXT: pcmpeqq %xmm1, %xmm0 31 ; SSE42-NEXT: pcmpeqq %xmm1, %xmm0 36 ; AVX-NEXT: vpcmpeqq %xmm1, %xmm0, %xmm0 41 ; XOP-NEXT: vpcomeqq %xmm1, %xmm0, %xmm0 51 ; SSE-NEXT: pcmpeqd %xmm1, %xmm0 56 ; AVX-NEXT: vpcmpeqd %xmm1, %xmm0, %xmm0 61 ; XOP-NEXT: vpcomeqd %xmm1, %xmm0, %xmm0 [all …]
|
D | vec_cmp_uint-128.ll | 19 ; SSE2-NEXT: pcmpeqd %xmm1, %xmm0 20 ; SSE2-NEXT: pshufd {{.*#+}} xmm1 = xmm0[1,0,3,2] 21 ; SSE2-NEXT: pand %xmm1, %xmm0 26 ; SSE41-NEXT: pcmpeqq %xmm1, %xmm0 31 ; SSE42-NEXT: pcmpeqq %xmm1, %xmm0 36 ; AVX-NEXT: vpcmpeqq %xmm1, %xmm0, %xmm0 41 ; XOP-NEXT: vpcomeqq %xmm1, %xmm0, %xmm0 51 ; SSE-NEXT: pcmpeqd %xmm1, %xmm0 56 ; AVX-NEXT: vpcmpeqd %xmm1, %xmm0, %xmm0 61 ; XOP-NEXT: vpcomeqd %xmm1, %xmm0, %xmm0 [all …]
|
D | vector-shift-lshr-128.ll | 19 ; SSE2-NEXT: pshufd {{.*#+}} xmm3 = xmm1[2,3,0,1] 22 ; SSE2-NEXT: psrlq %xmm1, %xmm0 30 ; SSE41-NEXT: psrlq %xmm1, %xmm2 31 ; SSE41-NEXT: pshufd {{.*#+}} xmm1 = xmm1[2,3,0,1] 32 ; SSE41-NEXT: psrlq %xmm1, %xmm0 38 ; AVX1-NEXT: vpsrlq %xmm1, %xmm0, %xmm2 39 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[2,3,0,1] 40 ; AVX1-NEXT: vpsrlq %xmm1, %xmm0, %xmm0 46 ; AVX2-NEXT: vpsrlvq %xmm1, %xmm0, %xmm0 52 ; XOPAVX1-NEXT: vpsubq %xmm1, %xmm2, %xmm1 [all …]
|
D | vector-shift-shl-128.ll | 19 ; SSE2-NEXT: pshufd {{.*#+}} xmm3 = xmm1[2,3,0,1] 22 ; SSE2-NEXT: psllq %xmm1, %xmm0 30 ; SSE41-NEXT: psllq %xmm1, %xmm2 31 ; SSE41-NEXT: pshufd {{.*#+}} xmm1 = xmm1[2,3,0,1] 32 ; SSE41-NEXT: psllq %xmm1, %xmm0 38 ; AVX1-NEXT: vpsllq %xmm1, %xmm0, %xmm2 39 ; AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[2,3,0,1] 40 ; AVX1-NEXT: vpsllq %xmm1, %xmm0, %xmm0 46 ; AVX2-NEXT: vpsllvq %xmm1, %xmm0, %xmm0 51 ; XOPAVX1-NEXT: vpshlq %xmm1, %xmm0, %xmm0 [all …]
|
D | vector-lzcnt-128.ll | 18 ; SSE2-NEXT: movd %rax, %xmm1 25 ; SSE2-NEXT: punpcklqdq {{.*#+}} xmm1 = xmm1[0],xmm0[0] 26 ; SSE2-NEXT: movdqa %xmm1, %xmm0 36 ; SSE3-NEXT: movd %rax, %xmm1 43 ; SSE3-NEXT: punpcklqdq {{.*#+}} xmm1 = xmm1[0],xmm0[0] 44 ; SSE3-NEXT: movdqa %xmm1, %xmm0 54 ; SSSE3-NEXT: movd %rax, %xmm1 61 ; SSSE3-NEXT: punpcklqdq {{.*#+}} xmm1 = xmm1[0],xmm0[0] 62 ; SSSE3-NEXT: movdqa %xmm1, %xmm0 72 ; SSE41-NEXT: movd %rax, %xmm1 [all …]
|
D | vec_int_to_fp.ll | 18 ; SSE-NEXT: cvtsi2sdq %rax, %xmm1 23 ; SSE-NEXT: unpcklpd {{.*#+}} xmm1 = xmm1[0],xmm0[0] 24 ; SSE-NEXT: movapd %xmm1, %xmm0 30 ; AVX-NEXT: vcvtsi2sdq %rax, %xmm0, %xmm1 34 ; AVX-NEXT: vunpcklpd {{.*#+}} xmm0 = xmm0[0],xmm1[0] 172 ; SSE-NEXT: movd %xmm1, %rax 174 ; SSE-NEXT: pshufd {{.*#+}} xmm0 = xmm1[2,3,0,1] 180 ; SSE-NEXT: movapd %xmm3, %xmm1 185 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm1 186 ; AVX1-NEXT: vpextrq $1, %xmm1, %rax [all …]
|
D | vector-rotate-128.ll | 20 ; SSE2-NEXT: psubq %xmm1, %xmm2 21 ; SSE2-NEXT: pshufd {{.*#+}} xmm3 = xmm1[2,3,0,1] 25 ; SSE2-NEXT: psllq %xmm1, %xmm3 28 ; SSE2-NEXT: movdqa %xmm0, %xmm1 29 ; SSE2-NEXT: psrlq %xmm3, %xmm1 31 ; SSE2-NEXT: movsd {{.*#+}} xmm1 = xmm0[0],xmm1[1] 32 ; SSE2-NEXT: orpd %xmm4, %xmm1 33 ; SSE2-NEXT: movapd %xmm1, %xmm0 39 ; SSE41-NEXT: psubq %xmm1, %xmm2 41 ; SSE41-NEXT: psllq %xmm1, %xmm3 [all …]
|
D | vector-trunc.ll | 12 ; SSE2-NEXT: pshufd {{.*#+}} xmm1 = xmm1[0,2,2,3] 14 ; SSE2-NEXT: punpcklqdq {{.*#+}} xmm0 = xmm0[0],xmm1[0] 16 ; SSE2-NEXT: pshufd {{.*#+}} xmm1 = xmm2[0,2,2,3] 17 ; SSE2-NEXT: punpcklqdq {{.*#+}} xmm1 = xmm1[0],xmm3[0] 22 ; SSSE3-NEXT: pshufd {{.*#+}} xmm1 = xmm1[0,2,2,3] 24 ; SSSE3-NEXT: punpcklqdq {{.*#+}} xmm0 = xmm0[0],xmm1[0] 26 ; SSSE3-NEXT: pshufd {{.*#+}} xmm1 = xmm2[0,2,2,3] 27 ; SSSE3-NEXT: punpcklqdq {{.*#+}} xmm1 = xmm1[0],xmm3[0] 32 ; SSE41-NEXT: pshufd {{.*#+}} xmm1 = xmm1[0,1,0,2] 34 ; SSE41-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1,2,3],xmm1[4,5,6,7] [all …]
|
D | sse-minmax.ll | 16 ; CHECK-NEXT: maxsd %xmm1, %xmm0 19 ; UNSAFE-NEXT: maxsd %xmm1, %xmm0 22 ; FINITE-NEXT: maxsd %xmm1, %xmm0 31 ; CHECK-NEXT: minsd %xmm1, %xmm0 34 ; UNSAFE-NEXT: minsd %xmm1, %xmm0 37 ; FINITE-NEXT: minsd %xmm1, %xmm0 46 ; CHECK-NEXT: minsd %xmm0, %xmm1 47 ; CHECK-NEXT: movap{{[sd]}} %xmm1, %xmm0 50 ; UNSAFE-NEXT: minsd %xmm1, %xmm0 53 ; FINITE-NEXT: minsd %xmm0, %xmm1 [all …]
|
D | vector-shift-ashr-256.ll | 26 ; AVX1-NEXT: vpsrlq %xmm1, %xmm3, %xmm4 27 ; AVX1-NEXT: vpshufd {{.*#+}} xmm5 = xmm1[2,3,0,1] 30 ; AVX1-NEXT: vpsrlq %xmm1, %xmm0, %xmm1 32 ; AVX1-NEXT: vpblendw {{.*#+}} xmm0 = xmm1[0,1,2,3],xmm0[4,5,6,7] 54 ; XOPAVX1-NEXT: vpsubq %xmm1, %xmm3, %xmm1 55 ; XOPAVX1-NEXT: vpshaq %xmm1, %xmm0, %xmm0 88 ; AVX1-NEXT: vpsrldq {{.*#+}} xmm3 = xmm1[12,13,14,15],zero,zero,zero,zero,zero,zero,zero,zero,z… 90 ; AVX1-NEXT: vpsrlq $32, %xmm1, %xmm4 93 ; AVX1-NEXT: vpunpckhdq {{.*#+}} xmm4 = xmm1[2],xmm5[2],xmm1[3],xmm5[3] 95 ; AVX1-NEXT: vpmovzxdq {{.*#+}} xmm1 = xmm1[0],zero,xmm1[1],zero [all …]
|
D | vector-shuffle-128-v2.ll | 65 ; SSE-NEXT: pshufd {{.*#+}} xmm0 = xmm1[0,1,0,1] 70 ; AVX1-NEXT: vpshufd {{.*#+}} xmm0 = xmm1[0,1,0,1] 75 ; AVX2-NEXT: vpbroadcastq %xmm1, %xmm0 80 ; AVX512VL-NEXT: vpbroadcastq %xmm1, %xmm0 88 ; SSE-NEXT: pshufd {{.*#+}} xmm0 = xmm1[2,3,0,1] 93 ; AVX-NEXT: vpshufd {{.*#+}} xmm0 = xmm1[2,3,0,1] 101 ; SSE-NEXT: pshufd {{.*#+}} xmm0 = xmm1[2,3,2,3] 106 ; AVX-NEXT: vpshufd {{.*#+}} xmm0 = xmm1[2,3,2,3] 170 ; SSE2-NEXT: movlhps {{.*#+}} xmm1 = xmm1[0,0] 171 ; SSE2-NEXT: movaps %xmm1, %xmm0 [all …]
|
D | vector-sext.ll | 49 … {{.*#+}} xmm1 = xmm1[8],xmm0[8],xmm1[9],xmm0[9],xmm1[10],xmm0[10],xmm1[11],xmm0[11],xmm1[12],xmm0… 50 ; SSE2-NEXT: psraw $8, %xmm1 58 … {{.*#+}} xmm1 = xmm1[8],xmm0[8],xmm1[9],xmm0[9],xmm1[10],xmm0[10],xmm1[11],xmm0[11],xmm1[12],xmm0… 59 ; SSSE3-NEXT: psraw $8, %xmm1 67 ; SSE41-NEXT: pmovsxbw %xmm0, %xmm1 73 ; AVX1-NEXT: vpmovsxbw %xmm0, %xmm1 88 ; X32-SSE41-NEXT: pmovsxbw %xmm0, %xmm1 134 …cklbw {{.*#+}} xmm1 = xmm1[0],xmm0[0],xmm1[1],xmm0[1],xmm1[2],xmm0[2],xmm1[3],xmm0[3],xmm1[4],xmm0… 135 ; SSE2-NEXT: punpcklwd {{.*#+}} xmm2 = xmm2[0],xmm1[0],xmm2[1],xmm1[1],xmm2[2],xmm1[2],xmm2[3],x… 139 ; SSE2-NEXT: punpcklwd {{.*#+}} xmm1 = xmm1[0],xmm0[0],xmm1[1],xmm0[1],xmm1[2],xmm0[2],xmm1[3],x… [all …]
|
D | combine-and.ll | 12 ; CHECK-NEXT: pxor %xmm1, %xmm1 13 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,3,4,5,6,7] 22 ; CHECK-NEXT: pxor %xmm1, %xmm1 23 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3],xmm1[4,5,6,7] 32 ; CHECK-NEXT: pxor %xmm1, %xmm1 33 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1,2,3],xmm0[4,5],xmm1[6,7] 42 ; CHECK-NEXT: pxor %xmm1, %xmm1 43 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1,2,3,4,5],xmm0[6,7] 52 ; CHECK-NEXT: pxor %xmm1, %xmm1 53 ; CHECK-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,3],xmm0[4,5],xmm1[6,7] [all …]
|
/external/boringssl/src/crypto/aes/asm/ |
D | vpaes-x86_64.pl | 79 ## Clobbers %xmm1-%xmm5, %r9, %r10, %r11, %rax 89 movdqa %xmm9, %xmm1 91 pandn %xmm0, %xmm1 93 psrld \$4, %xmm1 97 pshufb %xmm1, %xmm0 114 movdqa -0x40(%r11,%r10), %xmm1 # .Lk_mc_forward[] 121 pshufb %xmm1, %xmm0 # 0 = B 127 pshufb %xmm1, %xmm0 # 0 = 2B+C 134 movdqa %xmm9, %xmm1 # 1 : i 136 pandn %xmm0, %xmm1 # 1 = i<<4 [all …]
|
/external/libyuv/files/source/ |
D | compare_win.cc | 34 movdqu xmm1, [eax] in SumSquareError_SSE2() 38 movdqa xmm3, xmm1 // abs trick in SumSquareError_SSE2() 39 psubusb xmm1, xmm2 in SumSquareError_SSE2() 41 por xmm1, xmm2 in SumSquareError_SSE2() 42 movdqa xmm2, xmm1 in SumSquareError_SSE2() 43 punpcklbw xmm1, xmm5 in SumSquareError_SSE2() 45 pmaddwd xmm1, xmm1 in SumSquareError_SSE2() 47 paddd xmm0, xmm1 in SumSquareError_SSE2() 52 pshufd xmm1, xmm0, 0xee in SumSquareError_SSE2() 53 paddd xmm0, xmm1 in SumSquareError_SSE2() [all …]
|