/external/llvm-project/llvm/test/CodeGen/X86/ |
D | pr30430.ll | 14 ; CHECK-NEXT: vmovss {{.*#+}} xmm8 = mem[0],zero,zero,zero 15 ; CHECK-NEXT: vmovss {{.*#+}} xmm8 = mem[0],zero,zero,zero 16 ; CHECK-NEXT: vmovss {{.*#+}} xmm8 = mem[0],zero,zero,zero 17 ; CHECK-NEXT: vmovss {{.*#+}} xmm8 = mem[0],zero,zero,zero 18 ; CHECK-NEXT: vmovss {{.*#+}} xmm8 = mem[0],zero,zero,zero 19 ; CHECK-NEXT: vmovss {{.*#+}} xmm8 = mem[0],zero,zero,zero 20 ; CHECK-NEXT: vmovss {{.*#+}} xmm8 = mem[0],zero,zero,zero 21 ; CHECK-NEXT: vmovss {{.*#+}} xmm8 = mem[0],zero,zero,zero 22 ; CHECK-NEXT: vmovss %xmm0, {{[0-9]+}}(%rsp) 23 ; CHECK-NEXT: vmovss %xmm1, {{[0-9]+}}(%rsp) [all …]
|
D | avx1-logical-load-folding.ll | 11 ; X86-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 13 ; X86-NEXT: vmovss %xmm0, (%eax) 18 ; X64-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 20 ; X64-NEXT: vmovss %xmm0, (%rsi) 40 ; X86-NEXT: vmovss %xmm0, (%eax) 47 ; X64-NEXT: vmovss %xmm0, (%rsi) 67 ; X86-NEXT: vmovss %xmm0, (%eax) 74 ; X64-NEXT: vmovss %xmm0, (%rsi) 93 ; X86-NEXT: vmovss %xmm0, (%eax) 100 ; X64-NEXT: vmovss %xmm0, (%rsi)
|
D | pr46877.ll | 9 ; CHECK-NEXT: vmovss {{.*#+}} xmm14 = mem[0],zero,zero,zero 10 ; CHECK-NEXT: vmovss {{.*#+}} xmm10 = mem[0],zero,zero,zero 11 ; CHECK-NEXT: vmovss {{.*#+}} xmm13 = mem[0],zero,zero,zero 22 ; CHECK-NEXT: vmovss %xmm2, {{[-0-9]+}}(%r{{[sb]}}p) # 4-byte Spill 32 ; CHECK-NEXT: vmovss {{.*#+}} xmm5 = mem[0],zero,zero,zero 35 ; CHECK-NEXT: vmovss {{.*#+}} xmm7 = mem[0],zero,zero,zero 39 ; CHECK-NEXT: vmovss {{.*#+}} xmm9 = mem[0],zero,zero,zero 41 ; CHECK-NEXT: vmovss %xmm6, {{[-0-9]+}}(%r{{[sb]}}p) # 4-byte Spill 45 ; CHECK-NEXT: vmovss {{.*#+}} xmm5 = mem[0],zero,zero,zero 49 ; CHECK-NEXT: vmovss {{.*#+}} xmm3 = mem[0],zero,zero,zero [all …]
|
D | select-of-fp-constants.ll | 35 ; X64-AVX-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 54 ; X86-AVX2-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 63 ; X86-AVX512F-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 80 ; X64-AVX2-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 81 ; X64-AVX2-NEXT: vmovss {{.*#+}} xmm2 = mem[0],zero,zero,zero 88 ; X64-AVX512F-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 89 ; X64-AVX512F-NEXT: vmovss {{.*}}(%rip), %xmm0 {%k1}
|
D | vector-reduce-fmax.ll | 62 ; AVX512-NEXT: vmovss %xmm2, %xmm1, %xmm1 {%k1} 130 ; AVX512-NEXT: vmovss %xmm1, %xmm2, %xmm2 {%k1} 134 ; AVX512-NEXT: vmovss %xmm1, %xmm0, %xmm0 {%k1} 225 ; AVX512-NEXT: vmovss %xmm3, %xmm4, %xmm4 {%k1} 228 ; AVX512-NEXT: vmovss %xmm2, %xmm0, %xmm0 {%k1} 231 ; AVX512-NEXT: vmovss %xmm1, %xmm0, %xmm0 {%k1} 350 ; AVX512BW-NEXT: vmovss %xmm7, %xmm1, %xmm1 {%k1} 353 ; AVX512BW-NEXT: vmovss %xmm6, %xmm0, %xmm0 {%k1} 356 ; AVX512BW-NEXT: vmovss %xmm5, %xmm0, %xmm0 {%k1} 359 ; AVX512BW-NEXT: vmovss %xmm3, %xmm0, %xmm0 {%k1} [all …]
|
D | vector-reduce-fmin.ll | 54 ; AVX512-NEXT: vmovss %xmm2, %xmm1, %xmm1 {%k1} 146 ; AVX512-NEXT: vmovss %xmm3, %xmm4, %xmm4 {%k1} 149 ; AVX512-NEXT: vmovss %xmm2, %xmm0, %xmm0 {%k1} 152 ; AVX512-NEXT: vmovss %xmm1, %xmm0, %xmm0 {%k1} 271 ; AVX512BW-NEXT: vmovss %xmm7, %xmm1, %xmm1 {%k1} 274 ; AVX512BW-NEXT: vmovss %xmm6, %xmm0, %xmm0 {%k1} 277 ; AVX512BW-NEXT: vmovss %xmm5, %xmm0, %xmm0 {%k1} 280 ; AVX512BW-NEXT: vmovss %xmm3, %xmm0, %xmm0 {%k1} 283 ; AVX512BW-NEXT: vmovss %xmm4, %xmm0, %xmm0 {%k1} 286 ; AVX512BW-NEXT: vmovss %xmm2, %xmm0, %xmm0 {%k1} [all …]
|
D | gpr-to-mask.ll | 16 ; X86-64-NEXT: vmovss %xmm0, %xmm1, %xmm1 {%k1} 17 ; X86-64-NEXT: vmovss %xmm1, (%rsi) 22 ; X86-32-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 23 ; X86-32-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 28 ; X86-32-NEXT: vmovss {{.*#+}} xmm2 = mem[0],zero,zero,zero 32 ; X86-32-NEXT: vmovss {{.*#+}} xmm2 = mem[0],zero,zero,zero 35 ; X86-32-NEXT: vmovss %xmm1, %xmm0, %xmm0 {%k1} 36 ; X86-32-NEXT: vmovss %xmm0, (%eax) 76 ; X86-32-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 81 ; X86-32-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero [all …]
|
D | pr46820.ll | 17 ; CHECK-NEXT: vmovss {{.*#+}} xmm3 = mem[0],zero,zero,zero 18 ; CHECK-NEXT: vmovss %xmm3, 88(%rdi) 38 ; CHECK-NEXT: vmovss {{.*#+}} xmm2 = mem[0],zero,zero,zero 39 ; CHECK-NEXT: vmovss %xmm2, 88(%rdi)
|
D | vector-constrained-fp-intrinsics-fma.ll | 7 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 8 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 41 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 42 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 44 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 45 ; CHECK-NEXT: vmovss {{.*#+}} xmm2 = mem[0],zero,zero,zero 47 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 48 ; CHECK-NEXT: vmovss {{.*#+}} xmm3 = mem[0],zero,zero,zero
|
D | fma-scalar-memfold.ll | 20 ; AVX2-NEXT: vmovss (%rdi), %xmm0 # encoding: [0xc5,0xfa,0x10,0x07] 24 ; AVX2-NEXT: vmovss %xmm0, (%rdi) # encoding: [0xc5,0xfa,0x11,0x07] 29 ; AVX512-NEXT: vmovss (%rdi), %xmm0 # EVEX TO VEX Compression encoding: [0xc5,0xfa,0x10,0x07] 33 ; AVX512-NEXT: vmovss %xmm0, (%rdi) # EVEX TO VEX Compression encoding: [0xc5,0xfa,0x11,0x07] 57 ; AVX2-NEXT: vmovss (%rdi), %xmm0 # encoding: [0xc5,0xfa,0x10,0x07] 61 ; AVX2-NEXT: vmovss %xmm0, (%rdi) # encoding: [0xc5,0xfa,0x11,0x07] 66 ; AVX512-NEXT: vmovss (%rdi), %xmm0 # EVEX TO VEX Compression encoding: [0xc5,0xfa,0x10,0x07] 70 ; AVX512-NEXT: vmovss %xmm0, (%rdi) # EVEX TO VEX Compression encoding: [0xc5,0xfa,0x11,0x07] 94 ; AVX2-NEXT: vmovss (%rdi), %xmm0 # encoding: [0xc5,0xfa,0x10,0x07] 98 ; AVX2-NEXT: vmovss %xmm0, (%rdi) # encoding: [0xc5,0xfa,0x11,0x07] [all …]
|
D | pr38803.ll | 16 ; CHECK-NEXT: vmovss {{.*}}(%rip), %xmm0 {%k1} {z} 23 ; CHECK-NEXT: vmovss %xmm2, %xmm1, %xmm1 {%k1} 24 ; CHECK-NEXT: vmovss %xmm1, {{.*}}(%rip)
|
D | avx512-load-store.ll | 11 ; CHECK64-NEXT: vmovss %xmm2, %xmm1, %xmm0 {%k1} 18 ; CHECK32-NEXT: vmovss %xmm2, %xmm1, %xmm0 {%k1} 34 ; CHECK64-NEXT: vmovss %xmm1, %xmm0, %xmm0 {%k1} {z} 41 ; CHECK32-NEXT: vmovss %xmm1, %xmm0, %xmm0 {%k1} {z} 101 ; CHECK64-NEXT: vmovss %xmm0, (%rdi) {%k1} 109 ; CHECK32-NEXT: vmovss %xmm0, (%eax) {%k1} 148 ; CHECK64-NEXT: vmovss (%rsi), %xmm0 {%k1} 156 ; CHECK32-NEXT: vmovss (%eax), %xmm0 {%k1} 199 ; CHECK64-NEXT: vmovss (%rsi), %xmm0 {%k1} {z} 207 ; CHECK32-NEXT: vmovss (%eax), %xmm0 {%k1} {z} [all …]
|
D | vector-sqrt.ll | 32 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 33 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 37 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 40 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero
|
D | intersect-fma-fmf.ll | 8 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 10 ; CHECK-NEXT: vmovss %xmm0, (%esp)
|
D | function-subtarget-features.ll | 5 ; In this case avx has a vmovss instruction and otherwise we should be using movss 15 ; CHECK: vmovss 39 ; CHECK: vmovss
|
D | pr45833.ll | 22 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 60 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 64 ; CHECK-NEXT: vmovss {{.*#+}} xmm2 = mem[0],zero,zero,zero 115 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 119 ; CHECK-NEXT: vmovss {{.*#+}} xmm2 = mem[0],zero,zero,zero 173 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 177 ; CHECK-NEXT: vmovss {{.*#+}} xmm2 = mem[0],zero,zero,zero 182 ; CHECK-NEXT: vmovss {{.*#+}} xmm2 = mem[0],zero,zero,zero 232 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 236 ; CHECK-NEXT: vmovss {{.*#+}} xmm2 = mem[0],zero,zero,zero [all …]
|
D | fp-strict-scalar.ll | 98 ; AVX-X86-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 100 ; AVX-X86-NEXT: vmovss %xmm0, (%esp) 196 ; AVX-X86-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 198 ; AVX-X86-NEXT: vmovss %xmm0, (%esp) 294 ; AVX-X86-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 296 ; AVX-X86-NEXT: vmovss %xmm0, (%esp) 392 ; AVX-X86-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 394 ; AVX-X86-NEXT: vmovss %xmm0, (%esp) 438 ; AVX-X86-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 445 ; AVX-X64-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero [all …]
|
D | pr45563-2.ll | 23 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 48 ; CHECK-NEXT: vmovss %xmm1, 32(%rax) 67 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 111 ; CHECK-NEXT: vmovss %xmm0, 48(%rax) 130 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 134 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 196 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 200 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero 243 ; CHECK-NEXT: vmovss %xmm0, 64(%rax) 263 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero [all …]
|
D | fma4-scalar-memfold.ll | 12 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 14 ; CHECK-NEXT: vmovss %xmm0, (%rdi) 38 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = mem[0],zero,zero,zero 40 ; CHECK-NEXT: vmovss %xmm0, (%rdi)
|
D | fast-isel-select-pseudo-cmov.ll | 30 ; AVX512-NEXT: vmovss %xmm2, %xmm3, %xmm3 {%k1} 90 ; AVX512-ISEL-NEXT: vmovss %xmm0, %xmm1, %xmm1 {%k1} 131 ; AVX512-ISEL-NEXT: vmovss %xmm0, %xmm1, %xmm1 {%k1} 172 ; AVX512-ISEL-NEXT: vmovss %xmm0, %xmm1, %xmm1 {%k1} 213 ; AVX512-ISEL-NEXT: vmovss %xmm0, %xmm1, %xmm1 {%k1} 254 ; AVX512-ISEL-NEXT: vmovss %xmm0, %xmm1, %xmm1 {%k1} 295 ; AVX512-ISEL-NEXT: vmovss %xmm0, %xmm1, %xmm1 {%k1} 336 ; AVX512-ISEL-NEXT: vmovss %xmm0, %xmm1, %xmm1 {%k1} 377 ; AVX512-ISEL-NEXT: vmovss %xmm0, %xmm1, %xmm1 {%k1} 418 ; AVX512-ISEL-NEXT: vmovss %xmm0, %xmm1, %xmm1 {%k1} [all …]
|
D | extractelement-fp.ll | 17 ; X86-NEXT: vmovss %xmm0, (%esp) 65 ; X86-NEXT: vmovss %xmm0, (%esp) 109 ; X86-NEXT: vmovss %xmm0, (%esp) 153 ; X86-NEXT: vmovss %xmm0, (%esp) 197 ; X86-NEXT: vmovss %xmm0, (%esp) 239 ; X86-NEXT: vmovss %xmm1, {{[0-9]+}}(%esp) 240 ; X86-NEXT: vmovss %xmm0, (%esp) 359 ; X86-NEXT: vmovss %xmm0, {{[0-9]+}}(%esp) 409 ; X86-NEXT: vmovss %xmm0, (%esp) 451 ; X86-NEXT: vmovss %xmm0, (%esp) [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | fma-scalar-memfold.ll | 17 ; CHECK: vmovss (%rcx), %[[XMM:xmm[0-9]+]] 19 ; CHECK-NEXT: vmovss %[[XMM]], (%rcx) 42 ; CHECK: vmovss (%rcx), %[[XMM:xmm[0-9]+]] 44 ; CHECK-NEXT: vmovss %[[XMM]], (%rcx) 67 ; CHECK: vmovss (%rcx), %[[XMM:xmm[0-9]+]] 69 ; CHECK-NEXT: vmovss %[[XMM]], (%rcx) 92 ; CHECK: vmovss (%rcx), %[[XMM:xmm[0-9]+]] 94 ; CHECK-NEXT: vmovss %[[XMM]], (%rcx) 117 ; CHECK: vmovss (%rcx), %[[XMM:xmm[0-9]+]] 119 ; CHECK-NEXT: vmovss %[[XMM]], (%rcx) [all …]
|
D | function-subtarget-features.ll | 5 ; In this case avx has a vmovss instruction and otherwise we should be using movss 15 ; CHECK: vmovss 39 ; CHECK: vmovss
|
D | lower-vec-shift.ll | 21 ; AVX-NEXT: vmovss 24 ; AVX2-NEXT: vmovss 55 ; AVX-NEXT: vmovss 107 ; AVX-NEXT: vmovss
|
D | avx512-build-vector.ll | 19 ; CHECK-NEXT: vmovss {{.*#+}} xmm0 = xmm0[0],xmm2[1,2,3] 20 ; CHECK-NEXT: vmovss {{.*#+}} xmm1 = xmm1[0],xmm2[1,2,3]
|