/external/llvm/test/CodeGen/X86/ |
D | avx512-bugfix-26264.ll | 9 ; AVX512BW-NEXT: vmovupd (%rdi), %zmm1 {%k1} 11 ; AVX512BW-NEXT: vmovupd 128(%rdi), %zmm3 {%k2} 13 ; AVX512BW-NEXT: vmovupd 64(%rdi), %zmm2 {%k1} 15 ; AVX512BW-NEXT: vmovupd 192(%rdi), %zmm4 {%k1}
|
D | chain_order.ll | 8 ; CHECK: vmovupd %xmm{{.*}}, (%rdi) 9 ; CHECK-NEXT: vmovupd %xmm{{.*}}, 16(%rdi)
|
D | vec_fptrunc.ll | 60 ; X32-AVX-NEXT: vmovupd %xmm0, (%eax) 74 ; X64-AVX-NEXT: vmovupd %xmm0, (%rsi) 105 ; X32-AVX-NEXT: vmovupd %ymm0, (%eax) 126 ; X64-AVX-NEXT: vmovupd %ymm0, (%rsi)
|
D | vector-half-conversions.ll | 2398 ; AVX1-NEXT: vmovupd %ymm0, (%rsp) # 32-byte Spill 2410 ; AVX1-NEXT: vmovupd (%rsp), %ymm0 # 32-byte Reload 2445 ; AVX2-NEXT: vmovupd %ymm0, (%rsp) # 32-byte Spill 2457 ; AVX2-NEXT: vmovupd (%rsp), %ymm0 # 32-byte Reload 2492 ; AVX512-NEXT: vmovupd %ymm0, (%rsp) # 32-byte Spill 2502 ; AVX512-NEXT: vmovupd (%rsp), %ymm0 # 32-byte Reload 2541 ; AVX1-NEXT: vmovupd %ymm0, (%rsp) # 32-byte Spill 2553 ; AVX1-NEXT: vmovupd (%rsp), %ymm0 # 32-byte Reload 2589 ; AVX2-NEXT: vmovupd %ymm0, (%rsp) # 32-byte Spill 2601 ; AVX2-NEXT: vmovupd (%rsp), %ymm0 # 32-byte Reload [all …]
|
D | vec_extract-avx.ll | 147 ; X32-NEXT: vmovupd (%ecx), %xmm0 156 ; X64-NEXT: vmovupd (%rdi), %xmm0 199 ; X32-NEXT: vmovupd (%ecx), %xmm0 208 ; X64-NEXT: vmovupd (%rdi), %xmm0
|
D | merge-consecutive-loads-512.ll | 11 ; ALL-NEXT: vmovupd 16(%rdi), %ymm0 19 ; X32-AVX512F-NEXT: vmovupd 16(%eax), %ymm0 38 ; ALL-NEXT: vmovupd 32(%rdi), %ymm0 47 ; X32-AVX512F-NEXT: vmovupd 32(%eax), %ymm0 86 ; ALL-NEXT: vmovupd 16(%rdi), %zmm0 92 ; X32-AVX512F-NEXT: vmovupd 16(%eax), %zmm0 109 ; ALL-NEXT: vmovupd 8(%rdi), %xmm0 119 ; X32-AVX512F-NEXT: vmovupd 8(%eax), %xmm0 141 ; ALL-NEXT: vmovupd 8(%rdi), %zmm0 150 ; X32-AVX512F-NEXT: vmovupd 8(%eax), %zmm0
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/ |
D | chain_order.ll | 13 ; CHECK-NEXT: vmovupd (%rdi), %xmm1 14 ; CHECK-NEXT: vmovupd %xmm0, (%rdi) 16 ; CHECK-NEXT: vmovupd %xmm0, 16(%rdi)
|
D | pr34653.ll | 18 ; CHECK-NEXT: vmovupd {{[0-9]+}}(%rsp), %xmm0 21 ; CHECK-NEXT: vmovupd {{[0-9]+}}(%rsp), %ymm2 26 ; CHECK-NEXT: vmovupd {{[0-9]+}}(%rsp), %zmm7 27 ; CHECK-NEXT: vmovupd {{[0-9]+}}(%rsp), %zmm8 28 ; CHECK-NEXT: vmovupd {{[0-9]+}}(%rsp), %zmm9 29 ; CHECK-NEXT: vmovupd {{[0-9]+}}(%rsp), %zmm10
|
D | inline-asm-avx512f-v-constraint.ll | 68 ; CHECK: vmovupd %zmm1, %zmm0 69 …%0 = tail call <16 x float> asm "vmovupd $1, $0", "=v,v,~{dirflag},~{fpsr},~{flags}"(<16 x float> …
|
D | avx-intrinsics-x86_64.ll | 9 ; AVX-NEXT: vmovupd %ymm0, {{[-0-9]+}}(%r{{[sb]}}p) # 32-byte Spill 32 ; AVX-NEXT: vmovupd %ymm0, {{[-0-9]+}}(%r{{[sb]}}p) # 32-byte Spill
|
D | fast-isel-store.ll | 178 ; AVX32-NEXT: vmovupd %xmm0, (%rdi) 185 ; AVX64-NEXT: vmovupd %xmm0, (%eax) 360 ; AVX32-NEXT: vmovupd %ymm0, (%rdi) 367 ; AVX64-NEXT: vmovupd %ymm0, (%eax) 648 ; AVXONLY32-NEXT: vmovupd %ymm0, (%rdi) 649 ; AVXONLY32-NEXT: vmovupd %ymm1, 32(%rdi) 664 ; AVXONLY64-NEXT: vmovupd %ymm0, (%eax) 665 ; AVXONLY64-NEXT: vmovupd %ymm1, 32(%eax) 674 ; AVX51232-NEXT: vmovupd %zmm0, (%rdi) 681 ; AVX51264-NEXT: vmovupd %zmm0, (%eax)
|
D | vector-half-conversions.ll | 3048 ; AVX1-NEXT: vmovupd %ymm0, (%rsp) # 32-byte Spill 3085 ; AVX2-NEXT: vmovupd %ymm0, (%rsp) # 32-byte Spill 3097 ; AVX2-NEXT: vmovupd (%rsp), %ymm0 # 32-byte Reload 3122 ; AVX512-NEXT: vmovupd %ymm0, (%rsp) # 32-byte Spill 3134 ; AVX512-NEXT: vmovupd (%rsp), %ymm0 # 32-byte Reload 3164 ; AVX1-NEXT: vmovupd %ymm0, (%rsp) # 32-byte Spill 3202 ; AVX2-NEXT: vmovupd %ymm0, (%rsp) # 32-byte Spill 3214 ; AVX2-NEXT: vmovupd (%rsp), %ymm0 # 32-byte Reload 3240 ; AVX512-NEXT: vmovupd %ymm0, (%rsp) # 32-byte Spill 3252 ; AVX512-NEXT: vmovupd (%rsp), %ymm0 # 32-byte Reload [all …]
|
D | avx512-masked-memop-64-32.ll | 166 ; AVX512F-NEXT: vmovupd %zmm1, (%rdi) {%k1} 168 ; AVX512F-NEXT: vmovupd %zmm2, 64(%rdi) {%k1} 176 ; SKX-NEXT: vmovupd %zmm1, (%rdi) {%k1} 178 ; SKX-NEXT: vmovupd %zmm2, 64(%rdi) {%k1}
|
D | inline-asm-avx-v-constraint.ll | 131 ; CHECK: vmovupd %ymm1, %ymm0 133 …%0 = tail call <8 x float> asm "vmovupd $1, $0", "=v,v,~{dirflag},~{fpsr},~{flags}"(<8 x float> %_…
|
D | inline-asm-avx512vl-v-constraint.ll | 117 ; CHECK: vmovupd %ymm16, %ymm16 118 …%0 = tail call <8 x float> asm "vmovupd $1, $0", "=v,v,~{ymm0},~{ymm1},~{ymm2},~{ymm3},~{ymm4},~{y…
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/MC/X86/ |
D | AVX-32.s | 3949 vmovupd -485498096(%edx,%eax,4), %xmm1 label 3953 vmovupd 485498096(%edx,%eax,4), %xmm1 label 3957 vmovupd -485498096(%edx,%eax,4), %ymm4 label 3961 vmovupd 485498096(%edx,%eax,4), %ymm4 label 3965 vmovupd 485498096(%edx), %xmm1 label 3969 vmovupd 485498096(%edx), %ymm4 label 3973 vmovupd 485498096, %xmm1 label 3977 vmovupd 485498096, %ymm4 label 3981 vmovupd 64(%edx,%eax), %xmm1 label 3985 vmovupd 64(%edx,%eax), %ymm4 label [all …]
|
D | AVX-64.s | 8325 vmovupd 485498096, %xmm15 label 8329 vmovupd 485498096, %xmm6 label 8333 vmovupd 485498096, %ymm7 label 8337 vmovupd 485498096, %ymm9 label 8341 vmovupd -64(%rdx,%rax,4), %xmm15 label 8345 vmovupd 64(%rdx,%rax,4), %xmm15 label 8349 vmovupd -64(%rdx,%rax,4), %xmm6 label 8353 vmovupd 64(%rdx,%rax,4), %xmm6 label 8357 vmovupd -64(%rdx,%rax,4), %ymm7 label 8361 vmovupd 64(%rdx,%rax,4), %ymm7 label [all …]
|
D | x86-64-avx512f_vl.s | 1657 vmovupd %xmm22, %xmm24 1661 vmovupd %xmm22, %xmm24 {%k6} 1665 vmovupd %xmm22, %xmm24 {%k6} {z} 1669 vmovupd (%rcx), %xmm24 1673 vmovupd 291(%rax,%r14,8), %xmm24 1677 vmovupd 2032(%rdx), %xmm24 1681 vmovupd 2048(%rdx), %xmm24 1685 vmovupd -2048(%rdx), %xmm24 1689 vmovupd -2064(%rdx), %xmm24 1693 vmovupd %ymm25, %ymm30 [all …]
|
D | x86-32-avx.s | 736 vmovupd (%eax), %xmm2 740 vmovupd %xmm1, %xmm2 744 vmovupd %xmm1, (%eax) 2332 vmovupd (%eax), %ymm2 2336 vmovupd %ymm1, %ymm2 2340 vmovupd %ymm1, (%eax)
|
/external/llvm/test/MC/X86/ |
D | x86-64-avx512f_vl.s | 1657 vmovupd %xmm22, %xmm24 1661 vmovupd %xmm22, %xmm24 {%k6} 1665 vmovupd %xmm22, %xmm24 {%k6} {z} 1669 vmovupd (%rcx), %xmm24 1673 vmovupd 291(%rax,%r14,8), %xmm24 1677 vmovupd 2032(%rdx), %xmm24 1681 vmovupd 2048(%rdx), %xmm24 1685 vmovupd -2048(%rdx), %xmm24 1689 vmovupd -2064(%rdx), %xmm24 1693 vmovupd %ymm25, %ymm30 [all …]
|
D | x86-32-avx.s | 736 vmovupd (%eax), %xmm2 740 vmovupd %xmm1, %xmm2 744 vmovupd %xmm1, (%eax) 2332 vmovupd (%eax), %ymm2 2336 vmovupd %ymm1, %ymm2 2340 vmovupd %ymm1, (%eax)
|
/external/swiftshader/third_party/LLVM/test/MC/X86/ |
D | x86-32-avx.s | 720 vmovupd (%eax), %xmm2 724 vmovupd %xmm1, %xmm2 728 vmovupd %xmm1, (%eax) 2308 vmovupd (%eax), %ymm2 2312 vmovupd %ymm1, %ymm2 2316 vmovupd %ymm1, (%eax)
|
D | x86_64-avx-encoding.s | 697 vmovupd (%rax), %xmm12 701 vmovupd %xmm11, %xmm12 705 vmovupd %xmm11, (%rax) 2305 vmovupd (%rax), %ymm12 2309 vmovupd %ymm11, %ymm12 2313 vmovupd %ymm11, (%rax)
|
/external/capstone/suite/MC/X86/ |
D | x86-32-avx.s.cs | 185 0xc5,0xf9,0x10,0x10 = vmovupd (%eax), %xmm2 186 0xc5,0xf9,0x10,0xd1 = vmovupd %xmm1, %xmm2 187 0xc5,0xf9,0x11,0x08 = vmovupd %xmm1, (%eax) 584 0xc5,0xfd,0x10,0x10 = vmovupd (%eax), %ymm2 585 0xc5,0xfd,0x10,0xd1 = vmovupd %ymm1, %ymm2 586 0xc5,0xfd,0x11,0x08 = vmovupd %ymm1, (%eax)
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/tools/llvm-mca/X86/SkylakeServer/ |
D | resources-avx1.s | 438 vmovupd %xmm0, %xmm2 label 439 vmovupd %xmm0, (%rax) label 440 vmovupd (%rax), %xmm2 label 442 vmovupd %ymm0, %ymm2 label 443 vmovupd %ymm0, (%rax) label 444 vmovupd (%rax), %ymm2 label 1316 # CHECK-NEXT: 1 1 0.33 vmovupd %xmm0, %xmm2 1317 # CHECK-NEXT: 2 1 1.00 * vmovupd %xmm0, (%rax) 1318 # CHECK-NEXT: 1 6 0.50 * vmovupd (%rax), %xmm2 1319 # CHECK-NEXT: 1 1 0.33 vmovupd %ymm0, %ymm2 [all …]
|