/external/swiftshader/third_party/llvm-7.0/llvm/test/tools/llvm-mca/X86/BtVer2/ |
D | memcpy-like-test.s | 4 vmovaps (%rsi), %xmm0 label 5 vmovaps %xmm0, (%rdi) label 6 vmovaps 16(%rsi), %xmm0 label 7 vmovaps %xmm0, 16(%rdi) label 8 vmovaps 32(%rsi), %xmm0 label 9 vmovaps %xmm0, 32(%rdi) label 10 vmovaps 48(%rsi), %xmm0 label 11 vmovaps %xmm0, 48(%rdi) label 29 # CHECK-NEXT: 1 5 1.00 * vmovaps (%rsi), %xmm0 30 # CHECK-NEXT: 1 1 1.00 * vmovaps %xmm0, (%rdi) [all …]
|
D | load-store-alias.s | 4 vmovaps (%rsi), %xmm0 label 5 vmovaps %xmm0, (%rdi) label 6 vmovaps 16(%rsi), %xmm0 label 7 vmovaps %xmm0, 16(%rdi) label 8 vmovaps 32(%rsi), %xmm0 label 9 vmovaps %xmm0, 32(%rdi) label 10 vmovaps 48(%rsi), %xmm0 label 11 vmovaps %xmm0, 48(%rdi) label 29 # CHECK-NEXT: 1 5 1.00 * vmovaps (%rsi), %xmm0 30 # CHECK-NEXT: 1 1 1.00 * vmovaps %xmm0, (%rdi) [all …]
|
/external/swiftshader/third_party/llvm-7.0/llvm/test/CodeGen/X86/ |
D | var-permute-512.ll | 103 ; NOBW-NEXT: vmovaps %ymm1, {{[0-9]+}}(%rsp) 104 ; NOBW-NEXT: vmovaps %ymm0, {{[0-9]+}}(%rsp) 105 ; NOBW-NEXT: vmovaps %ymm1, {{[0-9]+}}(%rsp) 106 ; NOBW-NEXT: vmovaps %ymm0, {{[0-9]+}}(%rsp) 107 ; NOBW-NEXT: vmovaps %ymm1, {{[0-9]+}}(%rsp) 108 ; NOBW-NEXT: vmovaps %ymm0, {{[0-9]+}}(%rsp) 109 ; NOBW-NEXT: vmovaps %ymm1, {{[0-9]+}}(%rsp) 110 ; NOBW-NEXT: vmovaps %ymm0, {{[0-9]+}}(%rsp) 111 ; NOBW-NEXT: vmovaps %ymm1, {{[0-9]+}}(%rsp) 112 ; NOBW-NEXT: vmovaps %ymm0, {{[0-9]+}}(%rsp) [all …]
|
D | pr34592.ll | 14 ; CHECK-NEXT: vmovaps 240(%rbp), %ymm8 15 ; CHECK-NEXT: vmovaps 208(%rbp), %ymm9 16 ; CHECK-NEXT: vmovaps 176(%rbp), %ymm10 17 ; CHECK-NEXT: vmovaps 144(%rbp), %ymm11 18 ; CHECK-NEXT: vmovaps 112(%rbp), %ymm12 19 ; CHECK-NEXT: vmovaps 80(%rbp), %ymm13 20 ; CHECK-NEXT: vmovaps 48(%rbp), %ymm14 21 ; CHECK-NEXT: vmovaps 16(%rbp), %ymm15 31 ; CHECK-NEXT: vmovaps %ymm0, {{[0-9]+}}(%rsp) # 32-byte Spill 35 ; CHECK-NEXT: vmovaps %xmm2, %xmm9 [all …]
|
D | avx512-intel-ocl.ll | 54 ; WIN64-NEXT: vmovaps (%rcx), %zmm0 97 ; X32-NEXT: vmovaps %zmm1, {{[0-9]+}}(%esp) ## 64-byte Spill 114 ; WIN32-NEXT: vmovaps %zmm1, (%esp) # 64-byte Spill 132 ; WIN64-NEXT: vmovaps (%rdx), %zmm16 150 ; X64-NEXT: vmovaps %zmm1, %zmm16 193 ; WIN64-KNL-NEXT: vmovaps %zmm21, 1104(%rbp) # 64-byte Spill 194 ; WIN64-KNL-NEXT: vmovaps %zmm20, 992(%rbp) # 64-byte Spill 195 ; WIN64-KNL-NEXT: vmovaps %zmm19, 896(%rbp) # 64-byte Spill 196 ; WIN64-KNL-NEXT: vmovaps %zmm18, 832(%rbp) # 64-byte Spill 197 ; WIN64-KNL-NEXT: vmovaps %zmm17, 768(%rbp) # 64-byte Spill [all …]
|
D | fold-vector-sext-zext.ll | 14 ; X32-NEXT: vmovaps {{.*#+}} xmm0 = [0,4294967295,2,4294967293] 19 ; X64-NEXT: vmovaps {{.*#+}} xmm0 = [0,4294967295,2,4294967293] 32 ; X32-NEXT: vmovaps {{.*#+}} xmm0 = <u,4294967295,u,4294967293> 37 ; X64-NEXT: vmovaps {{.*#+}} xmm0 = <u,4294967295,u,4294967293> 50 ; X32-NEXT: vmovaps {{.*#+}} xmm0 = [0,4294967295,2,4294967293] 55 ; X64-NEXT: vmovaps {{.*#+}} xmm0 = [0,4294967295,2,4294967293] 68 ; X32-NEXT: vmovaps {{.*#+}} xmm0 = <u,4294967295,u,4294967293> 73 ; X64-NEXT: vmovaps {{.*#+}} xmm0 = <u,4294967295,u,4294967293> 86 ; X32-NEXT: vmovaps {{.*#+}} ymm0 = [0,0,4294967295,4294967295,2,0,4294967293,4294967295] 91 ; X64-NEXT: vmovaps {{.*#+}} ymm0 = [0,18446744073709551615,2,18446744073709551613] [all …]
|
D | pr34653.ll | 19 ; CHECK-NEXT: vmovaps %xmm0, %xmm1 23 ; CHECK-NEXT: vmovaps %xmm3, %xmm4 24 ; CHECK-NEXT: vmovaps %xmm2, %xmm5 25 ; CHECK-NEXT: vmovaps %xmm5, %xmm6 31 ; CHECK-NEXT: vmovaps %xmm11, %xmm12 33 ; CHECK-NEXT: vmovaps %xmm13, %xmm14 34 ; CHECK-NEXT: vmovaps %xmm10, %xmm15 35 ; CHECK-NEXT: vmovaps %xmm15, %xmm2 38 ; CHECK-NEXT: vmovaps %xmm0, {{[0-9]+}}(%rsp) # 16-byte Spill 41 ; CHECK-NEXT: vmovaps %xmm0, {{[0-9]+}}(%rsp) # 16-byte Spill [all …]
|
D | avx-vbroadcastf128.ll | 106 ; X32-NEXT: vmovaps (%ecx), %xmm1 108 ; X32-NEXT: vmovaps %xmm1, (%eax) 113 ; X64-NEXT: vmovaps (%rdi), %xmm1 115 ; X64-NEXT: vmovaps %xmm1, (%rsi) 128 ; X32-NEXT: vmovaps (%ecx), %xmm1 130 ; X32-NEXT: vmovaps %xmm1, (%eax) 135 ; X64-NEXT: vmovaps (%rdi), %xmm1 137 ; X64-NEXT: vmovaps %xmm1, (%rsi) 150 ; X32-NEXT: vmovaps (%ecx), %xmm1 152 ; X32-NEXT: vmovaps %xmm1, (%eax) [all …]
|
D | avx512-regcall-Mask.ll | 101 ; X32-NEXT: vmovaps {{.*#+}} xmm0 = [2,1,2,1] 103 ; X32-NEXT: vmovaps {{.*#+}} zmm0 = [2,1,2,1,2,1,2,1,2,1,2,1,2,1,2,1] 132 ; WIN64-NEXT: vmovaps %xmm7, {{[0-9]+}}(%rsp) # 16-byte Spill 134 ; WIN64-NEXT: vmovaps %xmm6, {{[0-9]+}}(%rsp) # 16-byte Spill 151 ; WIN64-NEXT: vmovaps {{[0-9]+}}(%rsp), %xmm6 # 16-byte Reload 152 ; WIN64-NEXT: vmovaps {{[0-9]+}}(%rsp), %xmm7 # 16-byte Reload 252 ; WIN64-NEXT: vmovaps %xmm7, {{[0-9]+}}(%rsp) # 16-byte Spill 254 ; WIN64-NEXT: vmovaps %xmm6, (%rsp) # 16-byte Spill 260 ; WIN64-NEXT: vmovaps (%rsp), %xmm6 # 16-byte Reload 261 ; WIN64-NEXT: vmovaps {{[0-9]+}}(%rsp), %xmm7 # 16-byte Reload [all …]
|
D | avx-intel-ocl.ll | 70 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 71 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 72 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 73 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 74 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 75 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 76 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 77 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 78 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 79 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill [all …]
|
D | avx-load-store.ll | 15 ; CHECK-NEXT: vmovaps (%rdi), %ymm0 17 ; CHECK-NEXT: vmovaps (%rsi), %ymm1 19 ; CHECK-NEXT: vmovaps (%rdx), %ymm2 23 ; CHECK-NEXT: vmovaps %ymm0, (%rbx) 25 ; CHECK-NEXT: vmovaps %ymm0, (%r15) 27 ; CHECK-NEXT: vmovaps %ymm0, (%r14) 39 ; CHECK_O0-NEXT: vmovaps (%rsi), %ymm1 53 ; CHECK_O0-NEXT: vmovaps %ymm1, (%rsi) 89 ; CHECK_O0-NEXT: vmovaps %xmm0, %xmm1 90 ; CHECK_O0-NEXT: vmovaps %xmm1, %xmm0 [all …]
|
D | subvector-broadcast.ll | 38 ; X32-AVX-NEXT: vmovaps %ymm0, %ymm1 50 ; X64-AVX-NEXT: vmovaps %ymm0, %ymm1 66 ; X32-AVX-NEXT: vmovaps (%eax), %ymm0 67 ; X32-AVX-NEXT: vmovaps %ymm0, %ymm1 78 ; X64-AVX-NEXT: vmovaps (%rdi), %ymm0 79 ; X64-AVX-NEXT: vmovaps %ymm0, %ymm1 123 ; X32-AVX-NEXT: vmovaps %ymm0, %ymm1 135 ; X64-AVX-NEXT: vmovaps %ymm0, %ymm1 151 ; X32-AVX-NEXT: vmovaps (%eax), %ymm0 152 ; X32-AVX-NEXT: vmovaps %ymm0, %ymm1 [all …]
|
D | sandybridge-loads.ll | 7 ; CHECK-NEXT: vmovaps (%rdi), %xmm0 9 ; CHECK-NEXT: vmovaps (%rsi), %ymm1 11 ; CHECK-NEXT: vmovaps (%rdx), %ymm2 15 ; CHECK-NEXT: vmovaps %ymm0, (%rax) 32 ; CHECK-NEXT: vmovaps (%rdi), %ymm0 33 ; CHECK-NEXT: vmovaps (%rsi), %ymm1 34 ; CHECK-NEXT: vmovaps %ymm0, (%rsi) 36 ; CHECK-NEXT: vmovaps %xmm1, (%rdi)
|
D | vaargs.ll | 12 ; CHECK-NEXT: vmovaps %xmm0, 48(%rsp) 13 ; CHECK-NEXT: vmovaps %xmm1, 64(%rsp) 14 ; CHECK-NEXT: vmovaps %xmm2, 80(%rsp) 15 ; CHECK-NEXT: vmovaps %xmm3, 96(%rsp) 16 ; CHECK-NEXT: vmovaps %xmm4, 112(%rsp) 17 ; CHECK-NEXT: vmovaps %xmm5, 128(%rsp) 18 ; CHECK-NEXT: vmovaps %xmm6, 144(%rsp) 19 ; CHECK-NEXT: vmovaps %xmm7, 160(%rsp)
|
D | widen_load-3.ll | 32 ; X86-AVX-NEXT: vmovaps (%ecx), %ymm0 33 ; X86-AVX-NEXT: vmovaps 32(%ecx), %ymm1 34 ; X86-AVX-NEXT: vmovaps %ymm0, (%eax) 38 ; X86-AVX-NEXT: vmovaps %xmm1, 32(%eax) 57 ; X64-AVX-NEXT: vmovaps (%rsi), %ymm0 58 ; X64-AVX-NEXT: vmovaps 32(%rsi), %ymm1 59 ; X64-AVX-NEXT: vmovaps %ymm0, (%rdi) 62 ; X64-AVX-NEXT: vmovaps %xmm1, 32(%rdi) 97 ; X86-AVX-NEXT: vmovaps %xmm1, 32(%eax) 98 ; X86-AVX-NEXT: vmovaps %ymm0, (%eax) [all …]
|
D | vector-extend-inreg.ll | 78 ; X32-AVX-NEXT: vmovaps %ymm1, {{[0-9]+}}(%esp) 79 ; X32-AVX-NEXT: vmovaps %ymm1, {{[0-9]+}}(%esp) 80 ; X32-AVX-NEXT: vmovaps %ymm1, {{[0-9]+}}(%esp) 81 ; X32-AVX-NEXT: vmovaps %ymm0, {{[0-9]+}}(%esp) 82 ; X32-AVX-NEXT: vmovaps %ymm1, {{[0-9]+}}(%esp) 83 ; X32-AVX-NEXT: vmovaps %ymm1, {{[0-9]+}}(%esp) 84 ; X32-AVX-NEXT: vmovaps %ymm1, (%esp) 85 ; X32-AVX-NEXT: vmovaps %ymm0, {{[0-9]+}}(%esp) 108 ; X64-AVX-NEXT: vmovaps %ymm1, {{[0-9]+}}(%rsp) 109 ; X64-AVX-NEXT: vmovaps %ymm1, {{[0-9]+}}(%rsp) [all …]
|
D | avx-vzeroupper.ll | 36 ; VZ-NEXT: vmovaps {{.*}}(%rip), %xmm0 39 ; VZ-NEXT: vmovaps %xmm0, {{.*}}(%rip) 41 ; VZ-NEXT: vmovaps %xmm0, {{.*}}(%rip) 50 ; FAST-ymm-zmm-NEXT: vmovaps {{.*}}(%rip), %xmm0 52 ; FAST-ymm-zmm-NEXT: vmovaps %xmm0, {{.*}}(%rip) 54 ; FAST-ymm-zmm-NEXT: vmovaps %xmm0, {{.*}}(%rip) 62 ; BTVER2-NEXT: vmovaps {{.*}}(%rip), %xmm0 65 ; BTVER2-NEXT: vmovaps %xmm0, {{.*}}(%rip) 67 ; BTVER2-NEXT: vmovaps %xmm0, {{.*}}(%rip) 109 ; VZ-NEXT: vmovaps %xmm0, (%rsp) # 16-byte Spill [all …]
|
/external/llvm/test/CodeGen/X86/ |
D | avx-load-store.ll | 4 ; CHECK: vmovaps 5 ; CHECK: vmovaps 6 ; CHECK: vmovaps 7 ; CHECK: vmovaps 8 ; CHECK: vmovaps 9 ; CHECK: vmovaps 48 ; CHECK: vmovaps %ymm 63 ; CHECK: vmovaps %ymm 82 ; CHECK: vmovaps %xmm 83 ; CHECK: vmovaps %xmm [all …]
|
D | avx-intel-ocl.ll | 70 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 71 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 72 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 73 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 74 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 75 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 76 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 77 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 78 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill 79 ; WIN64: vmovaps {{%ymm([6-9]|1[0-5])}}, {{.*(%rbp).*}} # 32-byte Spill [all …]
|
D | vaargs.ll | 12 ; CHECK-NEXT: vmovaps %xmm0, 48(%rsp) 13 ; CHECK-NEXT: vmovaps %xmm1, 64(%rsp) 14 ; CHECK-NEXT: vmovaps %xmm2, 80(%rsp) 15 ; CHECK-NEXT: vmovaps %xmm3, 96(%rsp) 16 ; CHECK-NEXT: vmovaps %xmm4, 112(%rsp) 17 ; CHECK-NEXT: vmovaps %xmm5, 128(%rsp) 18 ; CHECK-NEXT: vmovaps %xmm6, 144(%rsp) 19 ; CHECK-NEXT: vmovaps %xmm7, 160(%rsp)
|
D | avx512-bugfix-26264.ll | 16 ; AVX512BW-NEXT: vmovaps %zmm1, %zmm0 17 ; AVX512BW-NEXT: vmovaps %zmm2, %zmm1 18 ; AVX512BW-NEXT: vmovaps %zmm3, %zmm2 19 ; AVX512BW-NEXT: vmovaps %zmm4, %zmm3 37 ; AVX512BW-NEXT: vmovaps %zmm1, %zmm0 38 ; AVX512BW-NEXT: vmovaps %zmm2, %zmm1 39 ; AVX512BW-NEXT: vmovaps %zmm3, %zmm2 40 ; AVX512BW-NEXT: vmovaps %zmm4, %zmm3
|
D | fold-vector-sext-zext.ll | 13 ; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = [0,4294967295,2,4294967293] 26 ; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = <u,4294967295,u,4294967293> 39 ; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = [0,4294967295,2,4294967293] 52 ; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = <u,4294967295,u,4294967293> 65 ; CHECK-NEXT: vmovaps {{.*#+}} ymm0 = [0,18446744073709551615,2,18446744073709551613] 78 ; CHECK-NEXT: vmovaps {{.*#+}} ymm0 = <u,18446744073709551615,u,18446744073709551613> 91 ; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = <0,65535,2,65533,u,u,u,u> 108 ; CHECK-NEXT: vmovaps {{.*#+}} ymm0 = <0,4294967295,2,4294967293,u,u,u,u> 125 ; CHECK-NEXT: vmovaps {{.*#+}} xmm0 = <u,65535,u,65533,u,u,u,u> 142 ; CHECK-NEXT: vmovaps {{.*#+}} ymm0 = <0,u,2,u,u,u,u,u> [all …]
|
D | sandybridge-loads.ll | 4 ;CHECK: vmovaps 6 ;CHECK: vmovaps 24 ; CHECK: vmovaps 25 ; CHECK: vmovaps 27 ; CHECK: vmovaps 29 ; CHECK: vmovaps
|
D | fma-intrinsics-x86.ll | 13 ; CHECK-FMA-WIN-NEXT: vmovaps {{\(%rcx\), %xmm0|\(%r8\), %xmm1}} 14 ; CHECK-FMA-WIN-NEXT: vmovaps {{\(%rcx\), %xmm0|\(%r8\), %xmm1}} 30 ; CHECK-FMA-WIN-NEXT: vmovaps {{\(%rdx\), %xmm0|\(%r8\), %xmm1}} 31 ; CHECK-FMA-WIN-NEXT: vmovaps {{\(%rdx\), %xmm0|\(%r8\), %xmm1}} 35 ; CHECK-FMA-NEXT: vmovaps %xmm1, %xmm0 84 ; CHECK-FMA-WIN-NEXT: vmovaps (%{{(rcx|rdx)}}), %xmm{{0|1}} 85 ; CHECK-FMA-WIN-NEXT: vmovaps (%{{(rcx|rdx)}}), %xmm{{0|1}} 120 ; CHECK-FMA-WIN-NEXT: vmovaps (%{{(rcx|rdx)}}), %ymm{{0|1}} 121 ; CHECK-FMA-WIN-NEXT: vmovaps (%{{(rcx|rdx)}}), %ymm{{0|1}} 157 ; CHECK-FMA-WIN-NEXT: vmovaps {{\(%rcx\), %xmm0|\(%r8\), %xmm1}} [all …]
|
/external/swiftshader/third_party/LLVM/test/CodeGen/X86/ |
D | avx-load-store.ll | 4 ; CHECK: vmovaps 5 ; CHECK: vmovaps 6 ; CHECK: vmovaps 7 ; CHECK: vmovaps 8 ; CHECK: vmovaps 9 ; CHECK: vmovaps 44 ; CHECK: vmovaps %ymm 56 ; CHECK: vmovaps %ymm 73 ; CHECK: vmovaps %xmm 74 ; CHECK: vmovaps %xmm
|