/external/llvm-project/llvm/test/tools/llvm-mca/X86/BtVer2/ |
D | vbroadcast-operand-latency.s | 5 vbroadcastss (%rax), %ymm0 label 27 # CHECK-NEXT: 2 6 2.00 * vbroadcastss (%rax), %ymm0 52 … 2.00 1.00 1.00 1.00 - - - - - - vbroadcastss (%rax), %ymm0 59 # CHECK-NEXT: [0,1] .D=eeeeeeER . vbroadcastss (%rax), %ymm0 61 # CHECK-NEXT: [1,1] . D=eeeeeeER . vbroadcastss (%rax), %ymm0 63 # CHECK-NEXT: [2,1] . D=eeeeeeER vbroadcastss (%rax), %ymm0 73 # CHECK-NEXT: 1. 3 2.0 0.0 0.0 vbroadcastss (%rax), %ymm0
|
/external/llvm-project/llvm/test/tools/llvm-mca/X86/BdVer2/ |
D | vbroadcast-operand-latency.s | 5 vbroadcastss (%rax), %ymm0 label 27 # CHECK-NEXT: 2 6 2.00 * vbroadcastss (%rax), %ymm0 61 … - 1.00 1.00 - - - 0.50 0.50 - - vbroadcastss (%rax), %ymm0 68 # CHECK-NEXT: [0,1] DeeeeeeER . . vbroadcastss (%rax), %ymm0 70 # CHECK-NEXT: [1,1] .DeeeeeeER. . vbroadcastss (%rax), %ymm0 72 # CHECK-NEXT: [2,1] . D==eeeeeeER vbroadcastss (%rax), %ymm0 82 # CHECK-NEXT: 1. 3 1.7 1.0 0.0 vbroadcastss (%rax), %ymm0
|
/external/llvm-project/llvm/test/CodeGen/X86/ |
D | v8i1-masks.ll | 103 ; X32-AVX2-NEXT: vbroadcastss {{.*#+}} ymm1 = [1,1,1,1,1,1,1,1] 113 ; X64-AVX2-NEXT: vbroadcastss {{.*#+}} ymm1 = [1,1,1,1,1,1,1,1] 186 ; X32-AVX2-NEXT: vbroadcastss {{.*#+}} ymm1 = [-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-… 188 ; X32-AVX2-NEXT: vbroadcastss {{.*#+}} ymm2 = [1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,… 195 ; X64-AVX2-NEXT: vbroadcastss {{.*#+}} ymm1 = [-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-… 197 ; X64-AVX2-NEXT: vbroadcastss {{.*#+}} ymm2 = [1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,… 234 ; X32-AVX2-NEXT: vbroadcastss {{.*#+}} ymm1 = [-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-… 236 ; X32-AVX2-NEXT: vbroadcastss {{.*#+}} ymm2 = [1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,… 246 ; X64-AVX2-NEXT: vbroadcastss {{.*#+}} ymm1 = [-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-… 248 ; X64-AVX2-NEXT: vbroadcastss {{.*#+}} ymm2 = [1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,… [all …]
|
D | pr32368.ll | 24 ; AVX2-NEXT: vbroadcastss {{.*#+}} xmm1 = [4294967004,4294967004,4294967004,4294967004] 27 ; AVX2-NEXT: vbroadcastss {{.*#+}} xmm1 = [291,291,291,291] 33 ; AVX512-NEXT: vbroadcastss {{.*#+}} xmm1 = [4294967004,4294967004,4294967004,4294967004] 36 ; AVX512-NEXT: vbroadcastss {{.*#+}} xmm1 = [291,291,291,291] 71 ; AVX2-NEXT: vbroadcastss {{.*#+}} ymm1 = [4294967004,4294967004,4294967004,4294967004,429496700… 74 ; AVX2-NEXT: vbroadcastss {{.*#+}} ymm1 = [291,291,291,291,291,291,291,291] 80 ; AVX512-NEXT: vbroadcastss {{.*#+}} ymm1 = [4294967004,4294967004,4294967004,4294967004,4294967… 83 ; AVX512-NEXT: vbroadcastss {{.*#+}} ymm1 = [291,291,291,291,291,291,291,291] 129 ; AVX2-NEXT: vbroadcastss {{.*#+}} ymm2 = [4294967004,4294967004,4294967004,4294967004,429496700… 134 ; AVX2-NEXT: vbroadcastss {{.*#+}} ymm2 = [291,291,291,291,291,291,291,291]
|
D | combine-fcopysign.ll | 19 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm1 = [NaN,NaN,NaN,NaN] 34 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm1 = [NaN,NaN,NaN,NaN] 49 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm1 = [NaN,NaN,NaN,NaN] 66 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm1 = [-0.0E+0,-0.0E+0,-0.0E+0,-0.0E+0] 81 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm1 = [-0.0E+0,-0.0E+0,-0.0E+0,-0.0E+0] 96 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm1 = [-0.0E+0,-0.0E+0,-0.0E+0,-0.0E+0] 116 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm2 = [-0.0E+0,-0.0E+0,-0.0E+0,-0.0E+0] 118 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm2 = [NaN,NaN,NaN,NaN] 138 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm2 = [-0.0E+0,-0.0E+0,-0.0E+0,-0.0E+0] 140 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm2 = [NaN,NaN,NaN,NaN] [all …]
|
D | avx-vbroadcast.ll | 63 ; X32-NEXT: vbroadcastss (%eax), %ymm0 68 ; X64-NEXT: vbroadcastss (%rdi), %ymm0 83 ; X32-NEXT: vbroadcastss (%eax), %ymm0 88 ; X64-NEXT: vbroadcastss (%rdi), %ymm0 185 ; X32-NEXT: vbroadcastss (%eax), %ymm0 190 ; X64-NEXT: vbroadcastss (%rdi), %ymm0 205 ; X32-NEXT: vbroadcastss (%eax), %ymm0 210 ; X64-NEXT: vbroadcastss (%rdi), %ymm0 230 ; X32-NEXT: vbroadcastss (%ecx), %ymm0 236 ; X64-NEXT: vbroadcastss (%rdi), %ymm0 [all …]
|
D | combine-fabs.ll | 48 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm1 = [NaN,NaN,NaN,NaN] 64 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm1 = [NaN,NaN,NaN,NaN] 81 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm1 = [NaN,NaN,NaN,NaN] 97 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm1 = [NaN,NaN,NaN,NaN] 114 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm1 = [NaN,NaN,NaN,NaN] 130 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm1 = [NaN,NaN,NaN,NaN]
|
D | sqrt-fastmath-tune.ll | 67 ; BDW-NEXT: vbroadcastss {{.*#+}} xmm3 = [-3.0E+0,-3.0E+0,-3.0E+0,-3.0E+0] 69 ; BDW-NEXT: vbroadcastss {{.*#+}} xmm1 = [-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1] 72 ; BDW-NEXT: vbroadcastss {{.*#+}} xmm2 = [NaN,NaN,NaN,NaN] 74 ; BDW-NEXT: vbroadcastss {{.*#+}} xmm2 = [1.17549435E-38,1.17549435E-38,1.17549435E-38,1.1754943… 137 ; BDW-NEXT: vbroadcastss {{.*#+}} ymm3 = [-3.0E+0,-3.0E+0,-3.0E+0,-3.0E+0,-3.0E+0,-3.0E+0,-3.0E+… 139 ; BDW-NEXT: vbroadcastss {{.*#+}} ymm1 = [-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-… 142 ; BDW-NEXT: vbroadcastss {{.*#+}} ymm2 = [NaN,NaN,NaN,NaN,NaN,NaN,NaN,NaN] 144 ; BDW-NEXT: vbroadcastss {{.*#+}} ymm2 = [1.17549435E-38,1.17549435E-38,1.17549435E-38,1.1754943… 216 ; BDW-NEXT: vbroadcastss {{.*#+}} xmm3 = [-3.0E+0,-3.0E+0,-3.0E+0,-3.0E+0] 218 ; BDW-NEXT: vbroadcastss {{.*#+}} xmm1 = [-5.0E-1,-5.0E-1,-5.0E-1,-5.0E-1] [all …]
|
D | avx512vl-vbroadcast.ll | 13 ; CHECK-NEXT: vbroadcastss (%rsp), %ymm0 # 16-byte Folded Reload 32 ; CHECK-NEXT: vbroadcastss (%rsp), %xmm0 # 16-byte Folded Reload 66 ; CHECK-NEXT: vbroadcastss %xmm0, %ymm0 77 ; CHECK-NEXT: vbroadcastss %xmm1, %ymm0 {%k1} 90 ; CHECK-NEXT: vbroadcastss %xmm0, %ymm0 {%k1} {z} 102 ; CHECK-NEXT: vbroadcastss %xmm0, %xmm0 113 ; CHECK-NEXT: vbroadcastss %xmm1, %xmm0 {%k1} 126 ; CHECK-NEXT: vbroadcastss %xmm0, %xmm0 {%k1} {z}
|
D | avx2-vbroadcast.ll | 148 ; X32-NEXT: vbroadcastss (%eax), %xmm0 153 ; X64-NEXT: vbroadcastss (%rdi), %xmm0 168 ; X32-NEXT: vbroadcastss (%eax), %ymm0 173 ; X64-NEXT: vbroadcastss (%rdi), %ymm0 366 ; X32-NEXT: vbroadcastss 4(%eax), %xmm0 371 ; X64-NEXT: vbroadcastss 4(%rdi), %xmm0 383 ; X32-NEXT: vbroadcastss 12(%eax), %ymm0 388 ; X64-NEXT: vbroadcastss 12(%rdi), %ymm0 400 ; X32-NEXT: vbroadcastss 20(%eax), %ymm0 405 ; X64-NEXT: vbroadcastss 20(%rdi), %ymm0 [all …]
|
D | avx512-vbroadcast.ll | 28 ; ALL-NEXT: vbroadcastss %xmm0, %zmm0 37 ; ALL-NEXT: vbroadcastss %xmm0, %zmm0 48 ; ALL-NEXT: vbroadcastss %xmm0, %zmm1 {%k1} 62 ; ALL-NEXT: vbroadcastss %xmm0, %zmm0 {%k1} {z} 74 ; ALL-NEXT: vbroadcastss (%rdi), %zmm0 86 ; ALL-NEXT: vbroadcastss (%rdi), %zmm0 {%k1} 100 ; ALL-NEXT: vbroadcastss (%rdi), %zmm0 {%k1} {z} 193 ; ALL-NEXT: vbroadcastss %xmm0, %zmm0 202 ; ALL-NEXT: vbroadcastss %xmm0, %zmm0 264 ; ALL-NEXT: vbroadcastss %xmm0, %zmm0 [all …]
|
D | avx2-fma-fneg-combine.ll | 47 ; X32-NEXT: vbroadcastss {{.*#+}} xmm1 = [-0.0E+0,-0.0E+0,-0.0E+0,-0.0E+0] 54 ; X64-NEXT: vbroadcastss {{.*#+}} xmm1 = [-0.0E+0,-0.0E+0,-0.0E+0,-0.0E+0] 121 ; X32-NEXT: vbroadcastss {{[0-9]+}}(%esp), %ymm2 127 ; X64-NEXT: vbroadcastss %xmm0, %ymm0 142 ; X32-NEXT: vbroadcastss {{[0-9]+}}(%esp), %ymm2 148 ; X64-NEXT: vbroadcastss %xmm0, %ymm0
|
D | avx-splat.ll | 82 ; X86-NEXT: vbroadcastss {{[0-9]+}}(%esp), %ymm0 99 ; X64-NEXT: vbroadcastss {{[0-9]+}}(%rsp), %ymm0 131 ; X86-NEXT: vbroadcastss {{[0-9]+}}(%esp), %ymm0 204 ; X86-NEXT: vbroadcastss (%eax), %xmm0 209 ; X64-NEXT: vbroadcastss (%rdi), %xmm0 220 ; X86-NEXT: vbroadcastss 28(%eax), %ymm0 225 ; X64-NEXT: vbroadcastss 28(%rdi), %ymm0
|
D | avx512-bugfix-25270.ll | 14 ; CHECK-NEXT: vbroadcastss {{.*#+}} zmm1 = [4,4,4,4,4,4,4,4,4,4,4,4,4,4,4,4] 19 ; CHECK-NEXT: vbroadcastss {{.*#+}} zmm0 = [7,7,7,7,7,7,7,7,7,7,7,7,7,7,7,7]
|
D | x86-upgrade-avx-vbroadcast.ll | 13 ; CHECK-NEXT: vbroadcastss (%rdi), %xmm0 37 ; CHECK-NEXT: vbroadcastss (%rdi), %ymm0
|
D | splat-for-size.ll | 54 ; CHECK-NEXT: vbroadcastss {{.*#+}} xmm1 = [1.0E+0,1.0E+0,1.0E+0,1.0E+0] 64 ; CHECK-NEXT: vbroadcastss {{.*#+}} xmm1 = [1.0E+0,1.0E+0,1.0E+0,1.0E+0] 74 ; CHECK-NEXT: vbroadcastss {{.*#+}} ymm1 = [1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0… 84 ; CHECK-NEXT: vbroadcastss {{.*#+}} ymm1 = [1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0E+0,1.0… 169 ; AVX can't do integer splats, so fake it: use vbroadcastss to splat 32-bit value. 173 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm1 = [2,2,2,2] 189 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm1 = [2,2,2,2] 202 ; AVX can't do integer splats, so fake it: use vbroadcastss to splat 32-bit value. 207 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm2 = [2,2,2,2] 226 ; AVX-NEXT: vbroadcastss {{.*#+}} xmm2 = [2,2,2,2]
|
/external/llvm/test/CodeGen/X86/ |
D | avx-vbroadcast.ll | 35 ; X32-NEXT: vbroadcastss (%eax), %ymm0 40 ; X64-NEXT: vbroadcastss (%rdi), %ymm0 75 ; X32-NEXT: vbroadcastss (%eax), %ymm0 80 ; X64-NEXT: vbroadcastss (%rdi), %ymm0 97 ; X32-NEXT: vbroadcastss (%eax), %xmm0 102 ; X64-NEXT: vbroadcastss (%rdi), %xmm0 137 ; X32-NEXT: vbroadcastss (%eax), %xmm0 142 ; X64-NEXT: vbroadcastss (%rdi), %xmm0 176 ; X32-NEXT: vbroadcastss 12(%eax), %ymm0 181 ; X64-NEXT: vbroadcastss 12(%rdi), %ymm0 [all …]
|
D | avx2-vbroadcast.ll | 146 ; X32-NEXT: vbroadcastss (%eax), %xmm0 151 ; X64-NEXT: vbroadcastss (%rdi), %xmm0 166 ; X32-NEXT: vbroadcastss (%eax), %ymm0 171 ; X64-NEXT: vbroadcastss (%rdi), %ymm0 343 ; X32-NEXT: vbroadcastss 4(%eax), %xmm0 348 ; X64-NEXT: vbroadcastss 4(%rdi), %xmm0 360 ; X32-NEXT: vbroadcastss 12(%eax), %ymm0 365 ; X64-NEXT: vbroadcastss 12(%rdi), %ymm0 377 ; X32-NEXT: vbroadcastss 20(%eax), %ymm0 382 ; X64-NEXT: vbroadcastss 20(%rdi), %ymm0 [all …]
|
D | avx512vl-vbroadcast.ll | 14 ; CHECK-NEXT: vbroadcastss {{[0-9]+}}(%rsp), %ymm0 # 4-byte Folded Reload 33 ; CHECK-NEXT: vbroadcastss {{[0-9]+}}(%rsp), %xmm0 # 4-byte Folded Reload 66 ; CHECK-NEXT: vbroadcastss %xmm0, %ymm0 78 ; CHECK-NEXT: vbroadcastss %xmm1, %ymm0 {%k1} 92 ; CHECK-NEXT: vbroadcastss %xmm0, %ymm0 {%k1} {z} 104 ; CHECK-NEXT: vbroadcastss %xmm0, %xmm0 116 ; CHECK-NEXT: vbroadcastss %xmm1, %xmm0 {%k1} 130 ; CHECK-NEXT: vbroadcastss %xmm0, %xmm0 {%k1} {z}
|
D | splat-for-size.ll | 32 ; CHECK-NEXT: vbroadcastss {{.*}}(%rip), %xmm1 42 ; CHECK-NEXT: vbroadcastss {{.*}}(%rip), %ymm1 82 ; AVX can't do integer splats, so fake it: use vbroadcastss to splat 32-bit value. 86 ; AVX-NEXT: vbroadcastss {{.*}}(%rip), %xmm1 99 ; AVX can't do integer splats, so fake it: use vbroadcastss to splat 32-bit value. 104 ; AVX-NEXT: vbroadcastss {{.*}}(%rip), %xmm2
|
D | x86-upgrade-avx-vbroadcast.ll | 13 ; CHECK-NEXT: vbroadcastss (%rdi), %xmm0 37 ; CHECK-NEXT: vbroadcastss (%rdi), %ymm0
|
D | avx512-vbroadcast.ll | 28 ; ALL-NEXT: vbroadcastss %xmm0, %zmm0 37 ; ALL-NEXT: vbroadcastss %xmm0, %zmm0 49 ; ALL-NEXT: vbroadcastss %xmm0, %zmm1 {%k1} 64 ; ALL-NEXT: vbroadcastss %xmm0, %zmm0 {%k1} {z} 76 ; ALL-NEXT: vbroadcastss (%rdi), %zmm0 89 ; ALL-NEXT: vbroadcastss (%rdi), %zmm0 {%k1} 104 ; ALL-NEXT: vbroadcastss (%rdi), %zmm0 {%k1} {z} 210 ; ALL-NEXT: vbroadcastss %xmm0, %zmm0 273 ; ALL-NEXT: vbroadcastss %xmm0, %zmm0 343 ; ALL-NEXT: vbroadcastss %xmm0, %zmm0 [all …]
|
D | vector-gep.ll | 98 ;CHECK: vbroadcastss 111 ;CHECK: vbroadcastss 121 ;CHECK: vbroadcastss
|
/external/llvm-project/llvm/test/CodeGen/X86/avx512-shuffles/ |
D | broadcast-scalar-fp.ll | 249 ; CHECK-NEXT: vbroadcastss %xmm0, %xmm0 260 ; CHECK-NEXT: vbroadcastss %xmm0, %xmm1 {%k1} 275 ; CHECK-NEXT: vbroadcastss %xmm0, %xmm0 {%k1} {z} 288 ; CHECK-NEXT: vbroadcastss %xmm0, %xmm1 {%k1} 303 ; CHECK-NEXT: vbroadcastss %xmm0, %xmm0 {%k1} {z} 316 ; CHECK-NEXT: vbroadcastss %xmm0, %xmm1 {%k1} 331 ; CHECK-NEXT: vbroadcastss %xmm0, %xmm0 {%k1} {z} 344 ; CHECK-NEXT: vbroadcastss %xmm0, %xmm1 {%k1} 359 ; CHECK-NEXT: vbroadcastss %xmm0, %xmm0 {%k1} {z} 370 ; CHECK-NEXT: vbroadcastss %xmm0, %ymm0 [all …]
|
/external/llvm-project/llvm/test/tools/llvm-mca/X86/SkylakeClient/ |
D | bottleneck-analysis.s | 10 vbroadcastss %xmm0, %xmm0 61 # CHECK-NEXT: | 5. vbroadcastss %xmm0, %xmm0 93 # CHECK-NEXT: 1 1 1.00 vbroadcastss %xmm0, %xmm0 134 # CHECK-NEXT: - - - - - - - 1.00 - - vbroadcastss %x…
|