/external/llvm-project/llvm/test/Analysis/BasicAA/ |
D | sequential-gep.ll | 4 ; CHECK: NoAlias: i32* %gep1, i32* %gep2 9 %gep2 = getelementptr [8 x i32], [8 x i32]* %p, i32 2, i32 %add 14 ; CHECK: MayAlias: i32* %gep1, i32* %gep2 19 %gep2 = getelementptr [8 x i32], [8 x i32]* %p, i32 0, i32 %add 24 ; CHECK: MustAlias: i32* %gep1, i32* %gep2 29 %gep2 = getelementptr [8 x i32], [8 x i32]* %p, i32 0, i32 %add 34 ; CHECK: MayAlias: i32* %gep1, i32* %gep2 39 %gep2 = getelementptr [8 x i32], [8 x i32]* %p, i32 %add, i32 %add 44 ; CHECK: MayAlias: i32* %gep2, i64* %bc 49 %gep2 = getelementptr [8 x i32], [8 x i32]* %p, i32 2, i32 %add [all …]
|
D | vscale.ll | 7 ; CHECK-DAG: MayAlias: <vscale x 4 x i32>* %alloc, <vscale x 4 x i32>* %gep2 8 ; CHECK-DAG: MayAlias: <vscale x 4 x i32>* %gep1, <vscale x 4 x i32>* %gep2 12 %gep2 = getelementptr <vscale x 4 x i32>, <vscale x 4 x i32>* %alloc, i64 1 18 ; CHECK-DAG: MayAlias: <vscale x 4 x i32>* %alloc, <vscale x 4 x i32>* %gep2 19 ; TODO: AliasResult for gep1,gep2 can be improved as MustAlias 20 ; CHECK-DAG: MayAlias: <vscale x 4 x i32>* %gep1, <vscale x 4 x i32>* %gep2 24 %gep2 = getelementptr <vscale x 4 x i32>, <vscale x 4 x i32>* %alloc, i64 1 30 ; CHECK-DAG: MayAlias: <vscale x 4 x i32>* %alloc, i32* %gep2 31 ; CHECK-DAG: MayAlias: <vscale x 4 x i32>* %gep1, i32* %gep2 35 %gep2 = getelementptr <vscale x 4 x i32>, <vscale x 4 x i32>* %alloc, i64 0, i64 1 [all …]
|
/external/llvm/test/Analysis/BasicAA/ |
D | sequential-gep.ll | 4 ; CHECK: NoAlias: i32* %gep1, i32* %gep2 9 %gep2 = getelementptr [8 x i32], [8 x i32]* %p, i32 2, i32 %add 14 ; CHECK: PartialAlias: i32* %gep1, i32* %gep2 19 %gep2 = getelementptr [8 x i32], [8 x i32]* %p, i32 0, i32 %add 24 ; CHECK: MustAlias: i32* %gep1, i32* %gep2 29 %gep2 = getelementptr [8 x i32], [8 x i32]* %p, i32 0, i32 %add 34 ; CHECK: PartialAlias: i32* %gep1, i32* %gep2 39 %gep2 = getelementptr [8 x i32], [8 x i32]* %p, i32 %add, i32 %add 44 ; CHECK: PartialAlias: i32* %gep2, i64* %bc 49 %gep2 = getelementptr [8 x i32], [8 x i32]* %p, i32 2, i32 %add
|
/external/llvm-project/llvm/test/CodeGen/AArch64/ |
D | falkor-hwpf.ll | 8 ; CHECK: load i32, i32* %gep2, align 4, !falkor.strided.access !0 12 ; NOHWPF: load i32, i32* %gep2, align 4{{$}} 23 %gep2 = getelementptr inbounds i32, i32* %p2, i32 %iv 24 %load2 = load i32, i32* %gep2 37 ; CHECK: load i32, i32* %gep2, align 4{{$}} 41 ; NOHWPF: load i32, i32* %gep2, align 4{{$}} 65 %gep2 = getelementptr inbounds i32, i32* %p, i32 %iv1 66 %load2 = load i32, i32* %gep2 82 ; CHECK: load i32, i32* %gep2, align 4{{$}} 86 ; NOHWPF: load i32, i32* %gep2, align 4{{$}} [all …]
|
D | aarch64-interleaved-ld-combine.ll | 29 %gep2 = getelementptr inbounds <4 x float>, <4 x float>* %ptr, i64 3 33 %ld2 = load <4 x float>, <4 x float>* %gep2, align 16 46 store <4 x float> %m4_7, <4 x float>* %gep2, align 16 90 %gep2 = getelementptr inbounds <4 x float>, <4 x float>* %ptr, i64 %idx2 95 %ld2 = load <4 x float>, <4 x float>* %gep2, align 16 108 store <4 x float> %m4_7, <4 x float>* %gep2, align 16 150 %gep2 = getelementptr %struct.ilc, %struct.ilc* %ptr, i32 0, i32 1, i64 %idx2 156 %ld2 = load <4 x float>, <4 x float>* %gep2, align 4 169 store <4 x float> %m4_7, <4 x float>* %gep2, align 16 196 %gep2 = getelementptr inbounds <4 x float>, <4 x float>* %ptr, i64 %idx2 [all …]
|
D | ldst-opt.ll | 296 %gep2 = getelementptr inbounds %pre.struct.i32, %pre.struct.i32* %load2, i64 0, i32 2 299 %retptr = phi i32* [ %gep1, %if.then ], [ %gep2, %if.end ] 314 %gep2 = getelementptr inbounds %pre.struct.i64, %pre.struct.i64* %load2, i64 0, i32 2 317 %retptr = phi i64* [ %gep1, %if.then ], [ %gep2, %if.end ] 332 %gep2 = getelementptr inbounds %pre.struct.i128, %pre.struct.i128* %load2, i64 0, i32 2 335 %retptr = phi <2 x i64>* [ %gep1, %if.then ], [ %gep2, %if.end ] 350 %gep2 = getelementptr inbounds %pre.struct.float, %pre.struct.float* %load2, i64 0, i32 2 353 %retptr = phi float* [ %gep1, %if.then ], [ %gep2, %if.end ] 368 %gep2 = getelementptr inbounds %pre.struct.double, %pre.struct.double* %load2, i64 0, i32 2 371 %retptr = phi double* [ %gep1, %if.then ], [ %gep2, %if.end ] [all …]
|
/external/llvm-project/llvm/test/Transforms/InstCombine/ |
D | select-gep.ll | 12 %gep2 = getelementptr i32, i32* %q, i64 4 14 %select = select i1 %cmp, i32* %gep1, i32* %gep2 26 %gep2 = getelementptr i32, i32* %q, i64 4 28 %select = select i1 %cmp, i32* %gep1, i32* %gep2 40 %gep2 = getelementptr inbounds i32, i32* %q, i64 4 42 %select = select i1 %cmp, i32* %gep1, i32* %gep2 54 %gep2 = getelementptr inbounds i32, i32* %q, i64 4 56 %select = select i1 %cmp, i32* %gep1, i32* %gep2 68 %gep2 = getelementptr inbounds i32, i32* %p, i64 %y 70 %select = select i1 %cmp, i32* %gep1, i32* %gep2 [all …]
|
D | icmp-custom-dl.ll | 18 %gep2 = getelementptr inbounds i8, i8* %foo, i64 10 20 %cmp = icmp ult i8* %cast1, %gep2 36 %gep2 = getelementptr inbounds i8, i8 addrspace(1)* %foo, i64 10 38 %cmp = icmp ult i8 addrspace(1)* %cast1, %gep2 54 %gep2 = getelementptr inbounds i8, i8* %foo, i64 %j 56 %cmp = icmp ult i8* %cast1, %gep2 70 %gep2 = getelementptr inbounds i8, i8 addrspace(1)* %foo, i64 %j 72 %cmp = icmp ult i8 addrspace(1)* %cast1, %gep2 88 %gep2 = getelementptr inbounds i8, i8* %foo, i64 %j 90 %cmp = icmp ult i8* %cast1, %gep2 [all …]
|
/external/llvm-project/llvm/test/Transforms/SLPVectorizer/X86/ |
D | no_alternate_divrem.ll | 38 %gep2.0 = getelementptr i32, i32* %arr2, i32 0 39 %gep2.1 = getelementptr i32, i32* %arr2, i32 1 40 %gep2.2 = getelementptr i32, i32* %arr2, i32 2 41 %gep2.3 = getelementptr i32, i32* %arr2, i32 3 60 store i32 %res0, i32* %gep2.0 61 store i32 %res1, i32* %gep2.1 62 store i32 %res2, i32* %gep2.2 63 store i32 %res3, i32* %gep2.3 102 %gep2.0 = getelementptr i32, i32* %arr2, i32 0 103 %gep2.1 = getelementptr i32, i32* %arr2, i32 1 [all …]
|
/external/llvm-project/llvm/test/Transforms/CodeGenPrepare/ARM/ |
D | sink-addrmode.ll | 30 ; CHECK-NOT: phi i32* [ %gep1, %if.then ], [ %gep2, %if.else ] 41 %gep2 = getelementptr inbounds i32, i32* %ptr, i32 2 45 %phi = phi i32* [ %gep1, %if.then ], [ %gep2, %if.else ] 72 ; CHECK-NOT: phi i32* [ %gep1, %if.then ], [ %gep2, %if.else ] 83 %gep2 = getelementptr inbounds i32, i32* %ptr, i32 %off2 87 %phi = phi i32* [ %gep1, %if.then ], [ %gep2, %if.else ] 95 ; CHECK-NOT: phi i32* [ %gep1, %if.then ], [ %gep2, %if.else ] 106 %gep2 = getelementptr inbounds i32, i32* %ptr2, i32 1 110 %phi = phi i32* [ %gep1, %if.then ], [ %gep2, %if.else ] 118 ; CHECK-NOT: phi i32* [ %gep1, %if.then ], [ %gep2, %if.else ] [all …]
|
/external/llvm/test/Transforms/LoadStoreVectorizer/X86/ |
D | subchain-interleaved.ll | 16 %next.gep2 = getelementptr i32, i32* %ptr, i64 2 23 %l4 = load i32, i32* %next.gep2, align 4 36 %next.gep2 = getelementptr i32, i32* %ptr, i64 2 43 %l4 = load i32, i32* %next.gep2, align 4 55 %next.gep2 = getelementptr i32, i32* %ptr, i64 2 61 %l4 = load i32, i32* %next.gep2, align 4 74 %next.gep2 = getelementptr i32, i32* %ptr, i64 2 80 store i32 0, i32* %next.gep2, align 4 82 %l4 = load i32, i32* %next.gep2, align 4
|
/external/llvm-project/llvm/test/CodeGen/AMDGPU/ |
D | max3.ll | 11 %gep2 = getelementptr i32, i32 addrspace(1)* %cptr, i32 %tid 15 %c = load i32, i32 addrspace(1)* %gep2 30 %gep2 = getelementptr i32, i32 addrspace(1)* %cptr, i32 %tid 34 %c = load i32, i32 addrspace(1)* %gep2 54 %gep2 = getelementptr i16, i16 addrspace(1)* %cptr, i32 %tid 58 %c = load i16, i16 addrspace(1)* %gep2 78 %gep2 = getelementptr i16, i16 addrspace(1)* %cptr, i32 %tid 82 %c = load i16, i16 addrspace(1)* %gep2 102 %gep2 = getelementptr i8, i8 addrspace(1)* %cptr, i32 %tid 106 %c = load i8, i8 addrspace(1)* %gep2 [all …]
|
D | min3.ll | 11 %gep2 = getelementptr i32, i32 addrspace(1)* %cptr, i32 %tid 15 %c = load i32, i32 addrspace(1)* %gep2 30 %gep2 = getelementptr i32, i32 addrspace(1)* %cptr, i32 %tid 34 %c = load i32, i32 addrspace(1)* %gep2 51 %gep2 = getelementptr i32, i32 addrspace(1)* %cptr, i32 %tid 62 %c = load i32, i32 addrspace(1)* %gep2 85 %gep2 = getelementptr i32, i32 addrspace(1)* %cptr, i32 %tid 96 %c = load i32, i32 addrspace(1)* %gep2 124 %gep2 = getelementptr i16, i16 addrspace(1)* %cptr, i32 %tid 128 %c = load i16, i16 addrspace(1)* %gep2 [all …]
|
D | promote-alloca-to-lds-constantexpr-use.ll | 25 %gep2 = getelementptr inbounds [4 x i32], [4 x i32] addrspace(5)* %stack, i32 0, i32 2 29 store i32 99, i32 addrspace(5)* %gep2 52 %gep2 = getelementptr inbounds [4 x i32], [4 x i32] addrspace(5)* %stack, i32 0, i32 2 56 store i32 99, i32 addrspace(5)* %gep2 77 %gep2 = getelementptr inbounds [4 x i32], [4 x i32] addrspace(5)* %stack, i32 0, i32 2 81 store i32 99, i32 addrspace(5)* %gep2 101 %gep2 = getelementptr inbounds [4 x i32], [4 x i32] addrspace(5)* %stack, i32 0, i32 2 105 store i32 99, i32 addrspace(5)* %gep2 125 %gep2 = getelementptr inbounds [4 x i32], [4 x i32] addrspace(5)* %stack, i32 0, i32 2 129 store i32 99, i32 addrspace(5)* %gep2 [all …]
|
D | fmed3.ll | 167 %gep2 = getelementptr float, float addrspace(1)* %cptr, i32 %tid 171 %c = load volatile float, float addrspace(1)* %gep2 190 %gep2 = getelementptr float, float addrspace(1)* %cptr, i32 %tid 194 %c = load volatile float, float addrspace(1)* %gep2 213 %gep2 = getelementptr float, float addrspace(1)* %cptr, i32 %tid 217 %c = load volatile float, float addrspace(1)* %gep2 236 %gep2 = getelementptr float, float addrspace(1)* %cptr, i32 %tid 240 %c = load volatile float, float addrspace(1)* %gep2 265 %gep2 = getelementptr float, float addrspace(1)* %cptr, i32 %tid 269 %c = load volatile float, float addrspace(1)* %gep2 [all …]
|
D | local-stack-slot-offset.ll | 22 %gep2.store = getelementptr [513 x float], [513 x float] addrspace(5)* %m2, i32 0, i32 %idx2 23 store float %v2, float addrspace(5)* %gep2.store 30 %gep2.load = getelementptr [513 x float], [513 x float] addrspace(5)* %m2, i32 0, i32 512 31 %out2 = load float, float addrspace(5)* %gep2.load
|
/external/llvm/test/CodeGen/AArch64/ |
D | ldst-opt.ll | 295 %gep2 = getelementptr inbounds %pre.struct.i32, %pre.struct.i32* %load2, i64 0, i32 2 298 %retptr = phi i32* [ %gep1, %if.then ], [ %gep2, %if.end ] 313 %gep2 = getelementptr inbounds %pre.struct.i64, %pre.struct.i64* %load2, i64 0, i32 2 316 %retptr = phi i64* [ %gep1, %if.then ], [ %gep2, %if.end ] 331 %gep2 = getelementptr inbounds %pre.struct.i128, %pre.struct.i128* %load2, i64 0, i32 2 334 %retptr = phi <2 x i64>* [ %gep1, %if.then ], [ %gep2, %if.end ] 349 %gep2 = getelementptr inbounds %pre.struct.float, %pre.struct.float* %load2, i64 0, i32 2 352 %retptr = phi float* [ %gep1, %if.then ], [ %gep2, %if.end ] 367 %gep2 = getelementptr inbounds %pre.struct.double, %pre.struct.double* %load2, i64 0, i32 2 370 %retptr = phi double* [ %gep1, %if.then ], [ %gep2, %if.end ] [all …]
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | frem.ll | 20 %gep2 = getelementptr float, float addrspace(1)* %in2, i32 4 22 %r1 = load float, float addrspace(1)* %gep2, align 4 39 %gep2 = getelementptr float, float addrspace(1)* %in2, i32 4 41 %r1 = load float, float addrspace(1)* %gep2, align 4 85 %gep2 = getelementptr <2 x float>, <2 x float> addrspace(1)* %in2, i32 4 87 %r1 = load <2 x float>, <2 x float> addrspace(1)* %gep2, align 8 95 %gep2 = getelementptr <4 x float>, <4 x float> addrspace(1)* %in2, i32 4 97 %r1 = load <4 x float>, <4 x float> addrspace(1)* %gep2, align 16 105 %gep2 = getelementptr <2 x double>, <2 x double> addrspace(1)* %in2, i32 4 107 %r1 = load <2 x double>, <2 x double> addrspace(1)* %gep2, align 16
|
D | mad-sub.ll | 19 %gep2 = getelementptr float, float addrspace(1)* %ptr, i64 %add2 23 %c = load volatile float, float addrspace(1)* %gep2, align 4 43 %gep2 = getelementptr float, float addrspace(1)* %ptr, i64 %add2 47 %c = load volatile float, float addrspace(1)* %gep2, align 4 64 %gep2 = getelementptr double, double addrspace(1)* %ptr, i64 %add2 68 %c = load volatile double, double addrspace(1)* %gep2, align 8 88 %gep2 = getelementptr float, float addrspace(1)* %ptr, i64 %add2 92 %c = load volatile float, float addrspace(1)* %gep2, align 4 113 %gep2 = getelementptr float, float addrspace(1)* %ptr, i64 %add2 117 %c = load volatile float, float addrspace(1)* %gep2, align 4 [all …]
|
D | min3.ll | 11 %gep2 = getelementptr i32, i32 addrspace(1)* %cptr, i32 %tid 15 %c = load i32, i32 addrspace(1)* %gep2, align 4 30 %gep2 = getelementptr i32, i32 addrspace(1)* %cptr, i32 %tid 34 %c = load i32, i32 addrspace(1)* %gep2, align 4 51 %gep2 = getelementptr i32, i32 addrspace(1)* %cptr, i32 %tid 62 %c = load i32, i32 addrspace(1)* %gep2, align 4 85 %gep2 = getelementptr i32, i32 addrspace(1)* %cptr, i32 %tid 96 %c = load i32, i32 addrspace(1)* %gep2, align 4
|
/external/llvm-project/llvm/test/Transforms/LoadStoreVectorizer/X86/ |
D | subchain-interleaved.ll | 17 %next.gep2 = getelementptr i32, i32* %ptr, i64 2 24 %l4 = load i32, i32* %next.gep2, align 4 37 %next.gep2 = getelementptr i32, i32* %ptr, i64 2 44 %l4 = load i32, i32* %next.gep2, align 4 56 %next.gep2 = getelementptr i32, i32* %ptr, i64 2 62 %l4 = load i32, i32* %next.gep2, align 4 75 %next.gep2 = getelementptr i32, i32* %ptr, i64 2 81 store i32 0, i32* %next.gep2, align 4 83 %l4 = load i32, i32* %next.gep2, align 4
|
/external/llvm/test/Transforms/LoopStrengthReduce/ |
D | dont_insert_redundant_ops.ll | 17 %gep2 = getelementptr { i32, i32 }, { i32, i32 }* %P, i32 %INDVAR, i32 1 ; <i32*> [#uses=1] 18 store i32 0, i32* %gep2 33 %gep2 = getelementptr [2 x i32], [2 x i32]* %P, i32 %INDVAR, i64 1 ; <i32*> [#uses=1] 34 store i32 0, i32* %gep2
|
/external/llvm-project/llvm/test/Transforms/LoopStrengthReduce/ |
D | dont_insert_redundant_ops.ll | 17 %gep2 = getelementptr { i32, i32 }, { i32, i32 }* %P, i32 %INDVAR, i32 1 ; <i32*> [#uses=1] 18 store i32 0, i32* %gep2 33 %gep2 = getelementptr [2 x i32], [2 x i32]* %P, i32 %INDVAR, i64 1 ; <i32*> [#uses=1] 34 store i32 0, i32* %gep2
|
/external/llvm-project/llvm/test/Transforms/SROA/ |
D | fca.ll | 18 %gep2 = getelementptr inbounds { i32, i32 }, { i32, i32 }* %a, i32 0, i32 1 19 store i32 %y, i32* %gep2 42 %gep2 = getelementptr inbounds { i32, i32 }, { i32, i32 }* %a, i32 0, i32 1 43 store i32 %y, i32* %gep2
|
/external/llvm/test/Transforms/SROA/ |
D | fca.ll | 18 %gep2 = getelementptr inbounds { i32, i32 }, { i32, i32 }* %a, i32 0, i32 1 19 store i32 %y, i32* %gep2 42 %gep2 = getelementptr inbounds { i32, i32 }, { i32, i32 }* %a, i32 0, i32 1 43 store i32 %y, i32* %gep2
|