/external/llvm-project/llvm/test/Analysis/BasicAA/ |
D | gep-decomposition-limit.ll | 5 ; CHECK-DAG: MustAlias: i8* %gep.add5, i8* %gep.inc5 6 ; CHECK-DAG: NoAlias: i8* %gep.inc3, i8* %gep.inc5 7 ; CHECK-DAG: NoAlias: i8* %gep.inc4, i8* %gep.inc5 9 ; CHECK-DAG: MustAlias: i8* %gep.add6, i8* %gep.inc6 10 ; CHECK-DAG: NoAlias: i8* %gep.inc4, i8* %gep.inc6 11 ; CHECK-DAG: NoAlias: i8* %gep.inc5, i8* %gep.inc6 13 ; CHECK-DAG: MayAlias: i8* %gep.add7, i8* %gep.inc7 14 ; CHECK-DAG: MayAlias: i8* %gep.inc5, i8* %gep.inc7 15 ; CHECK-DAG: NoAlias: i8* %gep.inc6, i8* %gep.inc7 18 %gep.add5 = getelementptr i8, i8* %base, i64 5 [all …]
|
D | libfuncs.ll | 6 ; CHECK-NEXT: Just Ref: Ptr: i8* %a.gep.1 <-> %res = tail call i32 @memcmp(i8* %a, i8* %b, i64 4) 7 ; CHECK-NEXT: NoModRef: Ptr: i8* %a.gep.5 <-> %res = tail call i32 @memcmp(i8* %a, i8* %b, i64 4) 8 ; CHECK-NEXT: Just Ref: Ptr: i8* %b.gep.1 <-> %res = tail call i32 @memcmp(i8* %a, i8* %b, i64 4) 9 ; CHECK-NEXT: NoModRef: Ptr: i8* %b.gep.5 <-> %res = tail call i32 @memcmp(i8* %a, i8* %b, i64 4) 13 %a.gep.1 = getelementptr i8, i8* %a, i32 1 14 store i8 0, i8* %a.gep.1 15 %a.gep.5 = getelementptr i8, i8* %a, i32 5 16 store i8 1, i8* %a.gep.5 17 %b.gep.1 = getelementptr i8, i8* %b, i32 1 18 store i8 2, i8* %b.gep.1 [all …]
|
/external/llvm/test/Transforms/LoadStoreVectorizer/AMDGPU/ |
D | merge-stores.ll | 14 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 16 store i8 123, i8 addrspace(1)* %out.gep.1 24 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 26 store i8 123, i8 addrspace(1)* %out.gep.1 34 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 36 store i16 123, i16 addrspace(1)* %out.gep.1 44 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 46 store i16 0, i16 addrspace(1)* %out.gep.1 54 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 56 store i16 123, i16 addrspace(1)* %out.gep.1 [all …]
|
/external/llvm-project/llvm/test/Transforms/LoadStoreVectorizer/AMDGPU/ |
D | merge-stores.ll | 14 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 16 store i8 123, i8 addrspace(1)* %out.gep.1 24 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 26 store i8 123, i8 addrspace(1)* %out.gep.1 34 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 36 store i16 123, i16 addrspace(1)* %out.gep.1 44 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 46 store i16 0, i16 addrspace(1)* %out.gep.1 55 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 57 store i16 123, i16 addrspace(1)* %out.gep.1 [all …]
|
D | merge-stores-private.ll | 21 %out.gep.1 = getelementptr i32, i32 addrspace(5)* %out, i32 1 22 %out.gep.2 = getelementptr i32, i32 addrspace(5)* %out, i32 2 23 %out.gep.3 = getelementptr i32, i32 addrspace(5)* %out, i32 3 26 store i32 1, i32 addrspace(5)* %out.gep.1 27 store i32 23, i32 addrspace(5)* %out.gep.2 28 store i32 19, i32 addrspace(5)* %out.gep.3 34 ; ALIGNED: store i32 1, i32 addrspace(5)* %out.gep.1, align 1 35 ; ALIGNED: store i32 23, i32 addrspace(5)* %out.gep.2, align 1 36 ; ALIGNED: store i32 19, i32 addrspace(5)* %out.gep.3, align 1 48 %out.gep.1 = getelementptr i32, i32 addrspace(5)* %out, i32 1 [all …]
|
/external/llvm/test/CodeGen/AMDGPU/ |
D | merge-stores.ll | 19 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 21 store i8 123, i8 addrspace(1)* %out.gep.1 31 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 33 store i8 123, i8 addrspace(1)* %out.gep.1 41 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 43 store i16 123, i16 addrspace(1)* %out.gep.1 51 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 53 store i16 0, i16 addrspace(1)* %out.gep.1 63 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 65 store i16 123, i16 addrspace(1)* %out.gep.1 [all …]
|
D | mad-combine.ll | 36 %gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 37 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 38 %gep.2 = getelementptr float, float addrspace(1)* %gep.0, i32 2 39 %gep.out = getelementptr float, float addrspace(1)* %out, i32 %tid 41 %a = load volatile float, float addrspace(1)* %gep.0 42 %b = load volatile float, float addrspace(1)* %gep.1 43 %c = load volatile float, float addrspace(1)* %gep.2 47 store float %fma, float addrspace(1)* %gep.out 75 %gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 76 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 [all …]
|
D | commute-compares.ll | 13 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid 14 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid 15 %val = load i32, i32 addrspace(1)* %gep.in 18 store i32 %ext, i32 addrspace(1)* %gep.out 26 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid 27 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid 28 %val = load i32, i32 addrspace(1)* %gep.in 31 store i32 %ext, i32 addrspace(1)* %gep.out 41 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid 42 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid [all …]
|
D | fma-combine.ll | 18 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 19 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 20 %gep.2 = getelementptr double, double addrspace(1)* %gep.0, i32 2 21 %gep.out = getelementptr double, double addrspace(1)* %out, i32 %tid 23 %a = load volatile double, double addrspace(1)* %gep.0 24 %b = load volatile double, double addrspace(1)* %gep.1 25 %c = load volatile double, double addrspace(1)* %gep.2 29 store double %fma, double addrspace(1)* %gep.out 46 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 47 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 [all …]
|
D | fmuladd.ll | 41 %gep.0 = getelementptr float, float addrspace(1)* %out, i32 %tid 42 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 43 %gep.out = getelementptr float, float addrspace(1)* %out, i32 %tid 45 %r1 = load volatile float, float addrspace(1)* %gep.0 46 %r2 = load volatile float, float addrspace(1)* %gep.1 49 store float %r3, float addrspace(1)* %gep.out 60 %gep.0 = getelementptr float, float addrspace(1)* %out, i32 %tid 61 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 62 %gep.out = getelementptr float, float addrspace(1)* %out, i32 %tid 64 %r1 = load volatile float, float addrspace(1)* %gep.0 [all …]
|
D | madmk.ll | 17 %gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 18 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 19 %out.gep = getelementptr float, float addrspace(1)* %out, i32 %tid 21 %a = load volatile float, float addrspace(1)* %gep.0, align 4 22 %b = load volatile float, float addrspace(1)* %gep.1, align 4 26 store float %madmk, float addrspace(1)* %out.gep, align 4 41 %in.gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 42 %in.gep.1 = getelementptr float, float addrspace(1)* %in.gep.0, i32 1 43 %in.gep.2 = getelementptr float, float addrspace(1)* %in.gep.0, i32 2 45 %out.gep.0 = getelementptr float, float addrspace(1)* %out, i32 %tid [all …]
|
D | shift-and-i64-ubfe.ll | 13 %in.gep = getelementptr i64, i64 addrspace(1)* %in, i32 %id.x 14 %out.gep = getelementptr i64, i64 addrspace(1)* %out, i32 %id.x 15 %ld.64 = load i64, i64 addrspace(1)* %in.gep 18 store i64 %bit, i64 addrspace(1)* %out.gep 30 %in.gep = getelementptr i64, i64 addrspace(1)* %in, i32 %id.x 31 %out.gep = getelementptr i64, i64 addrspace(1)* %out, i32 %id.x 32 %ld.64 = load i64, i64 addrspace(1)* %in.gep 35 store i64 %bit, i64 addrspace(1)* %out.gep 46 %in.gep = getelementptr i64, i64 addrspace(1)* %in, i32 %id.x 47 %out.gep = getelementptr i64, i64 addrspace(1)* %out, i32 %id.x [all …]
|
D | fmax_legacy.f64.ll | 9 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 10 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 12 %a = load double, double addrspace(1)* %gep.0, align 8 13 %b = load double, double addrspace(1)* %gep.1, align 8 24 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 25 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 27 %a = load double, double addrspace(1)* %gep.0, align 8 28 %b = load double, double addrspace(1)* %gep.1, align 8 39 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 40 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 [all …]
|
/external/llvm-project/llvm/test/CodeGen/AMDGPU/ |
D | merge-stores.ll | 16 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 18 store i8 123, i8 addrspace(1)* %out.gep.1 28 %out.gep.1 = getelementptr i8, i8 addrspace(1)* %out, i32 1 30 store i8 123, i8 addrspace(1)* %out.gep.1 38 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 40 store i16 123, i16 addrspace(1)* %out.gep.1 48 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 50 store i16 0, i16 addrspace(1)* %out.gep.1 60 %out.gep.1 = getelementptr i16, i16 addrspace(1)* %out, i32 1 62 store i16 123, i16 addrspace(1)* %out.gep.1 [all …]
|
D | mad-combine.ll | 38 %gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 39 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 40 %gep.2 = getelementptr float, float addrspace(1)* %gep.0, i32 2 41 %gep.out = getelementptr float, float addrspace(1)* %out, i32 %tid 43 %a = load volatile float, float addrspace(1)* %gep.0 44 %b = load volatile float, float addrspace(1)* %gep.1 45 %c = load volatile float, float addrspace(1)* %gep.2 49 store float %fma, float addrspace(1)* %gep.out 77 %gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 78 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 [all …]
|
D | commute-compares.ll | 13 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid 14 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid 15 %val = load i32, i32 addrspace(1)* %gep.in 18 store i32 %ext, i32 addrspace(1)* %gep.out 26 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid 27 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid 28 %val = load i32, i32 addrspace(1)* %gep.in 31 store i32 %ext, i32 addrspace(1)* %gep.out 41 %gep.in = getelementptr i32, i32 addrspace(1)* %in, i32 %tid 42 %gep.out = getelementptr i32, i32 addrspace(1)* %out, i32 %tid [all …]
|
D | fma-combine.ll | 26 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 27 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 28 %gep.2 = getelementptr double, double addrspace(1)* %gep.0, i32 2 29 %gep.out = getelementptr double, double addrspace(1)* %out, i32 %tid 31 %a = load volatile double, double addrspace(1)* %gep.0 32 %b = load volatile double, double addrspace(1)* %gep.1 33 %c = load volatile double, double addrspace(1)* %gep.2 37 store double %fma, double addrspace(1)* %gep.out 54 %gep.0 = getelementptr double, double addrspace(1)* %in, i32 %tid 55 %gep.1 = getelementptr double, double addrspace(1)* %gep.0, i32 1 [all …]
|
D | icmp.i16.ll | 15 %a.gep = getelementptr inbounds i16, i16 addrspace(1)* %a.ptr, i64 %tid.ext 16 %b.gep = getelementptr inbounds i16, i16 addrspace(1)* %b.ptr, i64 %tid.ext 17 %out.gep = getelementptr inbounds i32, i32 addrspace(1)* %out, i64 %tid.ext 18 %a = load i16, i16 addrspace(1)* %a.gep 19 %b = load i16, i16 addrspace(1)* %b.gep 22 store i32 %tmp1, i32 addrspace(1)* %out.gep 33 %a.gep = getelementptr inbounds i16, i16 addrspace(1)* %a.ptr, i64 %tid.ext 34 %b.gep = getelementptr inbounds i16, i16 addrspace(1)* %b.ptr, i64 %tid.ext 35 %out.gep = getelementptr inbounds i32, i32 addrspace(1)* %out, i64 %tid.ext 36 %a = load i16, i16 addrspace(1)* %a.gep [all …]
|
D | madmk.ll | 17 %gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 18 %gep.1 = getelementptr float, float addrspace(1)* %gep.0, i32 1 19 %out.gep = getelementptr float, float addrspace(1)* %out, i32 %tid 21 %a = load volatile float, float addrspace(1)* %gep.0, align 4 22 %b = load volatile float, float addrspace(1)* %gep.1, align 4 26 store float %madmk, float addrspace(1)* %out.gep, align 4 41 %in.gep.0 = getelementptr float, float addrspace(1)* %in, i32 %tid 42 %in.gep.1 = getelementptr float, float addrspace(1)* %in.gep.0, i32 1 43 %in.gep.2 = getelementptr float, float addrspace(1)* %in.gep.0, i32 2 45 %out.gep.0 = getelementptr float, float addrspace(1)* %out, i32 %tid [all …]
|
D | add.i16.ll | 3 ; FIXME: Need to handle non-uniform case for function below (load without gep). 11 %gep.out = getelementptr inbounds i16, i16 addrspace(1)* %out, i32 %tid 12 %gep.in0 = getelementptr inbounds i16, i16 addrspace(1)* %in0, i32 %tid 13 %gep.in1 = getelementptr inbounds i16, i16 addrspace(1)* %in1, i32 %tid 14 %a = load volatile i16, i16 addrspace(1)* %gep.in0 15 %b = load volatile i16, i16 addrspace(1)* %gep.in1 21 ; FIXME: Need to handle non-uniform case for function below (load without gep). 28 %gep.out = getelementptr inbounds i16, i16 addrspace(1)* %out, i32 %tid 29 %gep.in0 = getelementptr inbounds i16, i16 addrspace(1)* %in0, i32 %tid 30 %a = load volatile i16, i16 addrspace(1)* %gep.in0 [all …]
|
D | shift-and-i64-ubfe.ll | 14 %in.gep = getelementptr i64, i64 addrspace(1)* %in, i32 %id.x 15 %out.gep = getelementptr i64, i64 addrspace(1)* %out, i32 %id.x 16 %ld.64 = load i64, i64 addrspace(1)* %in.gep 19 store i64 %bit, i64 addrspace(1)* %out.gep 32 %in.gep = getelementptr i64, i64 addrspace(1)* %in, i32 %id.x 33 %out.gep = getelementptr i64, i64 addrspace(1)* %out, i32 %id.x 34 %ld.64 = load i64, i64 addrspace(1)* %in.gep 37 store i64 %bit, i64 addrspace(1)* %out.gep 48 %in.gep = getelementptr i64, i64 addrspace(1)* %in, i32 %id.x 49 %out.gep = getelementptr i64, i64 addrspace(1)* %out, i32 %id.x [all …]
|
D | sub.i16.ll | 4 ; FIXME: Need to handle non-uniform case for function below (load without gep). 12 %gep.out = getelementptr inbounds i16, i16 addrspace(1)* %out, i32 %tid 13 %gep.in0 = getelementptr inbounds i16, i16 addrspace(1)* %in0, i32 %tid 14 %gep.in1 = getelementptr inbounds i16, i16 addrspace(1)* %in1, i32 %tid 15 %a = load volatile i16, i16 addrspace(1)* %gep.in0 16 %b = load volatile i16, i16 addrspace(1)* %gep.in1 22 ; FIXME: Need to handle non-uniform case for function below (load without gep). 29 %gep.out = getelementptr inbounds i16, i16 addrspace(1)* %out, i32 %tid 30 %gep.in0 = getelementptr inbounds i16, i16 addrspace(1)* %in0, i32 %tid 31 %a = load volatile i16, i16 addrspace(1)* %gep.in0 [all …]
|
/external/llvm-project/llvm/test/Transforms/SLPVectorizer/AArch64/ |
D | vectorizable-selects-min-max.ll | 32 %gep.1 = getelementptr inbounds i16, i16* %ptr, i16 1 33 %l.1 = load i16, i16* %gep.1 36 store i16 %s.1, i16* %gep.1, align 2 38 %gep.2 = getelementptr inbounds i16, i16* %ptr, i16 2 39 %l.2 = load i16, i16* %gep.2 42 store i16 %s.2, i16* %gep.2, align 2 44 %gep.3 = getelementptr inbounds i16, i16* %ptr, i16 3 45 %l.3 = load i16, i16* %gep.3 48 store i16 %s.3, i16* %gep.3, align 2 50 %gep.4 = getelementptr inbounds i16, i16* %ptr, i16 4 [all …]
|
D | vectorizable-selects-uniform-cmps.ll | 62 %gep.1 = getelementptr inbounds i16, i16* %ptr, i16 1 63 %l.1 = load i16, i16* %gep.1 66 store i16 %s.1, i16* %gep.1, align 2 68 %gep.2 = getelementptr inbounds i16, i16* %ptr, i16 2 69 %l.2 = load i16, i16* %gep.2 72 store i16 %s.2, i16* %gep.2, align 2 74 %gep.3 = getelementptr inbounds i16, i16* %ptr, i16 3 75 %l.3 = load i16, i16* %gep.3 78 store i16 %s.3, i16* %gep.3, align 2 80 %gep.4 = getelementptr inbounds i16, i16* %ptr, i16 4 [all …]
|
/external/llvm-project/llvm/test/Transforms/LoopVectorize/ |
D | alias-set-with-uncomputable-bounds.ll | 22 %gep.1 = getelementptr inbounds i32, i32* %ptr.3, i64 %iv 23 %offset.1 = load i32, i32* %gep.1, align 4 24 %gep.2 = getelementptr inbounds i32, i32* %ptr.2, i32 %offset.1 25 %lv = load i32, i32* %gep.2, align 4 26 %gep.3 = getelementptr inbounds i32, i32* %ptr.1, i64 %iv 27 store i32 %lv , i32* %gep.3, align 4 29 %gep.4 = getelementptr inbounds i32, i32* %ptr.1, i64 %offset.2 30 store i32 %lv, i32* %gep.4, align 4 58 %gep.1 = getelementptr inbounds i32, i32* %ptr.3, i64 %iv 59 %offset.1 = load i32, i32* %gep.1, align 4 [all …]
|