1; NOTE: Assertions have been autogenerated by utils/update_test_checks.py 2; RUN: opt -instcombine -S < %s | FileCheck %s 3 4target datalayout = "e-m:e-i64:64-f80:128-n8:16:32:64-S128-ni:4" 5target triple = "x86_64-unknown-linux-gnu" 6 7define i8 addrspace(4)* @f_0() { 8; CHECK-LABEL: @f_0( 9; CHECK-NEXT: ret i8 addrspace(4)* getelementptr (i8, i8 addrspace(4)* null, i64 50) 10; 11 %result = getelementptr i8, i8 addrspace(4)* null, i64 50 12 ret i8 addrspace(4)* %result 13} 14 15define i8 addrspace(3)* @f_1() { 16; inttoptr is fine here since addrspace(3) is integral. 17; CHECK-LABEL: @f_1( 18; CHECK-NEXT: ret i8 addrspace(3)* inttoptr (i64 50 to i8 addrspace(3)*) 19; 20 %result = getelementptr i8, i8 addrspace(3)* null, i64 50 21 ret i8 addrspace(3)* %result 22} 23 24define void @f_2(i8 addrspace(4)** %ptr0, i8 addrspace(4)** %ptr1) { 25; It is not okay to convert the load/store pair to load and store 26; integers, since pointers in address space 4 are non-integral. 27; CHECK-LABEL: @f_2( 28; CHECK-NEXT: entry: 29; CHECK-NEXT: [[VAL:%.*]] = load i8 addrspace(4)*, i8 addrspace(4)** [[PTR0:%.*]], align 8 30; CHECK-NEXT: store i8 addrspace(4)* [[VAL]], i8 addrspace(4)** [[PTR1:%.*]], align 8 31; CHECK-NEXT: ret void 32; 33entry: 34 %val = load i8 addrspace(4)*, i8 addrspace(4)** %ptr0 35 store i8 addrspace(4)* %val, i8 addrspace(4)** %ptr1 36 ret void 37} 38 39define void @f_3(i8 addrspace(3)** %ptr0, i8 addrspace(3)** %ptr1) { 40; It *is* okay to convert the load/store pair to load and store 41; integers, since pointers in address space 3 are integral. 42; CHECK-LABEL: @f_3( 43; CHECK-NEXT: entry: 44; CHECK-NEXT: [[VAL:%.*]] = load i8 addrspace(3)*, i8 addrspace(3)** [[PTR0:%.*]], align 8 45; CHECK-NEXT: store i8 addrspace(3)* [[VAL]], i8 addrspace(3)** [[PTR1:%.*]], align 8 46; CHECK-NEXT: ret void 47; 48entry: 49 %val = load i8 addrspace(3)*, i8 addrspace(3)** %ptr0 50 store i8 addrspace(3)* %val, i8 addrspace(3)** %ptr1 51 ret void 52} 53 54define i64 @g(i8 addrspace(4)** %gp) { 55; CHECK-LABEL: @g( 56; CHECK-NEXT: [[DOTPRE:%.*]] = load i8 addrspace(4)*, i8 addrspace(4)** [[GP:%.*]], align 8 57; CHECK-NEXT: [[V74:%.*]] = call i8 addrspace(4)* @alloc() 58; CHECK-NEXT: [[V77:%.*]] = getelementptr i8, i8 addrspace(4)* [[V74]], i64 -8 59; CHECK-NEXT: [[TMP1:%.*]] = bitcast i8 addrspace(4)* [[V77]] to i8 addrspace(4)* addrspace(4)* 60; CHECK-NEXT: [[TMP2:%.*]] = addrspacecast i8 addrspace(4)* addrspace(4)* [[TMP1]] to i8 addrspace(4)** 61; CHECK-NEXT: store i8 addrspace(4)* [[DOTPRE]], i8 addrspace(4)** [[TMP2]], align 8 62; CHECK-NEXT: [[TMP3:%.*]] = bitcast i8 addrspace(4)* [[V77]] to i64 addrspace(4)* 63; CHECK-NEXT: [[V80:%.*]] = addrspacecast i64 addrspace(4)* [[TMP3]] to i64* 64; CHECK-NEXT: [[V81:%.*]] = load i64, i64* [[V80]], align 8 65; CHECK-NEXT: ret i64 [[V81]] 66; 67 %.pre = load i8 addrspace(4)*, i8 addrspace(4)** %gp, align 8 68 %v74 = call i8 addrspace(4)* @alloc() 69 %v75 = addrspacecast i8 addrspace(4)* %v74 to i8* 70 %v76 = bitcast i8* %v75 to i8 addrspace(4)** 71 %v77 = getelementptr i8 addrspace(4)*, i8 addrspace(4)** %v76, i64 -1 72 store i8 addrspace(4)* %.pre, i8 addrspace(4)** %v77, align 8 73 %v80 = bitcast i8 addrspace(4)** %v77 to i64* 74 %v81 = load i64, i64* %v80, align 8 75 ret i64 %v81 76} 77 78define i64 @g2(i8* addrspace(4)* %gp) { 79; CHECK-LABEL: @g2( 80; CHECK-NEXT: [[DOTPRE:%.*]] = load i8*, i8* addrspace(4)* [[GP:%.*]], align 8 81; CHECK-NEXT: [[V74:%.*]] = call i8 addrspace(4)* @alloc() 82; CHECK-NEXT: [[V77:%.*]] = getelementptr i8, i8 addrspace(4)* [[V74]], i64 -8 83; CHECK-NEXT: [[TMP1:%.*]] = bitcast i8 addrspace(4)* [[V77]] to i8* addrspace(4)* 84; CHECK-NEXT: store i8* [[DOTPRE]], i8* addrspace(4)* [[TMP1]], align 8 85; CHECK-NEXT: [[V81_CAST:%.*]] = ptrtoint i8* [[DOTPRE]] to i64 86; CHECK-NEXT: ret i64 [[V81_CAST]] 87; 88 %.pre = load i8*, i8* addrspace(4)* %gp, align 8 89 %v74 = call i8 addrspace(4)* @alloc() 90 %v76 = bitcast i8 addrspace(4)* %v74 to i8* addrspace(4)* 91 %v77 = getelementptr i8*, i8* addrspace(4)* %v76, i64 -1 92 store i8* %.pre, i8* addrspace(4)* %v77, align 8 93 %v80 = bitcast i8* addrspace(4)* %v77 to i64 addrspace(4)* 94 %v81 = load i64, i64 addrspace(4)* %v80, align 8 95 ret i64 %v81 96} 97 98declare i8 addrspace(4)* @alloc() 99 100define i64 @f_4(i8 addrspace(4)* %v0) { 101; CHECK-LABEL: @f_4( 102; CHECK-NEXT: [[V6:%.*]] = call i64 bitcast (i64 (i64)* @f_5 to i64 (i8 addrspace(4)*)*)(i8 addrspace(4)* [[V0:%.*]]) 103; CHECK-NEXT: ret i64 [[V6]] 104; 105 %v5 = bitcast i64 (i64)* @f_5 to i64 (i8 addrspace(4)*)* 106 %v6 = call i64 %v5(i8 addrspace(4)* %v0) 107 ret i64 %v6 108} 109 110declare i64 @f_5(i64) 111