1; RUN: llc < %s -mtriple=arm64-apple-darwin | FileCheck %s 2 3define i8 @test_64bit_add(i16* %a, i64 %b) { 4; CHECK-LABEL: test_64bit_add: 5; CHECK: lsl [[REG:x[0-9]+]], x1, #1 6; CHECK: ldrb w0, [x0, [[REG]]] 7; CHECK: ret 8 %tmp1 = getelementptr inbounds i16, i16* %a, i64 %b 9 %tmp2 = load i16, i16* %tmp1 10 %tmp3 = trunc i16 %tmp2 to i8 11 ret i8 %tmp3 12} 13 14; These tests are trying to form SEXT and ZEXT operations that never leave i64 15; space, to make sure LLVM can adapt the offset register correctly. 16define void @ldst_8bit(i8* %base, i64 %offset) minsize { 17; CHECK-LABEL: ldst_8bit: 18 19 %off32.sext.tmp = shl i64 %offset, 32 20 %off32.sext = ashr i64 %off32.sext.tmp, 32 21 %addr8_sxtw = getelementptr i8, i8* %base, i64 %off32.sext 22 %val8_sxtw = load volatile i8, i8* %addr8_sxtw 23 %val32_signed = sext i8 %val8_sxtw to i32 24 store volatile i32 %val32_signed, i32* @var_32bit 25; CHECK: ldrsb {{w[0-9]+}}, [{{x[0-9]+}}, {{w[0-9]+}}, sxtw] 26 27 %addrint_uxtw = ptrtoint i8* %base to i64 28 %offset_uxtw = and i64 %offset, 4294967295 29 %addrint1_uxtw = add i64 %addrint_uxtw, %offset_uxtw 30 %addr_uxtw = inttoptr i64 %addrint1_uxtw to i8* 31 %val8_uxtw = load volatile i8, i8* %addr_uxtw 32 %newval8 = add i8 %val8_uxtw, 1 33 store volatile i8 %newval8, i8* @var_8bit 34; CHECK: ldrb {{w[0-9]+}}, [{{x[0-9]+}}, {{w[0-9]+}}, uxtw] 35 36 ret void 37} 38 39 40define void @ldst_16bit(i16* %base, i64 %offset) minsize { 41; CHECK-LABEL: ldst_16bit: 42 43 %addrint_uxtw = ptrtoint i16* %base to i64 44 %offset_uxtw = and i64 %offset, 4294967295 45 %addrint1_uxtw = add i64 %addrint_uxtw, %offset_uxtw 46 %addr_uxtw = inttoptr i64 %addrint1_uxtw to i16* 47 %val8_uxtw = load volatile i16, i16* %addr_uxtw 48 %newval8 = add i16 %val8_uxtw, 1 49 store volatile i16 %newval8, i16* @var_16bit 50; CHECK: ldrh {{w[0-9]+}}, [{{x[0-9]+}}, {{w[0-9]+}}, uxtw] 51 52 %base_sxtw = ptrtoint i16* %base to i64 53 %offset_sxtw.tmp = shl i64 %offset, 32 54 %offset_sxtw = ashr i64 %offset_sxtw.tmp, 32 55 %addrint_sxtw = add i64 %base_sxtw, %offset_sxtw 56 %addr_sxtw = inttoptr i64 %addrint_sxtw to i16* 57 %val16_sxtw = load volatile i16, i16* %addr_sxtw 58 %val64_signed = sext i16 %val16_sxtw to i64 59 store volatile i64 %val64_signed, i64* @var_64bit 60; CHECK: ldrsh {{x[0-9]+}}, [{{x[0-9]+}}, {{w[0-9]+}}, sxtw] 61 62 63 %base_uxtwN = ptrtoint i16* %base to i64 64 %offset_uxtwN = and i64 %offset, 4294967295 65 %offset2_uxtwN = shl i64 %offset_uxtwN, 1 66 %addrint_uxtwN = add i64 %base_uxtwN, %offset2_uxtwN 67 %addr_uxtwN = inttoptr i64 %addrint_uxtwN to i16* 68 %val32 = load volatile i32, i32* @var_32bit 69 %val16_trunc32 = trunc i32 %val32 to i16 70 store volatile i16 %val16_trunc32, i16* %addr_uxtwN 71; CHECK: strh {{w[0-9]+}}, [{{x[0-9]+}}, {{w[0-9]+}}, uxtw #1] 72 ret void 73} 74 75define void @ldst_32bit(i32* %base, i64 %offset) minsize { 76; CHECK-LABEL: ldst_32bit: 77 78 %addrint_uxtw = ptrtoint i32* %base to i64 79 %offset_uxtw = and i64 %offset, 4294967295 80 %addrint1_uxtw = add i64 %addrint_uxtw, %offset_uxtw 81 %addr_uxtw = inttoptr i64 %addrint1_uxtw to i32* 82 %val32_uxtw = load volatile i32, i32* %addr_uxtw 83 %newval32 = add i32 %val32_uxtw, 1 84 store volatile i32 %newval32, i32* @var_32bit 85; CHECK: ldr {{w[0-9]+}}, [{{x[0-9]+}}, {{w[0-9]+}}, uxtw] 86 87 %base_sxtw = ptrtoint i32* %base to i64 88 %offset_sxtw.tmp = shl i64 %offset, 32 89 %offset_sxtw = ashr i64 %offset_sxtw.tmp, 32 90 %addrint_sxtw = add i64 %base_sxtw, %offset_sxtw 91 %addr_sxtw = inttoptr i64 %addrint_sxtw to i32* 92 %val32_sxtw = load volatile i32, i32* %addr_sxtw 93 %val64_signed = sext i32 %val32_sxtw to i64 94 store volatile i64 %val64_signed, i64* @var_64bit 95; CHECK: ldrsw {{x[0-9]+}}, [{{x[0-9]+}}, {{w[0-9]+}}, sxtw] 96 97 98 %base_uxtwN = ptrtoint i32* %base to i64 99 %offset_uxtwN = and i64 %offset, 4294967295 100 %offset2_uxtwN = shl i64 %offset_uxtwN, 2 101 %addrint_uxtwN = add i64 %base_uxtwN, %offset2_uxtwN 102 %addr_uxtwN = inttoptr i64 %addrint_uxtwN to i32* 103 %val32 = load volatile i32, i32* @var_32bit 104 store volatile i32 %val32, i32* %addr_uxtwN 105; CHECK: str {{w[0-9]+}}, [{{x[0-9]+}}, {{w[0-9]+}}, uxtw #2] 106 ret void 107} 108 109define void @ldst_64bit(i64* %base, i64 %offset) minsize { 110; CHECK-LABEL: ldst_64bit: 111 112 %addrint_uxtw = ptrtoint i64* %base to i64 113 %offset_uxtw = and i64 %offset, 4294967295 114 %addrint1_uxtw = add i64 %addrint_uxtw, %offset_uxtw 115 %addr_uxtw = inttoptr i64 %addrint1_uxtw to i64* 116 %val64_uxtw = load volatile i64, i64* %addr_uxtw 117 %newval8 = add i64 %val64_uxtw, 1 118 store volatile i64 %newval8, i64* @var_64bit 119; CHECK: ldr {{x[0-9]+}}, [{{x[0-9]+}}, {{w[0-9]+}}, uxtw] 120 121 %base_sxtw = ptrtoint i64* %base to i64 122 %offset_sxtw.tmp = shl i64 %offset, 32 123 %offset_sxtw = ashr i64 %offset_sxtw.tmp, 32 124 %addrint_sxtw = add i64 %base_sxtw, %offset_sxtw 125 %addr_sxtw = inttoptr i64 %addrint_sxtw to i64* 126 %val64_sxtw = load volatile i64, i64* %addr_sxtw 127 store volatile i64 %val64_sxtw, i64* @var_64bit 128; CHECK: ldr {{x[0-9]+}}, [{{x[0-9]+}}, {{w[0-9]+}}, sxtw] 129 130 131 %base_uxtwN = ptrtoint i64* %base to i64 132 %offset_uxtwN = and i64 %offset, 4294967295 133 %offset2_uxtwN = shl i64 %offset_uxtwN, 3 134 %addrint_uxtwN = add i64 %base_uxtwN, %offset2_uxtwN 135 %addr_uxtwN = inttoptr i64 %addrint_uxtwN to i64* 136 %val64 = load volatile i64, i64* @var_64bit 137 store volatile i64 %val64, i64* %addr_uxtwN 138; CHECK: str {{x[0-9]+}}, [{{x[0-9]+}}, {{w[0-9]+}}, uxtw #3] 139 ret void 140} 141 142@var_8bit = global i8 0 143@var_16bit = global i16 0 144@var_32bit = global i32 0 145@var_64bit = global i64 0 146