1; RUN: llc -o - %s | FileCheck %s 2; Check that we reserve an emergency spill slot, even if we added an extra 3; CSR spill for the values used by the swiftself parameter. 4; CHECK-LABEL: func: 5; CHECK: str [[REG:x[0-9]+]], [sp, #8] 6; CHECK: add [[REG]], sp, #248 7; CHECK: str xzr, [{{\s*}}[[REG]], #32760] 8; CHECK: ldr [[REG]], [sp, #8] 9target triple = "arm64-apple-ios" 10 11@ptr8 = external global i8* 12@ptr64 = external global i64 13 14define hidden swiftcc void @func(i8* swiftself %arg) #0 { 15bb: 16 %stack0 = alloca i8*, i32 5000, align 8 17 %stack1 = alloca i8*, i32 32, align 8 18 19 %v0 = load volatile i64, i64* @ptr64, align 8 20 %v1 = load volatile i64, i64* @ptr64, align 8 21 %v2 = load volatile i64, i64* @ptr64, align 8 22 %v3 = load volatile i64, i64* @ptr64, align 8 23 %v4 = load volatile i64, i64* @ptr64, align 8 24 %v5 = load volatile i64, i64* @ptr64, align 8 25 %v6 = load volatile i64, i64* @ptr64, align 8 26 %v7 = load volatile i64, i64* @ptr64, align 8 27 %v8 = load volatile i64, i64* @ptr64, align 8 28 %v9 = load volatile i64, i64* @ptr64, align 8 29 %v10 = load volatile i64, i64* @ptr64, align 8 30 %v11 = load volatile i64, i64* @ptr64, align 8 31 %v12 = load volatile i64, i64* @ptr64, align 8 32 %v13 = load volatile i64, i64* @ptr64, align 8 33 %v14 = load volatile i64, i64* @ptr64, align 8 34 %v15 = load volatile i64, i64* @ptr64, align 8 35 %v16 = load volatile i64, i64* @ptr64, align 8 36 %v17 = load volatile i64, i64* @ptr64, align 8 37 %v18 = load volatile i64, i64* @ptr64, align 8 38 %v19 = load volatile i64, i64* @ptr64, align 8 39 %v20 = load volatile i64, i64* @ptr64, align 8 40 %v21 = load volatile i64, i64* @ptr64, align 8 41 %v22 = load volatile i64, i64* @ptr64, align 8 42 %v23 = load volatile i64, i64* @ptr64, align 8 43 %v24 = load volatile i64, i64* @ptr64, align 8 44 %v25 = load volatile i64, i64* @ptr64, align 8 45 46 ; this should exceed stack-relative addressing limits and need an emergency 47 ; spill slot. 48 %s = getelementptr inbounds i8*, i8** %stack0, i64 4092 49 store volatile i8* null, i8** %s 50 store volatile i8* null, i8** %stack1 51 52 store volatile i64 %v0, i64* @ptr64, align 8 53 store volatile i64 %v1, i64* @ptr64, align 8 54 store volatile i64 %v2, i64* @ptr64, align 8 55 store volatile i64 %v3, i64* @ptr64, align 8 56 store volatile i64 %v4, i64* @ptr64, align 8 57 store volatile i64 %v5, i64* @ptr64, align 8 58 store volatile i64 %v6, i64* @ptr64, align 8 59 store volatile i64 %v7, i64* @ptr64, align 8 60 store volatile i64 %v8, i64* @ptr64, align 8 61 store volatile i64 %v9, i64* @ptr64, align 8 62 store volatile i64 %v10, i64* @ptr64, align 8 63 store volatile i64 %v11, i64* @ptr64, align 8 64 store volatile i64 %v12, i64* @ptr64, align 8 65 store volatile i64 %v13, i64* @ptr64, align 8 66 store volatile i64 %v14, i64* @ptr64, align 8 67 store volatile i64 %v15, i64* @ptr64, align 8 68 store volatile i64 %v16, i64* @ptr64, align 8 69 store volatile i64 %v17, i64* @ptr64, align 8 70 store volatile i64 %v18, i64* @ptr64, align 8 71 store volatile i64 %v19, i64* @ptr64, align 8 72 store volatile i64 %v20, i64* @ptr64, align 8 73 store volatile i64 %v21, i64* @ptr64, align 8 74 store volatile i64 %v22, i64* @ptr64, align 8 75 store volatile i64 %v23, i64* @ptr64, align 8 76 store volatile i64 %v24, i64* @ptr64, align 8 77 store volatile i64 %v25, i64* @ptr64, align 8 78 79 ; use swiftself parameter late so it stays alive throughout the function. 80 store volatile i8* %arg, i8** @ptr8 81 ret void 82} 83