1; RUN: llc < %s -stack-symbol-ordering=0 -tailcallopt -relocation-model=static -code-model=medium -stack-alignment=8 -mtriple=x86_64-linux-gnu -mcpu=opteron | FileCheck %s 2 3; Check the HiPE calling convention works (x86-64) 4 5define void @zap(i64 %a, i64 %b) nounwind { 6entry: 7 ; CHECK: movq %rsi, %rax 8 ; CHECK-NEXT: movl $8, %ecx 9 ; CHECK-NEXT: movl $9, %r8d 10 ; CHECK-NEXT: movq %rdi, %rsi 11 ; CHECK-NEXT: movq %rax, %rdx 12 ; CHECK-NEXT: callq addfour 13 %0 = call cc 11 {i64, i64, i64} @addfour(i64 undef, i64 undef, i64 %a, i64 %b, i64 8, i64 9) 14 %res = extractvalue {i64, i64, i64} %0, 2 15 16 ; CHECK: movl $1, %edx 17 ; CHECK-NEXT: movl $2, %ecx 18 ; CHECK-NEXT: movl $3, %r8d 19 ; CHECK-NEXT: movq %rax, %r9 20 ; CHECK: callq foo 21 tail call void @foo(i64 undef, i64 undef, i64 1, i64 2, i64 3, i64 %res) nounwind 22 ret void 23} 24 25define cc 11 {i64, i64, i64} @addfour(i64 %hp, i64 %p, i64 %x, i64 %y, i64 %z, i64 %w) nounwind { 26entry: 27 ; CHECK: leaq (%rsi,%rdx), %rax 28 ; CHECK-NEXT: addq %rcx, %rax 29 ; CHECK-NEXT: addq %r8, %rax 30 %0 = add i64 %x, %y 31 %1 = add i64 %0, %z 32 %2 = add i64 %1, %w 33 34 ; CHECK: ret 35 %res = insertvalue {i64, i64, i64} undef, i64 %2, 2 36 ret {i64, i64, i64} %res 37} 38 39define cc 11 void @foo(i64 %hp, i64 %p, i64 %arg0, i64 %arg1, i64 %arg2, i64 %arg3) nounwind { 40entry: 41 ; CHECK: movq %r15, 40(%rsp) 42 ; CHECK-NEXT: movq %rbp, 32(%rsp) 43 ; CHECK-NEXT: movq %rsi, 24(%rsp) 44 ; CHECK-NEXT: movq %rdx, 16(%rsp) 45 ; CHECK-NEXT: movq %rcx, 8(%rsp) 46 ; CHECK-NEXT: movq %r8, (%rsp) 47 %hp_var = alloca i64 48 %p_var = alloca i64 49 %arg0_var = alloca i64 50 %arg1_var = alloca i64 51 %arg2_var = alloca i64 52 %arg3_var = alloca i64 53 store i64 %hp, i64* %hp_var 54 store i64 %p, i64* %p_var 55 store i64 %arg0, i64* %arg0_var 56 store i64 %arg1, i64* %arg1_var 57 store i64 %arg2, i64* %arg2_var 58 store i64 %arg3, i64* %arg3_var 59 60 ; Loads are reading values just writen from corresponding register and are therefore noops. 61 %0 = load i64, i64* %hp_var 62 %1 = load i64, i64* %p_var 63 %2 = load i64, i64* %arg0_var 64 %3 = load i64, i64* %arg1_var 65 %4 = load i64, i64* %arg2_var 66 %5 = load i64, i64* %arg3_var 67 ; CHECK: jmp bar 68 tail call cc 11 void @bar(i64 %0, i64 %1, i64 %2, i64 %3, i64 %4, i64 %5) nounwind 69 ret void 70} 71 72define cc 11 void @baz() nounwind { 73 %tmp_clos = load i64, i64* @clos 74 %tmp_clos2 = inttoptr i64 %tmp_clos to i64* 75 %indirect_call = bitcast i64* %tmp_clos2 to void (i64, i64, i64)* 76 ; CHECK: movl $42, %esi 77 ; CHECK-NEXT: jmpq *(%rax) 78 tail call cc 11 void %indirect_call(i64 undef, i64 undef, i64 42) nounwind 79 ret void 80} 81 82; Sanity-check the tail call sequence. Number of arguments was chosen as to 83; expose a bug where the tail call sequence clobbered the stack. 84define cc 11 { i64, i64, i64 } @tailcaller(i64 %hp, i64 %p) #0 { 85 ; CHECK: movl $15, %esi 86 ; CHECK-NEXT: movl $31, %edx 87 ; CHECK-NEXT: movl $47, %ecx 88 ; CHECK-NEXT: movl $63, %r8d 89 ; CHECK-NEXT: popq %rax 90 ; CHECK-NEXT: .cfi_def_cfa_offset 16 91 ; CHECK-NEXT: jmp tailcallee 92 %ret = tail call cc11 { i64, i64, i64 } @tailcallee(i64 %hp, i64 %p, i64 15, 93 i64 31, i64 47, i64 63, i64 79) #1 94 ret { i64, i64, i64 } %ret 95} 96 97!hipe.literals = !{ !0, !1, !2 } 98!0 = !{ !"P_NSP_LIMIT", i32 160 } 99!1 = !{ !"X86_LEAF_WORDS", i32 24 } 100!2 = !{ !"AMD64_LEAF_WORDS", i32 24 } 101@clos = external constant i64 102declare cc 11 void @bar(i64, i64, i64, i64, i64, i64) 103declare cc 11 { i64, i64, i64 } @tailcallee(i64, i64, i64, i64, i64, i64, i64) 104