llvm/llvm/test/CodeGen/X86/segmented-stacks-dynamic.ll

; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc < %s -mcpu=generic -mtriple=i686-linux -verify-machineinstrs | FileCheck %s -check-prefix=X86
; RUN: llc < %s -mcpu=generic -mtriple=x86_64-linux -verify-machineinstrs | FileCheck %s -check-prefix=X64
; RUN: llc < %s -mcpu=generic -mtriple=x86_64-linux-gnux32 -verify-machineinstrs | FileCheck %s -check-prefix=X32ABI
; RUN: llc < %s -mcpu=generic -mtriple=i686-linux -filetype=obj
; RUN: llc < %s -mcpu=generic -mtriple=x86_64-linux -filetype=obj
; RUN: llc < %s -mcpu=generic -mtriple=x86_64-linux-gnux32 -filetype=obj

; Just to prevent the alloca from being optimized away
declare void @dummy_use(ptr, i32)

define i32 @test_basic(i32 %l) #0 {
; X86-LABEL: test_basic:
; X86:       # %bb.0:
; X86-NEXT:    cmpl %gs:48, %esp
; X86-NEXT:    jbe .LBB0_1
; X86-NEXT:  .LBB0_2:
; X86-NEXT:    pushl %ebp
; X86-NEXT:    .cfi_def_cfa_offset 8
; X86-NEXT:    .cfi_offset %ebp, -8
; X86-NEXT:    movl %esp, %ebp
; X86-NEXT:    .cfi_def_cfa_register %ebp
; X86-NEXT:    pushl %esi
; X86-NEXT:    pushl %eax
; X86-NEXT:    .cfi_offset %esi, -12
; X86-NEXT:    movl 8(%ebp), %esi
; X86-NEXT:    leal 15(,%esi,4), %ecx
; X86-NEXT:    andl $-16, %ecx
; X86-NEXT:    movl %esp, %eax
; X86-NEXT:    subl %ecx, %eax
; X86-NEXT:    cmpl %eax, %gs:48
; X86-NEXT:    jg .LBB0_4
; X86-NEXT:  # %bb.3:
; X86-NEXT:    movl %eax, %esp
; X86-NEXT:    jmp .LBB0_5
; X86-NEXT:  .LBB0_4:
; X86-NEXT:    subl $12, %esp
; X86-NEXT:    pushl %ecx
; X86-NEXT:    calll __morestack_allocate_stack_space
; X86-NEXT:    addl $16, %esp
; X86-NEXT:  .LBB0_5:
; X86-NEXT:    subl $8, %esp
; X86-NEXT:    pushl %esi
; X86-NEXT:    pushl %eax
; X86-NEXT:    calll dummy_use@PLT
; X86-NEXT:    addl $16, %esp
; X86-NEXT:    testl %esi, %esi
; X86-NEXT:    je .LBB0_6
; X86-NEXT:  # %bb.8: # %false
; X86-NEXT:    decl %esi
; X86-NEXT:    subl $12, %esp
; X86-NEXT:    pushl %esi
; X86-NEXT:    calll test_basic@PLT
; X86-NEXT:    jmp .LBB0_7
; X86-NEXT:  .LBB0_6: # %true
; X86-NEXT:    xorl %eax, %eax
; X86-NEXT:  .LBB0_7: # %true
; X86-NEXT:    leal -4(%ebp), %esp
; X86-NEXT:    popl %esi
; X86-NEXT:    popl %ebp
; X86-NEXT:    .cfi_def_cfa %esp, 4
; X86-NEXT:    retl
; X86-NEXT:  .LBB0_1:
; X86-NEXT:    .cfi_restore %ebp
; X86-NEXT:    .cfi_restore %esi
; X86-NEXT:    pushl $4
; X86-NEXT:    pushl $12
; X86-NEXT:    calll __morestack
; X86-NEXT:    retl
; X86-NEXT:    jmp .LBB0_2
;
; X64-LABEL: test_basic:
; X64:       # %bb.0:
; X64-NEXT:    cmpq %fs:112, %rsp
; X64-NEXT:    jbe .LBB0_1
; X64-NEXT:  .LBB0_2:
; X64-NEXT:    pushq %rbp
; X64-NEXT:    .cfi_def_cfa_offset 16
; X64-NEXT:    .cfi_offset %rbp, -16
; X64-NEXT:    movq %rsp, %rbp
; X64-NEXT:    .cfi_def_cfa_register %rbp
; X64-NEXT:    pushq %rbx
; X64-NEXT:    pushq %rax
; X64-NEXT:    .cfi_offset %rbx, -24
; X64-NEXT:    movl %edi, %ebx
; X64-NEXT:    movl %edi, %eax
; X64-NEXT:    leaq 15(,%rax,4), %rax
; X64-NEXT:    andq $-16, %rax
; X64-NEXT:    movq %rsp, %rdi
; X64-NEXT:    subq %rax, %rdi
; X64-NEXT:    cmpq %rdi, %fs:112
; X64-NEXT:    jg .LBB0_4
; X64-NEXT:  # %bb.3:
; X64-NEXT:    movq %rdi, %rsp
; X64-NEXT:    jmp .LBB0_5
; X64-NEXT:  .LBB0_4:
; X64-NEXT:    movq %rax, %rdi
; X64-NEXT:    callq __morestack_allocate_stack_space
; X64-NEXT:    movq %rax, %rdi
; X64-NEXT:  .LBB0_5:
; X64-NEXT:    movl %ebx, %esi
; X64-NEXT:    callq dummy_use@PLT
; X64-NEXT:    testl %ebx, %ebx
; X64-NEXT:    je .LBB0_6
; X64-NEXT:  # %bb.8: # %false
; X64-NEXT:    decl %ebx
; X64-NEXT:    movl %ebx, %edi
; X64-NEXT:    callq test_basic@PLT
; X64-NEXT:    jmp .LBB0_7
; X64-NEXT:  .LBB0_6: # %true
; X64-NEXT:    xorl %eax, %eax
; X64-NEXT:  .LBB0_7: # %true
; X64-NEXT:    leaq -8(%rbp), %rsp
; X64-NEXT:    popq %rbx
; X64-NEXT:    popq %rbp
; X64-NEXT:    .cfi_def_cfa %rsp, 8
; X64-NEXT:    retq
; X64-NEXT:  .LBB0_1:
; X64-NEXT:    .cfi_restore %rbx
; X64-NEXT:    .cfi_restore %rbp
; X64-NEXT:    movl $24, %r10d
; X64-NEXT:    movl $0, %r11d
; X64-NEXT:    callq __morestack
; X64-NEXT:    retq
; X64-NEXT:    jmp .LBB0_2
;
; X32ABI-LABEL: test_basic:
; X32ABI:       # %bb.0:
; X32ABI-NEXT:    cmpl %fs:64, %esp
; X32ABI-NEXT:    jbe .LBB0_1
; X32ABI-NEXT:  .LBB0_2:
; X32ABI-NEXT:    pushq %rbp
; X32ABI-NEXT:    .cfi_def_cfa_offset 16
; X32ABI-NEXT:    .cfi_offset %rbp, -16
; X32ABI-NEXT:    movl %esp, %ebp
; X32ABI-NEXT:    .cfi_def_cfa_register %rbp
; X32ABI-NEXT:    pushq %rbx
; X32ABI-NEXT:    pushq %rax
; X32ABI-NEXT:    .cfi_offset %rbx, -24
; X32ABI-NEXT:    movl %edi, %ebx
; X32ABI-NEXT:    leal 15(,%rbx,4), %eax
; X32ABI-NEXT:    andl $-16, %eax
; X32ABI-NEXT:    movl %esp, %edi
; X32ABI-NEXT:    subl %eax, %edi
; X32ABI-NEXT:    cmpl %edi, %fs:64
; X32ABI-NEXT:    jg .LBB0_4
; X32ABI-NEXT:  # %bb.3:
; X32ABI-NEXT:    movl %edi, %esp
; X32ABI-NEXT:    jmp .LBB0_5
; X32ABI-NEXT:  .LBB0_4:
; X32ABI-NEXT:    movl %eax, %edi
; X32ABI-NEXT:    callq __morestack_allocate_stack_space
; X32ABI-NEXT:    movl %eax, %edi
; X32ABI-NEXT:  .LBB0_5:
; X32ABI-NEXT:    movl %ebx, %esi
; X32ABI-NEXT:    callq dummy_use@PLT
; X32ABI-NEXT:    testl %ebx, %ebx
; X32ABI-NEXT:    je .LBB0_6
; X32ABI-NEXT:  # %bb.8: # %false
; X32ABI-NEXT:    decl %ebx
; X32ABI-NEXT:    movl %ebx, %edi
; X32ABI-NEXT:    callq test_basic@PLT
; X32ABI-NEXT:    jmp .LBB0_7
; X32ABI-NEXT:  .LBB0_6: # %true
; X32ABI-NEXT:    xorl %eax, %eax
; X32ABI-NEXT:  .LBB0_7: # %true
; X32ABI-NEXT:    leal -8(%ebp), %esp
; X32ABI-NEXT:    popq %rbx
; X32ABI-NEXT:    popq %rbp
; X32ABI-NEXT:    .cfi_def_cfa %rsp, 8
; X32ABI-NEXT:    retq
; X32ABI-NEXT:  .LBB0_1:
; X32ABI-NEXT:    .cfi_def_cfa_register 4294967294
; X32ABI-NEXT:    .cfi_restore %rbx
; X32ABI-NEXT:    .cfi_restore %rbp
; X32ABI-NEXT:    movl $24, %r10d
; X32ABI-NEXT:    movl $0, %r11d
; X32ABI-NEXT:    callq __morestack
; X32ABI-NEXT:    retq
; X32ABI-NEXT:    jmp .LBB0_2
        %mem = alloca i32, i32 %l
        call void @dummy_use (ptr %mem, i32 %l)
        %terminate = icmp eq i32 %l, 0
        br i1 %terminate, label %true, label %false

true:
        ret i32 0

false:
        %newlen = sub i32 %l, 1
        %retvalue = call i32 @test_basic(i32 %newlen)
        ret i32 %retvalue



















}

attributes #0 = { "split-stack" }