Home | History | Annotate | Download | only in X86
      1 ; RUN: llc < %s -mtriple=i686-linux -segmented-stacks | FileCheck %s -check-prefix=X32
      2 ; RUN: llc < %s -mtriple=x86_64-linux  -segmented-stacks | FileCheck %s -check-prefix=X64
      3 
      4 ; Just to prevent the alloca from being optimized away
      5 declare void @dummy_use(i32*, i32)
      6 
      7 define i32 @test_basic(i32 %l) {
      8         %mem = alloca i32, i32 %l
      9         call void @dummy_use (i32* %mem, i32 %l)
     10         %terminate = icmp eq i32 %l, 0
     11         br i1 %terminate, label %true, label %false
     12 
     13 true:
     14         ret i32 0
     15 
     16 false:
     17         %newlen = sub i32 %l, 1
     18         %retvalue = call i32 @test_basic(i32 %newlen)
     19         ret i32 %retvalue
     20 
     21 ; X32:      test_basic:
     22 
     23 ; X32:      leal -12(%esp), %ecx
     24 ; X32-NEXT: cmpl %gs:48, %ecx
     25 
     26 ; X32:      subl $8, %esp
     27 ; X32-NEXT: pushl $4
     28 ; X32-NEXT: pushl $12
     29 ; X32-NEXT: calll __morestack
     30 ; X32-NEXT: addl $8, %esp
     31 ; X32-NEXT: ret 
     32 
     33 ; X32:      movl %eax, %esp
     34 
     35 ; X32:      subl $12, %esp
     36 ; X32-NEXT: pushl %ecx
     37 ; X32-NEXT: calll __morestack_allocate_stack_space
     38 ; X32-NEXT: addl $16, %esp
     39 
     40 ; X64:      test_basic:
     41 
     42 ; X64:      leaq -24(%rsp), %r11
     43 ; X64-NEXT: cmpq %fs:112, %r11
     44 
     45 ; X64:      movabsq $24, %r10
     46 ; X64-NEXT: movabsq $0, %r11
     47 ; X64-NEXT: callq __morestack
     48 ; X64-NEXT: ret
     49 
     50 ; X64:      movq %rsp, %rax
     51 ; X64-NEXT: subq %rcx, %rax
     52 ; X64-NEXT: cmpq %rax, %fs:112
     53 
     54 ; X64:      movq %rax, %rsp
     55 
     56 ; X64:      movq %rcx, %rdi
     57 ; X64-NEXT: callq __morestack_allocate_stack_space
     58 
     59 }
     60 
     61 define i32 @test_nested(i32 * nest %closure, i32 %other) {
     62        %addend = load i32 * %closure
     63        %result = add i32 %other, %addend
     64        ret i32 %result
     65 
     66 ; X32:      leal (%esp), %edx
     67 ; X32-NEXT: cmpl %gs:48, %edx
     68 
     69 
     70 ; X32:      subl $8, %esp
     71 ; X32-NEXT: pushl $4
     72 ; X32-NEXT: pushl $0
     73 ; X32-NEXT: calll __morestack
     74 ; X32-NEXT: addl $8, %esp
     75 ; X32-NEXT: ret
     76 
     77 ; X64:      leaq (%rsp), %r11
     78 ; X64-NEXT: cmpq %fs:112, %r11
     79 
     80 ; X64:      movq %r10, %rax
     81 ; X64-NEXT: movabsq $0, %r10
     82 ; X64-NEXT: movabsq $0, %r11
     83 ; X64-NEXT: callq __morestack
     84 ; X64-NEXT: ret
     85 ; X64:      movq %rax, %r10
     86 
     87 }
     88