Home | History | Annotate | Download | only in X86
      1 ; RUN: llc < %s -mtriple=x86_64-apple-darwin -mcpu=corei7     | FileCheck --check-prefix=SSE %s
      2 ; RUN: llc < %s -mtriple=x86_64-apple-darwin -mcpu=corei7-avx | FileCheck --check-prefix=AVX %s
      3 
      4 define preserve_allcc void @preserve_allcc1() nounwind {
      5 entry:
      6 ;SSE-LABEL: preserve_allcc1
      7 ;SSE:       pushq %r10
      8 ;SSE-NEXT:  pushq %r9
      9 ;SSE-NEXT:  pushq %r8
     10 ;SSE-NEXT:  pushq %rdi
     11 ;SSE-NEXT:  pushq %rsi
     12 ;SSE-NEXT:  pushq %rdx
     13 ;SSE-NEXT:  pushq %rcx
     14 ;SSE-NEXT:  pushq %rax
     15 ;SSE-NEXT:  pushq %rbp
     16 ;SSE-NEXT:  pushq %r15
     17 ;SSE-NEXT:  pushq %r14
     18 ;SSE-NEXT:  pushq %r13
     19 ;SSE-NEXT:  pushq %r12
     20 ;SSE-NEXT:  pushq %rbx
     21 ;SSE:       movaps %xmm15
     22 ;SSE-NEXT:  movaps %xmm14
     23 ;SSE-NEXT:  movaps %xmm13
     24 ;SSE-NEXT:  movaps %xmm12
     25 ;SSE-NEXT:  movaps %xmm11
     26 ;SSE-NEXT:  movaps %xmm10
     27 ;SSE-NEXT:  movaps %xmm9
     28 ;SSE-NEXT:  movaps %xmm8
     29 ;SSE-NEXT:  movaps %xmm7
     30 ;SSE-NEXT:  movaps %xmm6
     31 ;SSE-NEXT:  movaps %xmm5
     32 ;SSE-NEXT:  movaps %xmm4
     33 ;SSE-NEXT:  movaps %xmm3
     34 ;SSE-NEXT:  movaps %xmm2
     35 ;SSE-NEXT:  movaps %xmm1
     36 ;SSE-NEXT:  movaps %xmm0
     37 ;AVX-LABEL: preserve_allcc1
     38 ;AVX:       pushq %r10
     39 ;AVX-NEXT:  pushq %r9
     40 ;AVX-NEXT:  pushq %r8
     41 ;AVX-NEXT:  pushq %rdi
     42 ;AVX-NEXT:  pushq %rsi
     43 ;AVX-NEXT:  pushq %rdx
     44 ;AVX-NEXT:  pushq %rcx
     45 ;AVX-NEXT:  pushq %rax
     46 ;AVX-NEXT:  pushq %rbp
     47 ;AVX-NEXT:  pushq %r15
     48 ;AVX-NEXT:  pushq %r14
     49 ;AVX-NEXT:  pushq %r13
     50 ;AVX-NEXT:  pushq %r12
     51 ;AVX-NEXT:  pushq %rbx
     52 ;AVX:       vmovups %ymm15
     53 ;AVX-NEXT:  vmovups %ymm14
     54 ;AVX-NEXT:  vmovups %ymm13
     55 ;AVX-NEXT:  vmovups %ymm12
     56 ;AVX-NEXT:  vmovups %ymm11
     57 ;AVX-NEXT:  vmovups %ymm10
     58 ;AVX-NEXT:  vmovups %ymm9
     59 ;AVX-NEXT:  vmovups %ymm8
     60 ;AVX-NEXT:  vmovups %ymm7
     61 ;AVX-NEXT:  vmovups %ymm6
     62 ;AVX-NEXT:  vmovups %ymm5
     63 ;AVX-NEXT:  vmovups %ymm4
     64 ;AVX-NEXT:  vmovups %ymm3
     65 ;AVX-NEXT:  vmovups %ymm2
     66 ;AVX-NEXT:  vmovups %ymm1
     67 ;AVX-NEXT:  vmovups %ymm0
     68   call void asm sideeffect "", "~{rax},~{rbx},~{rcx},~{rdx},~{rsi},~{rdi},~{r8},~{r9},~{r10},~{r11},~{r12},~{r13},~{r14},~{r15},~{rbp},~{xmm0},~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15}"()
     69   ret void
     70 }
     71 
     72 ; Make sure only R11 is saved before the call
     73 declare preserve_allcc void @bar(i64, i64, double, double)
     74 define void @preserve_allcc2() nounwind {
     75 entry:
     76 ;SSE-LABEL: preserve_allcc2
     77 ;SSE:       movq %r11, [[REG:%[a-z0-9]+]]
     78 ;SSE-NOT:   movaps %xmm
     79 ;SSE:       movq [[REG]], %r11
     80   %a0 = call i64 asm sideeffect "", "={rax}"() nounwind
     81   %a1 = call i64 asm sideeffect "", "={rcx}"() nounwind
     82   %a2 = call i64 asm sideeffect "", "={rdx}"() nounwind
     83   %a3 = call i64 asm sideeffect "", "={r8}"() nounwind
     84   %a4 = call i64 asm sideeffect "", "={r9}"() nounwind
     85   %a5 = call i64 asm sideeffect "", "={r10}"() nounwind
     86   %a6 = call i64 asm sideeffect "", "={r11}"() nounwind
     87   %a10 = call <2 x double> asm sideeffect "", "={xmm2}"() nounwind
     88   %a11 = call <2 x double> asm sideeffect "", "={xmm3}"() nounwind
     89   %a12 = call <2 x double> asm sideeffect "", "={xmm4}"() nounwind
     90   %a13 = call <2 x double> asm sideeffect "", "={xmm5}"() nounwind
     91   %a14 = call <2 x double> asm sideeffect "", "={xmm6}"() nounwind
     92   %a15 = call <2 x double> asm sideeffect "", "={xmm7}"() nounwind
     93   %a16 = call <2 x double> asm sideeffect "", "={xmm8}"() nounwind
     94   %a17 = call <2 x double> asm sideeffect "", "={xmm9}"() nounwind
     95   %a18 = call <2 x double> asm sideeffect "", "={xmm10}"() nounwind
     96   %a19 = call <2 x double> asm sideeffect "", "={xmm11}"() nounwind
     97   %a20 = call <2 x double> asm sideeffect "", "={xmm12}"() nounwind
     98   %a21 = call <2 x double> asm sideeffect "", "={xmm13}"() nounwind
     99   %a22 = call <2 x double> asm sideeffect "", "={xmm14}"() nounwind
    100   %a23 = call <2 x double> asm sideeffect "", "={xmm15}"() nounwind
    101   call preserve_allcc void @bar(i64 1, i64 2, double 3.0, double 4.0)
    102   call void asm sideeffect "", "{rax},{rcx},{rdx},{r8},{r9},{r10},{r11},{xmm2},{xmm3},{xmm4},{xmm5},{xmm6},{xmm7},{xmm8},{xmm9},{xmm10},{xmm11},{xmm12},{xmm13},{xmm14},{xmm15}"(i64 %a0, i64 %a1, i64 %a2, i64 %a3, i64 %a4, i64 %a5, i64 %a6, <2 x double> %a10, <2 x double> %a11, <2 x double> %a12, <2 x double> %a13, <2 x double> %a14, <2 x double> %a15, <2 x double> %a16, <2 x double> %a17, <2 x double> %a18, <2 x double> %a19, <2 x double> %a20, <2 x double> %a21, <2 x double> %a22, <2 x double> %a23)
    103   ret void
    104 }
    105