1 ; RUN: llc < %s -mtriple=x86_64-apple-darwin -mcpu=corei7 | FileCheck --check-prefix=SSE %s 2 ; RUN: llc < %s -mtriple=x86_64-apple-darwin -mcpu=corei7-avx | FileCheck --check-prefix=AVX %s 3 4 define preserve_allcc void @preserve_allcc1() nounwind { 5 entry: 6 ;SSE-LABEL: preserve_allcc1 7 ;SSE: pushq %r10 8 ;SSE-NEXT: pushq %r9 9 ;SSE-NEXT: pushq %r8 10 ;SSE-NEXT: pushq %rdi 11 ;SSE-NEXT: pushq %rsi 12 ;SSE-NEXT: pushq %rdx 13 ;SSE-NEXT: pushq %rcx 14 ;SSE-NEXT: pushq %rax 15 ;SSE-NEXT: pushq %rbp 16 ;SSE-NEXT: pushq %r15 17 ;SSE-NEXT: pushq %r14 18 ;SSE-NEXT: pushq %r13 19 ;SSE-NEXT: pushq %r12 20 ;SSE-NEXT: pushq %rbx 21 ;SSE: movaps %xmm15 22 ;SSE-NEXT: movaps %xmm14 23 ;SSE-NEXT: movaps %xmm13 24 ;SSE-NEXT: movaps %xmm12 25 ;SSE-NEXT: movaps %xmm11 26 ;SSE-NEXT: movaps %xmm10 27 ;SSE-NEXT: movaps %xmm9 28 ;SSE-NEXT: movaps %xmm8 29 ;SSE-NEXT: movaps %xmm7 30 ;SSE-NEXT: movaps %xmm6 31 ;SSE-NEXT: movaps %xmm5 32 ;SSE-NEXT: movaps %xmm4 33 ;SSE-NEXT: movaps %xmm3 34 ;SSE-NEXT: movaps %xmm2 35 ;SSE-NEXT: movaps %xmm1 36 ;SSE-NEXT: movaps %xmm0 37 ;AVX-LABEL: preserve_allcc1 38 ;AVX: pushq %r10 39 ;AVX-NEXT: pushq %r9 40 ;AVX-NEXT: pushq %r8 41 ;AVX-NEXT: pushq %rdi 42 ;AVX-NEXT: pushq %rsi 43 ;AVX-NEXT: pushq %rdx 44 ;AVX-NEXT: pushq %rcx 45 ;AVX-NEXT: pushq %rax 46 ;AVX-NEXT: pushq %rbp 47 ;AVX-NEXT: pushq %r15 48 ;AVX-NEXT: pushq %r14 49 ;AVX-NEXT: pushq %r13 50 ;AVX-NEXT: pushq %r12 51 ;AVX-NEXT: pushq %rbx 52 ;AVX: vmovups %ymm15 53 ;AVX-NEXT: vmovups %ymm14 54 ;AVX-NEXT: vmovups %ymm13 55 ;AVX-NEXT: vmovups %ymm12 56 ;AVX-NEXT: vmovups %ymm11 57 ;AVX-NEXT: vmovups %ymm10 58 ;AVX-NEXT: vmovups %ymm9 59 ;AVX-NEXT: vmovups %ymm8 60 ;AVX-NEXT: vmovups %ymm7 61 ;AVX-NEXT: vmovups %ymm6 62 ;AVX-NEXT: vmovups %ymm5 63 ;AVX-NEXT: vmovups %ymm4 64 ;AVX-NEXT: vmovups %ymm3 65 ;AVX-NEXT: vmovups %ymm2 66 ;AVX-NEXT: vmovups %ymm1 67 ;AVX-NEXT: vmovups %ymm0 68 call void asm sideeffect "", "~{rax},~{rbx},~{rcx},~{rdx},~{rsi},~{rdi},~{r8},~{r9},~{r10},~{r11},~{r12},~{r13},~{r14},~{r15},~{rbp},~{xmm0},~{xmm1},~{xmm2},~{xmm3},~{xmm4},~{xmm5},~{xmm6},~{xmm7},~{xmm8},~{xmm9},~{xmm10},~{xmm11},~{xmm12},~{xmm13},~{xmm14},~{xmm15}"() 69 ret void 70 } 71 72 ; Make sure only R11 is saved before the call 73 declare preserve_allcc void @bar(i64, i64, double, double) 74 define void @preserve_allcc2() nounwind { 75 entry: 76 ;SSE-LABEL: preserve_allcc2 77 ;SSE: movq %r11, [[REG:%[a-z0-9]+]] 78 ;SSE-NOT: movaps %xmm 79 ;SSE: movq [[REG]], %r11 80 %a0 = call i64 asm sideeffect "", "={rax}"() nounwind 81 %a1 = call i64 asm sideeffect "", "={rcx}"() nounwind 82 %a2 = call i64 asm sideeffect "", "={rdx}"() nounwind 83 %a3 = call i64 asm sideeffect "", "={r8}"() nounwind 84 %a4 = call i64 asm sideeffect "", "={r9}"() nounwind 85 %a5 = call i64 asm sideeffect "", "={r10}"() nounwind 86 %a6 = call i64 asm sideeffect "", "={r11}"() nounwind 87 %a10 = call <2 x double> asm sideeffect "", "={xmm2}"() nounwind 88 %a11 = call <2 x double> asm sideeffect "", "={xmm3}"() nounwind 89 %a12 = call <2 x double> asm sideeffect "", "={xmm4}"() nounwind 90 %a13 = call <2 x double> asm sideeffect "", "={xmm5}"() nounwind 91 %a14 = call <2 x double> asm sideeffect "", "={xmm6}"() nounwind 92 %a15 = call <2 x double> asm sideeffect "", "={xmm7}"() nounwind 93 %a16 = call <2 x double> asm sideeffect "", "={xmm8}"() nounwind 94 %a17 = call <2 x double> asm sideeffect "", "={xmm9}"() nounwind 95 %a18 = call <2 x double> asm sideeffect "", "={xmm10}"() nounwind 96 %a19 = call <2 x double> asm sideeffect "", "={xmm11}"() nounwind 97 %a20 = call <2 x double> asm sideeffect "", "={xmm12}"() nounwind 98 %a21 = call <2 x double> asm sideeffect "", "={xmm13}"() nounwind 99 %a22 = call <2 x double> asm sideeffect "", "={xmm14}"() nounwind 100 %a23 = call <2 x double> asm sideeffect "", "={xmm15}"() nounwind 101 call preserve_allcc void @bar(i64 1, i64 2, double 3.0, double 4.0) 102 call void asm sideeffect "", "{rax},{rcx},{rdx},{r8},{r9},{r10},{r11},{xmm2},{xmm3},{xmm4},{xmm5},{xmm6},{xmm7},{xmm8},{xmm9},{xmm10},{xmm11},{xmm12},{xmm13},{xmm14},{xmm15}"(i64 %a0, i64 %a1, i64 %a2, i64 %a3, i64 %a4, i64 %a5, i64 %a6, <2 x double> %a10, <2 x double> %a11, <2 x double> %a12, <2 x double> %a13, <2 x double> %a14, <2 x double> %a15, <2 x double> %a16, <2 x double> %a17, <2 x double> %a18, <2 x double> %a19, <2 x double> %a20, <2 x double> %a21, <2 x double> %a22, <2 x double> %a23) 103 ret void 104 } 105