Home | History | Annotate | Download | only in AArch64
      1 ; RUN: llc < %s -march=arm64 -aarch64-stp-suppress=false -verify-machineinstrs -mcpu=cyclone | FileCheck %s
      2 ; RUN: llc < %s -march=arm64 -aarch64-unscaled-mem-op=true\
      3 ; RUN:   -verify-machineinstrs -mcpu=cyclone | FileCheck -check-prefix=STUR_CHK %s
      4 
      5 ; CHECK: stp_int
      6 ; CHECK: stp w0, w1, [x2]
      7 define void @stp_int(i32 %a, i32 %b, i32* nocapture %p) nounwind {
      8   store i32 %a, i32* %p, align 4
      9   %add.ptr = getelementptr inbounds i32* %p, i64 1
     10   store i32 %b, i32* %add.ptr, align 4
     11   ret void
     12 }
     13 
     14 ; CHECK: stp_long
     15 ; CHECK: stp x0, x1, [x2]
     16 define void @stp_long(i64 %a, i64 %b, i64* nocapture %p) nounwind {
     17   store i64 %a, i64* %p, align 8
     18   %add.ptr = getelementptr inbounds i64* %p, i64 1
     19   store i64 %b, i64* %add.ptr, align 8
     20   ret void
     21 }
     22 
     23 ; CHECK: stp_float
     24 ; CHECK: stp s0, s1, [x0]
     25 define void @stp_float(float %a, float %b, float* nocapture %p) nounwind {
     26   store float %a, float* %p, align 4
     27   %add.ptr = getelementptr inbounds float* %p, i64 1
     28   store float %b, float* %add.ptr, align 4
     29   ret void
     30 }
     31 
     32 ; CHECK: stp_double
     33 ; CHECK: stp d0, d1, [x0]
     34 define void @stp_double(double %a, double %b, double* nocapture %p) nounwind {
     35   store double %a, double* %p, align 8
     36   %add.ptr = getelementptr inbounds double* %p, i64 1
     37   store double %b, double* %add.ptr, align 8
     38   ret void
     39 }
     40 
     41 ; Test the load/store optimizer---combine ldurs into a ldp, if appropriate
     42 define void @stur_int(i32 %a, i32 %b, i32* nocapture %p) nounwind {
     43 ; STUR_CHK: stur_int
     44 ; STUR_CHK: stp w{{[0-9]+}}, {{w[0-9]+}}, [x{{[0-9]+}}, #-8]
     45 ; STUR_CHK-NEXT: ret
     46   %p1 = getelementptr inbounds i32* %p, i32 -1
     47   store i32 %a, i32* %p1, align 2
     48   %p2 = getelementptr inbounds i32* %p, i32 -2
     49   store i32 %b, i32* %p2, align 2
     50   ret void
     51 }
     52 
     53 define void @stur_long(i64 %a, i64 %b, i64* nocapture %p) nounwind {
     54 ; STUR_CHK: stur_long
     55 ; STUR_CHK: stp x{{[0-9]+}}, {{x[0-9]+}}, [x{{[0-9]+}}, #-16]
     56 ; STUR_CHK-NEXT: ret
     57   %p1 = getelementptr inbounds i64* %p, i32 -1
     58   store i64 %a, i64* %p1, align 2
     59   %p2 = getelementptr inbounds i64* %p, i32 -2
     60   store i64 %b, i64* %p2, align 2
     61   ret void
     62 }
     63 
     64 define void @stur_float(float %a, float %b, float* nocapture %p) nounwind {
     65 ; STUR_CHK: stur_float
     66 ; STUR_CHK: stp s{{[0-9]+}}, {{s[0-9]+}}, [x{{[0-9]+}}, #-8]
     67 ; STUR_CHK-NEXT: ret
     68   %p1 = getelementptr inbounds float* %p, i32 -1
     69   store float %a, float* %p1, align 2
     70   %p2 = getelementptr inbounds float* %p, i32 -2
     71   store float %b, float* %p2, align 2
     72   ret void
     73 }
     74 
     75 define void @stur_double(double %a, double %b, double* nocapture %p) nounwind {
     76 ; STUR_CHK: stur_double
     77 ; STUR_CHK: stp d{{[0-9]+}}, {{d[0-9]+}}, [x{{[0-9]+}}, #-16]
     78 ; STUR_CHK-NEXT: ret
     79   %p1 = getelementptr inbounds double* %p, i32 -1
     80   store double %a, double* %p1, align 2
     81   %p2 = getelementptr inbounds double* %p, i32 -2
     82   store double %b, double* %p2, align 2
     83   ret void
     84 }
     85 
     86 define void @splat_v4i32(i32 %v, i32 *%p) {
     87 entry:
     88 
     89 ; CHECK-LABEL: splat_v4i32
     90 ; CHECK-DAG: stp w0, w0, [x1]
     91 ; CHECK-DAG: stp w0, w0, [x1, #8]
     92 ; CHECK: ret
     93 
     94   %p17 = insertelement <4 x i32> undef, i32 %v, i32 0
     95   %p18 = insertelement <4 x i32> %p17, i32 %v, i32 1
     96   %p19 = insertelement <4 x i32> %p18, i32 %v, i32 2
     97   %p20 = insertelement <4 x i32> %p19, i32 %v, i32 3
     98   %p21 = bitcast i32* %p to <4 x i32>*
     99   store <4 x i32> %p20, <4 x i32>* %p21, align 4
    100   ret void
    101 }
    102