Home | History | Annotate | Download | only in SystemZ
      1 ; Test moves between FPRs and GPRs.
      2 ;
      3 ; RUN: llc < %s -mtriple=s390x-linux-gnu | FileCheck %s
      4 
      5 declare i64 @foo()
      6 declare double @bar()
      7 @dptr = external global double
      8 @iptr = external global i64
      9 
     10 ; Test 32-bit moves from GPRs to FPRs.  The GPR must be moved into the high
     11 ; 32 bits of the FPR.
     12 define float @f1(i32 %a) {
     13 ; CHECK-LABEL: f1:
     14 ; CHECK: sllg [[REGISTER:%r[0-5]]], %r2, 32
     15 ; CHECK: ldgr %f0, [[REGISTER]]
     16   %res = bitcast i32 %a to float
     17   ret float %res
     18 }
     19 
     20 ; Like f1, but create a situation where the shift can be folded with
     21 ; surrounding code.
     22 define float @f2(i64 %big) {
     23 ; CHECK-LABEL: f2:
     24 ; CHECK: risbg [[REGISTER:%r[0-5]]], %r2, 0, 159, 31
     25 ; CHECK: ldgr %f0, [[REGISTER]]
     26   %shift = lshr i64 %big, 1
     27   %a = trunc i64 %shift to i32
     28   %res = bitcast i32 %a to float
     29   ret float %res
     30 }
     31 
     32 ; Another example of the same thing.
     33 define float @f3(i64 %big) {
     34 ; CHECK-LABEL: f3:
     35 ; CHECK: risbg [[REGISTER:%r[0-5]]], %r2, 0, 159, 2
     36 ; CHECK: ldgr %f0, [[REGISTER]]
     37   %shift = ashr i64 %big, 30
     38   %a = trunc i64 %shift to i32
     39   %res = bitcast i32 %a to float
     40   ret float %res
     41 }
     42 
     43 ; Like f1, but the value to transfer is already in the high 32 bits.
     44 define float @f4(i64 %big) {
     45 ; CHECK-LABEL: f4:
     46 ; CHECK-NOT: %r2
     47 ; CHECK: nilf %r2, 0
     48 ; CHECK-NOT: %r2
     49 ; CHECK: ldgr %f0, %r2
     50   %shift = ashr i64 %big, 32
     51   %a = trunc i64 %shift to i32
     52   %res = bitcast i32 %a to float
     53   ret float %res
     54 }
     55 
     56 ; Test 64-bit moves from GPRs to FPRs.
     57 define double @f5(i64 %a) {
     58 ; CHECK-LABEL: f5:
     59 ; CHECK: ldgr %f0, %r2
     60   %res = bitcast i64 %a to double
     61   ret double %res
     62 }
     63 
     64 ; Test 128-bit moves from GPRs to FPRs.  i128 isn't a legitimate type,
     65 ; so this goes through memory.
     66 ; FIXME: it would be better to use one MVC here.
     67 define void @f6(fp128 *%a, i128 *%b) {
     68 ; CHECK-LABEL: f6:
     69 ; CHECK: lg
     70 ; CHECK: mvc
     71 ; CHECK: stg
     72 ; CHECK: br %r14
     73   %val = load i128 *%b
     74   %res = bitcast i128 %val to fp128
     75   store fp128 %res, fp128 *%a
     76   ret void
     77 }
     78 
     79 ; Test 32-bit moves from FPRs to GPRs.  The high 32 bits of the FPR should
     80 ; be moved into the low 32 bits of the GPR.
     81 define i32 @f7(float %a) {
     82 ; CHECK-LABEL: f7:
     83 ; CHECK: lgdr [[REGISTER:%r[0-5]]], %f0
     84 ; CHECK: srlg %r2, [[REGISTER]], 32
     85   %res = bitcast float %a to i32
     86   ret i32 %res
     87 }
     88 
     89 ; Test 64-bit moves from FPRs to GPRs.
     90 define i64 @f8(double %a) {
     91 ; CHECK-LABEL: f8:
     92 ; CHECK: lgdr %r2, %f0
     93   %res = bitcast double %a to i64
     94   ret i64 %res
     95 }
     96 
     97 ; Test 128-bit moves from FPRs to GPRs, with the same restriction as f6.
     98 define void @f9(fp128 *%a, i128 *%b) {
     99 ; CHECK-LABEL: f9:
    100 ; CHECK: ld
    101 ; CHECK: ld
    102 ; CHECK: std
    103 ; CHECK: std
    104   %val = load fp128 *%a
    105   %res = bitcast fp128 %val to i128
    106   store i128 %res, i128 *%b
    107   ret void
    108 }
    109 
    110 ; Test cases where the destination of an LGDR needs to be spilled.
    111 ; We shouldn't have any integer stack stores or floating-point loads.
    112 define void @f10(double %extra) {
    113 ; CHECK-LABEL: f10:
    114 ; CHECK: dptr
    115 ; CHECK-NOT: stg {{.*}}(%r15)
    116 ; CHECK: %loop
    117 ; CHECK-NOT: ld {{.*}}(%r15)
    118 ; CHECK: %exit
    119 ; CHECK: br %r14
    120 entry:
    121   %double0 = load volatile double *@dptr
    122   %biased0 = fadd double %double0, %extra
    123   %int0 = bitcast double %biased0 to i64
    124   %double1 = load volatile double *@dptr
    125   %biased1 = fadd double %double1, %extra
    126   %int1 = bitcast double %biased1 to i64
    127   %double2 = load volatile double *@dptr
    128   %biased2 = fadd double %double2, %extra
    129   %int2 = bitcast double %biased2 to i64
    130   %double3 = load volatile double *@dptr
    131   %biased3 = fadd double %double3, %extra
    132   %int3 = bitcast double %biased3 to i64
    133   %double4 = load volatile double *@dptr
    134   %biased4 = fadd double %double4, %extra
    135   %int4 = bitcast double %biased4 to i64
    136   %double5 = load volatile double *@dptr
    137   %biased5 = fadd double %double5, %extra
    138   %int5 = bitcast double %biased5 to i64
    139   %double6 = load volatile double *@dptr
    140   %biased6 = fadd double %double6, %extra
    141   %int6 = bitcast double %biased6 to i64
    142   %double7 = load volatile double *@dptr
    143   %biased7 = fadd double %double7, %extra
    144   %int7 = bitcast double %biased7 to i64
    145   %double8 = load volatile double *@dptr
    146   %biased8 = fadd double %double8, %extra
    147   %int8 = bitcast double %biased8 to i64
    148   %double9 = load volatile double *@dptr
    149   %biased9 = fadd double %double9, %extra
    150   %int9 = bitcast double %biased9 to i64
    151   br label %loop
    152 
    153 loop:
    154   %start = call i64 @foo()
    155   %or0 = or i64 %start, %int0
    156   %or1 = or i64 %or0, %int1
    157   %or2 = or i64 %or1, %int2
    158   %or3 = or i64 %or2, %int3
    159   %or4 = or i64 %or3, %int4
    160   %or5 = or i64 %or4, %int5
    161   %or6 = or i64 %or5, %int6
    162   %or7 = or i64 %or6, %int7
    163   %or8 = or i64 %or7, %int8
    164   %or9 = or i64 %or8, %int9
    165   store i64 %or9, i64 *@iptr
    166   %cont = icmp ne i64 %start, 1
    167   br i1 %cont, label %loop, label %exit
    168 
    169 exit:
    170   ret void
    171 }
    172 
    173 ; ...likewise LDGR, with the requirements the other way around.
    174 define void @f11(i64 %mask) {
    175 ; CHECK-LABEL: f11:
    176 ; CHECK: iptr
    177 ; CHECK-NOT: std {{.*}}(%r15)
    178 ; CHECK: %loop
    179 ; CHECK-NOT: lg {{.*}}(%r15)
    180 ; CHECK: %exit
    181 ; CHECK: br %r14
    182 entry:
    183   %int0 = load volatile i64 *@iptr
    184   %masked0 = and i64 %int0, %mask
    185   %double0 = bitcast i64 %masked0 to double
    186   %int1 = load volatile i64 *@iptr
    187   %masked1 = and i64 %int1, %mask
    188   %double1 = bitcast i64 %masked1 to double
    189   %int2 = load volatile i64 *@iptr
    190   %masked2 = and i64 %int2, %mask
    191   %double2 = bitcast i64 %masked2 to double
    192   %int3 = load volatile i64 *@iptr
    193   %masked3 = and i64 %int3, %mask
    194   %double3 = bitcast i64 %masked3 to double
    195   %int4 = load volatile i64 *@iptr
    196   %masked4 = and i64 %int4, %mask
    197   %double4 = bitcast i64 %masked4 to double
    198   %int5 = load volatile i64 *@iptr
    199   %masked5 = and i64 %int5, %mask
    200   %double5 = bitcast i64 %masked5 to double
    201   %int6 = load volatile i64 *@iptr
    202   %masked6 = and i64 %int6, %mask
    203   %double6 = bitcast i64 %masked6 to double
    204   %int7 = load volatile i64 *@iptr
    205   %masked7 = and i64 %int7, %mask
    206   %double7 = bitcast i64 %masked7 to double
    207   %int8 = load volatile i64 *@iptr
    208   %masked8 = and i64 %int8, %mask
    209   %double8 = bitcast i64 %masked8 to double
    210   %int9 = load volatile i64 *@iptr
    211   %masked9 = and i64 %int9, %mask
    212   %double9 = bitcast i64 %masked9 to double
    213   br label %loop
    214 
    215 loop:
    216   %start = call double @bar()
    217   %add0 = fadd double %start, %double0
    218   %add1 = fadd double %add0, %double1
    219   %add2 = fadd double %add1, %double2
    220   %add3 = fadd double %add2, %double3
    221   %add4 = fadd double %add3, %double4
    222   %add5 = fadd double %add4, %double5
    223   %add6 = fadd double %add5, %double6
    224   %add7 = fadd double %add6, %double7
    225   %add8 = fadd double %add7, %double8
    226   %add9 = fadd double %add8, %double9
    227   store double %add9, double *@dptr
    228   %cont = fcmp one double %start, 1.0
    229   br i1 %cont, label %loop, label %exit
    230 
    231 exit:
    232   ret void
    233 }
    234 
    235 ; Test cases where the source of an LDGR needs to be spilled.
    236 ; We shouldn't have any integer stack stores or floating-point loads.
    237 define void @f12() {
    238 ; CHECK-LABEL: f12:
    239 ; CHECK: %loop
    240 ; CHECK-NOT: std {{.*}}(%r15)
    241 ; CHECK: %exit
    242 ; CHECK: foo@PLT
    243 ; CHECK-NOT: lg {{.*}}(%r15)
    244 ; CHECK: foo@PLT
    245 ; CHECK: br %r14
    246 entry:
    247   br label %loop
    248 
    249 loop:
    250   %int0 = phi i64 [ 0, %entry ], [ %add0, %loop ]
    251   %int1 = phi i64 [ 0, %entry ], [ %add1, %loop ]
    252   %int2 = phi i64 [ 0, %entry ], [ %add2, %loop ]
    253   %int3 = phi i64 [ 0, %entry ], [ %add3, %loop ]
    254   %int4 = phi i64 [ 0, %entry ], [ %add4, %loop ]
    255   %int5 = phi i64 [ 0, %entry ], [ %add5, %loop ]
    256   %int6 = phi i64 [ 0, %entry ], [ %add6, %loop ]
    257   %int7 = phi i64 [ 0, %entry ], [ %add7, %loop ]
    258   %int8 = phi i64 [ 0, %entry ], [ %add8, %loop ]
    259   %int9 = phi i64 [ 0, %entry ], [ %add9, %loop ]
    260 
    261   %bias = call i64 @foo()
    262   %add0 = add i64 %int0, %bias
    263   %add1 = add i64 %int1, %bias
    264   %add2 = add i64 %int2, %bias
    265   %add3 = add i64 %int3, %bias
    266   %add4 = add i64 %int4, %bias
    267   %add5 = add i64 %int5, %bias
    268   %add6 = add i64 %int6, %bias
    269   %add7 = add i64 %int7, %bias
    270   %add8 = add i64 %int8, %bias
    271   %add9 = add i64 %int9, %bias
    272   %cont = icmp ne i64 %bias, 1
    273   br i1 %cont, label %loop, label %exit
    274 
    275 exit:
    276   %unused1 = call i64 @foo()
    277   %factor = load volatile double *@dptr
    278 
    279   %conv0 = bitcast i64 %add0 to double
    280   %mul0 = fmul double %conv0, %factor
    281   store volatile double %mul0, double *@dptr
    282   %conv1 = bitcast i64 %add1 to double
    283   %mul1 = fmul double %conv1, %factor
    284   store volatile double %mul1, double *@dptr
    285   %conv2 = bitcast i64 %add2 to double
    286   %mul2 = fmul double %conv2, %factor
    287   store volatile double %mul2, double *@dptr
    288   %conv3 = bitcast i64 %add3 to double
    289   %mul3 = fmul double %conv3, %factor
    290   store volatile double %mul3, double *@dptr
    291   %conv4 = bitcast i64 %add4 to double
    292   %mul4 = fmul double %conv4, %factor
    293   store volatile double %mul4, double *@dptr
    294   %conv5 = bitcast i64 %add5 to double
    295   %mul5 = fmul double %conv5, %factor
    296   store volatile double %mul5, double *@dptr
    297   %conv6 = bitcast i64 %add6 to double
    298   %mul6 = fmul double %conv6, %factor
    299   store volatile double %mul6, double *@dptr
    300   %conv7 = bitcast i64 %add7 to double
    301   %mul7 = fmul double %conv7, %factor
    302   store volatile double %mul7, double *@dptr
    303   %conv8 = bitcast i64 %add8 to double
    304   %mul8 = fmul double %conv8, %factor
    305   store volatile double %mul8, double *@dptr
    306   %conv9 = bitcast i64 %add9 to double
    307   %mul9 = fmul double %conv9, %factor
    308   store volatile double %mul9, double *@dptr
    309 
    310   %unused2 = call i64 @foo()
    311 
    312   ret void
    313 }
    314 
    315 ; ...likewise LGDR, with the requirements the other way around.
    316 define void @f13() {
    317 ; CHECK-LABEL: f13:
    318 ; CHECK: %loop
    319 ; CHECK-NOT: stg {{.*}}(%r15)
    320 ; CHECK: %exit
    321 ; CHECK: foo@PLT
    322 ; CHECK-NOT: ld {{.*}}(%r15)
    323 ; CHECK: foo@PLT
    324 ; CHECK: br %r14
    325 entry:
    326   br label %loop
    327 
    328 loop:
    329   %double0 = phi double [ 1.0, %entry ], [ %mul0, %loop ]
    330   %double1 = phi double [ 1.0, %entry ], [ %mul1, %loop ]
    331   %double2 = phi double [ 1.0, %entry ], [ %mul2, %loop ]
    332   %double3 = phi double [ 1.0, %entry ], [ %mul3, %loop ]
    333   %double4 = phi double [ 1.0, %entry ], [ %mul4, %loop ]
    334   %double5 = phi double [ 1.0, %entry ], [ %mul5, %loop ]
    335   %double6 = phi double [ 1.0, %entry ], [ %mul6, %loop ]
    336   %double7 = phi double [ 1.0, %entry ], [ %mul7, %loop ]
    337   %double8 = phi double [ 1.0, %entry ], [ %mul8, %loop ]
    338   %double9 = phi double [ 1.0, %entry ], [ %mul9, %loop ]
    339 
    340   %factor = call double @bar()
    341   %mul0 = fmul double %double0, %factor
    342   %mul1 = fmul double %double1, %factor
    343   %mul2 = fmul double %double2, %factor
    344   %mul3 = fmul double %double3, %factor
    345   %mul4 = fmul double %double4, %factor
    346   %mul5 = fmul double %double5, %factor
    347   %mul6 = fmul double %double6, %factor
    348   %mul7 = fmul double %double7, %factor
    349   %mul8 = fmul double %double8, %factor
    350   %mul9 = fmul double %double9, %factor
    351   %cont = fcmp one double %factor, 1.0
    352   br i1 %cont, label %loop, label %exit
    353 
    354 exit:
    355   %unused1 = call i64 @foo()
    356   %bias = load volatile i64 *@iptr
    357 
    358   %conv0 = bitcast double %mul0 to i64
    359   %add0 = add i64 %conv0, %bias
    360   store volatile i64 %add0, i64 *@iptr
    361   %conv1 = bitcast double %mul1 to i64
    362   %add1 = add i64 %conv1, %bias
    363   store volatile i64 %add1, i64 *@iptr
    364   %conv2 = bitcast double %mul2 to i64
    365   %add2 = add i64 %conv2, %bias
    366   store volatile i64 %add2, i64 *@iptr
    367   %conv3 = bitcast double %mul3 to i64
    368   %add3 = add i64 %conv3, %bias
    369   store volatile i64 %add3, i64 *@iptr
    370   %conv4 = bitcast double %mul4 to i64
    371   %add4 = add i64 %conv4, %bias
    372   store volatile i64 %add4, i64 *@iptr
    373   %conv5 = bitcast double %mul5 to i64
    374   %add5 = add i64 %conv5, %bias
    375   store volatile i64 %add5, i64 *@iptr
    376   %conv6 = bitcast double %mul6 to i64
    377   %add6 = add i64 %conv6, %bias
    378   store volatile i64 %add6, i64 *@iptr
    379   %conv7 = bitcast double %mul7 to i64
    380   %add7 = add i64 %conv7, %bias
    381   store volatile i64 %add7, i64 *@iptr
    382   %conv8 = bitcast double %mul8 to i64
    383   %add8 = add i64 %conv8, %bias
    384   store volatile i64 %add8, i64 *@iptr
    385   %conv9 = bitcast double %mul9 to i64
    386   %add9 = add i64 %conv9, %bias
    387   store volatile i64 %add9, i64 *@iptr
    388 
    389   %unused2 = call i64 @foo()
    390 
    391   ret void
    392 }
    393