Home | History | Annotate | Download | only in SystemZ
      1 ; Test 64-bit byteswaps from memory to registers.
      2 ;
      3 ; RUN: llc < %s -mtriple=s390x-linux-gnu | FileCheck %s
      4 
      5 declare i64 @llvm.bswap.i64(i64 %a)
      6 
      7 ; Check LRVG with no displacement.
      8 define i64 @f1(i64 *%src) {
      9 ; CHECK-LABEL: f1:
     10 ; CHECK: lrvg %r2, 0(%r2)
     11 ; CHECK: br %r14
     12   %a = load i64 *%src
     13   %swapped = call i64 @llvm.bswap.i64(i64 %a)
     14   ret i64 %swapped
     15 }
     16 
     17 ; Check the high end of the aligned LRVG range.
     18 define i64 @f2(i64 *%src) {
     19 ; CHECK-LABEL: f2:
     20 ; CHECK: lrvg %r2, 524280(%r2)
     21 ; CHECK: br %r14
     22   %ptr = getelementptr i64 *%src, i64 65535
     23   %a = load i64 *%ptr
     24   %swapped = call i64 @llvm.bswap.i64(i64 %a)
     25   ret i64 %swapped
     26 }
     27 
     28 ; Check the next doubleword up, which needs separate address logic.
     29 ; Other sequences besides this one would be OK.
     30 define i64 @f3(i64 *%src) {
     31 ; CHECK-LABEL: f3:
     32 ; CHECK: agfi %r2, 524288
     33 ; CHECK: lrvg %r2, 0(%r2)
     34 ; CHECK: br %r14
     35   %ptr = getelementptr i64 *%src, i64 65536
     36   %a = load i64 *%ptr
     37   %swapped = call i64 @llvm.bswap.i64(i64 %a)
     38   ret i64 %swapped
     39 }
     40 
     41 ; Check the high end of the negative aligned LRVG range.
     42 define i64 @f4(i64 *%src) {
     43 ; CHECK-LABEL: f4:
     44 ; CHECK: lrvg %r2, -8(%r2)
     45 ; CHECK: br %r14
     46   %ptr = getelementptr i64 *%src, i64 -1
     47   %a = load i64 *%ptr
     48   %swapped = call i64 @llvm.bswap.i64(i64 %a)
     49   ret i64 %swapped
     50 }
     51 
     52 ; Check the low end of the LRVG range.
     53 define i64 @f5(i64 *%src) {
     54 ; CHECK-LABEL: f5:
     55 ; CHECK: lrvg %r2, -524288(%r2)
     56 ; CHECK: br %r14
     57   %ptr = getelementptr i64 *%src, i64 -65536
     58   %a = load i64 *%ptr
     59   %swapped = call i64 @llvm.bswap.i64(i64 %a)
     60   ret i64 %swapped
     61 }
     62 
     63 ; Check the next doubleword down, which needs separate address logic.
     64 ; Other sequences besides this one would be OK.
     65 define i64 @f6(i64 *%src) {
     66 ; CHECK-LABEL: f6:
     67 ; CHECK: agfi %r2, -524296
     68 ; CHECK: lrvg %r2, 0(%r2)
     69 ; CHECK: br %r14
     70   %ptr = getelementptr i64 *%src, i64 -65537
     71   %a = load i64 *%ptr
     72   %swapped = call i64 @llvm.bswap.i64(i64 %a)
     73   ret i64 %swapped
     74 }
     75 
     76 ; Check that LRVG allows an index.
     77 define i64 @f7(i64 %src, i64 %index) {
     78 ; CHECK-LABEL: f7:
     79 ; CHECK: lrvg %r2, 524287({{%r3,%r2|%r2,%r3}})
     80 ; CHECK: br %r14
     81   %add1 = add i64 %src, %index
     82   %add2 = add i64 %add1, 524287
     83   %ptr = inttoptr i64 %add2 to i64 *
     84   %a = load i64 *%ptr
     85   %swapped = call i64 @llvm.bswap.i64(i64 %a)
     86   ret i64 %swapped
     87 }
     88 
     89 ; Check that volatile accesses do not use LRVG, which might access the
     90 ; storage multple times.
     91 define i64 @f8(i64 *%src) {
     92 ; CHECK-LABEL: f8:
     93 ; CHECK: lg [[REG:%r[0-5]]], 0(%r2)
     94 ; CHECK: lrvgr %r2, [[REG]]
     95 ; CHECK: br %r14
     96   %a = load volatile i64 *%src
     97   %swapped = call i64 @llvm.bswap.i64(i64 %a)
     98   ret i64 %swapped
     99 }
    100 
    101 ; Test a case where we spill the source of at least one LRVGR.  We want
    102 ; to use LRVG if possible.
    103 define void @f9(i64 *%ptr) {
    104 ; CHECK-LABEL: f9:
    105 ; CHECK: lrvg {{%r[0-9]+}}, 160(%r15)
    106 ; CHECK: br %r14
    107   %val0 = load volatile i64 *%ptr
    108   %val1 = load volatile i64 *%ptr
    109   %val2 = load volatile i64 *%ptr
    110   %val3 = load volatile i64 *%ptr
    111   %val4 = load volatile i64 *%ptr
    112   %val5 = load volatile i64 *%ptr
    113   %val6 = load volatile i64 *%ptr
    114   %val7 = load volatile i64 *%ptr
    115   %val8 = load volatile i64 *%ptr
    116   %val9 = load volatile i64 *%ptr
    117   %val10 = load volatile i64 *%ptr
    118   %val11 = load volatile i64 *%ptr
    119   %val12 = load volatile i64 *%ptr
    120   %val13 = load volatile i64 *%ptr
    121   %val14 = load volatile i64 *%ptr
    122   %val15 = load volatile i64 *%ptr
    123 
    124   %swapped0 = call i64 @llvm.bswap.i64(i64 %val0)
    125   %swapped1 = call i64 @llvm.bswap.i64(i64 %val1)
    126   %swapped2 = call i64 @llvm.bswap.i64(i64 %val2)
    127   %swapped3 = call i64 @llvm.bswap.i64(i64 %val3)
    128   %swapped4 = call i64 @llvm.bswap.i64(i64 %val4)
    129   %swapped5 = call i64 @llvm.bswap.i64(i64 %val5)
    130   %swapped6 = call i64 @llvm.bswap.i64(i64 %val6)
    131   %swapped7 = call i64 @llvm.bswap.i64(i64 %val7)
    132   %swapped8 = call i64 @llvm.bswap.i64(i64 %val8)
    133   %swapped9 = call i64 @llvm.bswap.i64(i64 %val9)
    134   %swapped10 = call i64 @llvm.bswap.i64(i64 %val10)
    135   %swapped11 = call i64 @llvm.bswap.i64(i64 %val11)
    136   %swapped12 = call i64 @llvm.bswap.i64(i64 %val12)
    137   %swapped13 = call i64 @llvm.bswap.i64(i64 %val13)
    138   %swapped14 = call i64 @llvm.bswap.i64(i64 %val14)
    139   %swapped15 = call i64 @llvm.bswap.i64(i64 %val15)
    140 
    141   store volatile i64 %val0, i64 *%ptr
    142   store volatile i64 %val1, i64 *%ptr
    143   store volatile i64 %val2, i64 *%ptr
    144   store volatile i64 %val3, i64 *%ptr
    145   store volatile i64 %val4, i64 *%ptr
    146   store volatile i64 %val5, i64 *%ptr
    147   store volatile i64 %val6, i64 *%ptr
    148   store volatile i64 %val7, i64 *%ptr
    149   store volatile i64 %val8, i64 *%ptr
    150   store volatile i64 %val9, i64 *%ptr
    151   store volatile i64 %val10, i64 *%ptr
    152   store volatile i64 %val11, i64 *%ptr
    153   store volatile i64 %val12, i64 *%ptr
    154   store volatile i64 %val13, i64 *%ptr
    155   store volatile i64 %val14, i64 *%ptr
    156   store volatile i64 %val15, i64 *%ptr
    157 
    158   store volatile i64 %swapped0, i64 *%ptr
    159   store volatile i64 %swapped1, i64 *%ptr
    160   store volatile i64 %swapped2, i64 *%ptr
    161   store volatile i64 %swapped3, i64 *%ptr
    162   store volatile i64 %swapped4, i64 *%ptr
    163   store volatile i64 %swapped5, i64 *%ptr
    164   store volatile i64 %swapped6, i64 *%ptr
    165   store volatile i64 %swapped7, i64 *%ptr
    166   store volatile i64 %swapped8, i64 *%ptr
    167   store volatile i64 %swapped9, i64 *%ptr
    168   store volatile i64 %swapped10, i64 *%ptr
    169   store volatile i64 %swapped11, i64 *%ptr
    170   store volatile i64 %swapped12, i64 *%ptr
    171   store volatile i64 %swapped13, i64 *%ptr
    172   store volatile i64 %swapped14, i64 *%ptr
    173   store volatile i64 %swapped15, i64 *%ptr
    174 
    175   ret void
    176 }
    177