Home | History | Annotate | Download | only in NVPTX
      1 ; RUN: llc < %s -march=nvptx -mcpu=sm_20 | FileCheck %s
      2 
      3 target datalayout = "e-p:32:32:32-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v16:16:16-v32:32:32-v64:64:64-v128:128:128-n16:32:64"
      4 
      5 
      6 define void @one(i64 %a, i64 %b, i64* %p1, i64* %p2) {
      7 ; CHECK: cvt.s64.s8
      8 ; CHECK: cvt.s64.s8
      9 entry:
     10   %sext = shl i64 %a, 56
     11   %conv1 = ashr exact i64 %sext, 56
     12   %sext1 = shl i64 %b, 56
     13   %conv4 = ashr exact i64 %sext1, 56
     14   %shr = ashr i64 %a, 16
     15   %shr9 = ashr i64 %b, 16
     16   %add = add nsw i64 %conv4, %conv1
     17   store i64 %add, i64* %p1, align 8
     18   %add17 = add nsw i64 %shr9, %shr
     19   store i64 %add17, i64* %p2, align 8
     20   ret void
     21 }
     22 
     23 
     24 define void @two(i64 %a, i64 %b, i64* %p1, i64* %p2) {
     25 entry:
     26 ; CHECK: cvt.s64.s32
     27 ; CHECK: cvt.s64.s32
     28   %sext = shl i64 %a, 32
     29   %conv1 = ashr exact i64 %sext, 32
     30   %sext1 = shl i64 %b, 32
     31   %conv4 = ashr exact i64 %sext1, 32
     32   %shr = ashr i64 %a, 16
     33   %shr9 = ashr i64 %b, 16
     34   %add = add nsw i64 %conv4, %conv1
     35   store i64 %add, i64* %p1, align 8
     36   %add17 = add nsw i64 %shr9, %shr
     37   store i64 %add17, i64* %p2, align 8
     38   ret void
     39 }
     40 
     41 
     42 define void @three(i64 %a, i64 %b, i64* %p1, i64* %p2) {
     43 entry:
     44 ; CHECK: cvt.s64.s16
     45 ; CHECK: cvt.s64.s16
     46   %sext = shl i64 %a, 48
     47   %conv1 = ashr exact i64 %sext, 48
     48   %sext1 = shl i64 %b, 48
     49   %conv4 = ashr exact i64 %sext1, 48
     50   %shr = ashr i64 %a, 16
     51   %shr9 = ashr i64 %b, 16
     52   %add = add nsw i64 %conv4, %conv1
     53   store i64 %add, i64* %p1, align 8
     54   %add17 = add nsw i64 %shr9, %shr
     55   store i64 %add17, i64* %p2, align 8
     56   ret void
     57 }
     58 
     59 
     60 define void @four(i32 %a, i32 %b, i32* %p1, i32* %p2) {
     61 entry:
     62 ; CHECK: cvt.s32.s8
     63 ; CHECK: cvt.s32.s8
     64   %sext = shl i32 %a, 24
     65   %conv1 = ashr exact i32 %sext, 24
     66   %sext1 = shl i32 %b, 24
     67   %conv4 = ashr exact i32 %sext1, 24
     68   %shr = ashr i32 %a, 16
     69   %shr9 = ashr i32 %b, 16
     70   %add = add nsw i32 %conv4, %conv1
     71   store i32 %add, i32* %p1, align 4
     72   %add17 = add nsw i32 %shr9, %shr
     73   store i32 %add17, i32* %p2, align 4
     74   ret void
     75 }
     76 
     77 
     78 define void @five(i32 %a, i32 %b, i32* %p1, i32* %p2) {
     79 entry:
     80 ; CHECK: cvt.s32.s16
     81 ; CHECK: cvt.s32.s16
     82   %sext = shl i32 %a, 16
     83   %conv1 = ashr exact i32 %sext, 16
     84   %sext1 = shl i32 %b, 16
     85   %conv4 = ashr exact i32 %sext1, 16
     86   %shr = ashr i32 %a, 16
     87   %shr9 = ashr i32 %b, 16
     88   %add = add nsw i32 %conv4, %conv1
     89   store i32 %add, i32* %p1, align 4
     90   %add17 = add nsw i32 %shr9, %shr
     91   store i32 %add17, i32* %p2, align 4
     92   ret void
     93 }
     94 
     95 
     96 define void @six(i16 %a, i16 %b, i16* %p1, i16* %p2) {
     97 entry:
     98 ; CHECK: cvt.s16.s8
     99 ; CHECK: cvt.s16.s8
    100   %sext = shl i16 %a, 8
    101   %conv1 = ashr exact i16 %sext, 8
    102   %sext1 = shl i16 %b, 8
    103   %conv4 = ashr exact i16 %sext1, 8
    104   %shr = ashr i16 %a, 8
    105   %shr9 = ashr i16 %b, 8
    106   %add = add nsw i16 %conv4, %conv1
    107   store i16 %add, i16* %p1, align 4
    108   %add17 = add nsw i16 %shr9, %shr
    109   store i16 %add17, i16* %p2, align 4
    110   ret void
    111 }
    112