Home | History | Annotate | Download | only in AMDGPU
      1 ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -verify-machineinstrs < %s | FileCheck -check-prefixes=GCN,SI,FUNC %s
      2 ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=tonga -verify-machineinstrs < %s | FileCheck -check-prefixes=GCN,VI,FUNC %s
      3 ; RUN: llc -amdgpu-scalarize-global-loads=false -march=amdgcn -mcpu=gfx900 -verify-machineinstrs < %s | FileCheck -check-prefixes=GCN,GFX9,FUNC %s
      4 ; RUN: llc -amdgpu-scalarize-global-loads=false -march=r600 -mcpu=cypress -verify-machineinstrs < %s | FileCheck -check-prefixes=EG,FUNC %s
      5 
      6 ; FUNC-LABEL: {{^}}s_uaddo_i64_zext:
      7 ; GCN: s_add_u32
      8 ; GCN: s_addc_u32
      9 ; GCN: v_cmp_lt_u64_e32 vcc
     10 
     11 ; EG: ADDC_UINT
     12 ; EG: ADDC_UINT
     13 define amdgpu_kernel void @s_uaddo_i64_zext(i64 addrspace(1)* %out, i64 %a, i64 %b) #0 {
     14   %uadd = call { i64, i1 } @llvm.uadd.with.overflow.i64(i64 %a, i64 %b)
     15   %val = extractvalue { i64, i1 } %uadd, 0
     16   %carry = extractvalue { i64, i1 } %uadd, 1
     17   %ext = zext i1 %carry to i64
     18   %add2 = add i64 %val, %ext
     19   store i64 %add2, i64 addrspace(1)* %out, align 8
     20   ret void
     21 }
     22 
     23 ; FIXME: Could do scalar
     24 
     25 ; FUNC-LABEL: {{^}}s_uaddo_i32:
     26 ; SI: v_add_i32_e32 v{{[0-9]+}}, vcc, s{{[0-9]+}}, v{{[0-9]+}}
     27 ; VI: v_add_u32_e32 v{{[0-9]+}}, vcc, s{{[0-9]+}}, v{{[0-9]+}}
     28 ; GFX9: v_add_co_u32_e32 v{{[0-9]+}}, vcc, s{{[0-9]+}}, v{{[0-9]+}}
     29 
     30 ; GCN: v_cndmask_b32_e64 v{{[0-9]+}}, 0, 1, vcc
     31 
     32 ; EG: ADDC_UINT
     33 ; EG: ADD_INT
     34 define amdgpu_kernel void @s_uaddo_i32(i32 addrspace(1)* %out, i1 addrspace(1)* %carryout, i32 %a, i32 %b) #0 {
     35   %uadd = call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 %a, i32 %b)
     36   %val = extractvalue { i32, i1 } %uadd, 0
     37   %carry = extractvalue { i32, i1 } %uadd, 1
     38   store i32 %val, i32 addrspace(1)* %out, align 4
     39   store i1 %carry, i1 addrspace(1)* %carryout
     40   ret void
     41 }
     42 
     43 ; FUNC-LABEL: {{^}}v_uaddo_i32:
     44 ; SI: v_add_i32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
     45 ; VI: v_add_u32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
     46 ; GFX9: v_add_co_u32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
     47 
     48 ; GCN: v_cndmask_b32_e64 v{{[0-9]+}}, 0, 1, vcc
     49 
     50 ; EG: ADDC_UINT
     51 ; EG: ADD_INT
     52 define amdgpu_kernel void @v_uaddo_i32(i32 addrspace(1)* %out, i1 addrspace(1)* %carryout, i32 addrspace(1)* %a.ptr, i32 addrspace(1)* %b.ptr) #0 {
     53   %tid = call i32 @llvm.amdgcn.workitem.id.x()
     54   %tid.ext = sext i32 %tid to i64
     55   %a.gep = getelementptr inbounds i32, i32 addrspace(1)* %a.ptr
     56   %b.gep = getelementptr inbounds i32, i32 addrspace(1)* %b.ptr
     57   %a = load i32, i32 addrspace(1)* %a.gep, align 4
     58   %b = load i32, i32 addrspace(1)* %b.gep, align 4
     59   %uadd = call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 %a, i32 %b)
     60   %val = extractvalue { i32, i1 } %uadd, 0
     61   %carry = extractvalue { i32, i1 } %uadd, 1
     62   store i32 %val, i32 addrspace(1)* %out, align 4
     63   store i1 %carry, i1 addrspace(1)* %carryout
     64   ret void
     65 }
     66 
     67 ; FUNC-LABEL: {{^}}v_uaddo_i32_novcc:
     68 ; SI: v_add_i32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
     69 ; VI: v_add_u32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
     70 ; GFX9: v_add_co_u32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
     71 
     72 ; GCN: v_cndmask_b32_e64 v{{[0-9]+}}, 0, 1, vcc
     73 
     74 ; EG: ADDC_UINT
     75 ; EG: ADD_INT
     76 define amdgpu_kernel void @v_uaddo_i32_novcc(i32 addrspace(1)* %out, i1 addrspace(1)* %carryout, i32 addrspace(1)* %a.ptr, i32 addrspace(1)* %b.ptr) #0 {
     77   %tid = call i32 @llvm.amdgcn.workitem.id.x()
     78   %tid.ext = sext i32 %tid to i64
     79   %a.gep = getelementptr inbounds i32, i32 addrspace(1)* %a.ptr
     80   %b.gep = getelementptr inbounds i32, i32 addrspace(1)* %b.ptr
     81   %a = load i32, i32 addrspace(1)* %a.gep, align 4
     82   %b = load i32, i32 addrspace(1)* %b.gep, align 4
     83   %uadd = call { i32, i1 } @llvm.uadd.with.overflow.i32(i32 %a, i32 %b)
     84   %val = extractvalue { i32, i1 } %uadd, 0
     85   %carry = extractvalue { i32, i1 } %uadd, 1
     86   store volatile i32 %val, i32 addrspace(1)* %out, align 4
     87   call void asm sideeffect "", "~{VCC}"() #0
     88   store volatile i1 %carry, i1 addrspace(1)* %carryout
     89   ret void
     90 }
     91 
     92 ; FUNC-LABEL: {{^}}s_uaddo_i64:
     93 ; GCN: s_add_u32
     94 ; GCN: s_addc_u32
     95 
     96 ; EG: ADDC_UINT
     97 ; EG: ADD_INT
     98 define amdgpu_kernel void @s_uaddo_i64(i64 addrspace(1)* %out, i1 addrspace(1)* %carryout, i64 %a, i64 %b) #0 {
     99   %uadd = call { i64, i1 } @llvm.uadd.with.overflow.i64(i64 %a, i64 %b)
    100   %val = extractvalue { i64, i1 } %uadd, 0
    101   %carry = extractvalue { i64, i1 } %uadd, 1
    102   store i64 %val, i64 addrspace(1)* %out, align 8
    103   store i1 %carry, i1 addrspace(1)* %carryout
    104   ret void
    105 }
    106 
    107 ; FUNC-LABEL: {{^}}v_uaddo_i64:
    108 ; SI: v_add_i32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
    109 ; SI: v_addc_u32_e32 v{{[0-9]+}}, vcc,
    110 
    111 ; VI: v_add_u32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
    112 ; VI: v_addc_u32_e32 v{{[0-9]+}}, vcc,
    113 
    114 ; GFX9: v_add_co_u32_e32 v{{[0-9]+}}, vcc, v{{[0-9]+}}, v{{[0-9]+}}
    115 ; GFX9: v_addc_co_u32_e32 v{{[0-9]+}}, vcc,
    116 
    117 ; EG: ADDC_UINT
    118 ; EG: ADD_INT
    119 define amdgpu_kernel void @v_uaddo_i64(i64 addrspace(1)* %out, i1 addrspace(1)* %carryout, i64 addrspace(1)* %a.ptr, i64 addrspace(1)* %b.ptr) #0 {
    120   %tid = call i32 @llvm.amdgcn.workitem.id.x()
    121   %tid.ext = sext i32 %tid to i64
    122   %a.gep = getelementptr inbounds i64, i64 addrspace(1)* %a.ptr
    123   %b.gep = getelementptr inbounds i64, i64 addrspace(1)* %b.ptr
    124   %a = load i64, i64 addrspace(1)* %a.gep
    125   %b = load i64, i64 addrspace(1)* %b.gep
    126   %uadd = call { i64, i1 } @llvm.uadd.with.overflow.i64(i64 %a, i64 %b)
    127   %val = extractvalue { i64, i1 } %uadd, 0
    128   %carry = extractvalue { i64, i1 } %uadd, 1
    129   store i64 %val, i64 addrspace(1)* %out
    130   store i1 %carry, i1 addrspace(1)* %carryout
    131   ret void
    132 }
    133 
    134 ; FUNC-LABEL: {{^}}v_uaddo_i16:
    135 ; VI: v_add_u16_e32
    136 ; VI: v_cmp_lt_u16_e32
    137 
    138 ; GFX9: v_add_u16_e32
    139 ; GFX9: v_cmp_lt_u16_e32
    140 define amdgpu_kernel void @v_uaddo_i16(i16 addrspace(1)* %out, i1 addrspace(1)* %carryout, i16 addrspace(1)* %a.ptr, i16 addrspace(1)* %b.ptr) #0 {
    141   %tid = call i32 @llvm.amdgcn.workitem.id.x()
    142   %tid.ext = sext i32 %tid to i64
    143   %a.gep = getelementptr inbounds i16, i16 addrspace(1)* %a.ptr
    144   %b.gep = getelementptr inbounds i16, i16 addrspace(1)* %b.ptr
    145   %a = load i16, i16 addrspace(1)* %a.gep
    146   %b = load i16, i16 addrspace(1)* %b.gep
    147   %uadd = call { i16, i1 } @llvm.uadd.with.overflow.i16(i16 %a, i16 %b)
    148   %val = extractvalue { i16, i1 } %uadd, 0
    149   %carry = extractvalue { i16, i1 } %uadd, 1
    150   store i16 %val, i16 addrspace(1)* %out
    151   store i1 %carry, i1 addrspace(1)* %carryout
    152   ret void
    153 }
    154 
    155 declare i32 @llvm.amdgcn.workitem.id.x() #1
    156 declare { i16, i1 } @llvm.uadd.with.overflow.i16(i16, i16) #1
    157 declare { i32, i1 } @llvm.uadd.with.overflow.i32(i32, i32) #1
    158 declare { i64, i1 } @llvm.uadd.with.overflow.i64(i64, i64) #1
    159 
    160 attributes #0 = { nounwind }
    161 attributes #1 = { nounwind readnone }
    162