Home | History | Annotate | Download | only in AMDGPU
      1 ; RUN: llc -march=amdgcn -mcpu=gfx900 -mattr=-flat-for-global -verify-machineinstrs < %s | FileCheck -allow-deprecated-dag-overlap -check-prefix=GCN -check-prefix=GFX9 %s
      2 ; RUN: llc -march=amdgcn -mcpu=tonga -mattr=-flat-for-global -verify-machineinstrs < %s | FileCheck -allow-deprecated-dag-overlap -check-prefix=GCN -check-prefix=VI -check-prefix=CIVI %s
      3 ; RUN: llc -march=amdgcn -mcpu=bonaire -verify-machineinstrs < %s | FileCheck -allow-deprecated-dag-overlap -check-prefix=GCN -check-prefix=CI -check-prefix=CIVI %s
      4 
      5 ; FIXME: Should be same on CI/VI
      6 ; GCN-LABEL: {{^}}s_ashr_v2i16:
      7 ; GFX9: s_load_dword [[LHS:s[0-9]+]]
      8 ; GFX9: s_load_dword [[RHS:s[0-9]+]]
      9 ; GFX9: v_mov_b32_e32 [[VLHS:v[0-9]+]], [[LHS]]
     10 ; GFX9: v_pk_ashrrev_i16 [[RESULT:v[0-9]+]], [[RHS]], [[VLHS]]
     11 
     12 ; CIVI: s_load_dword [[LHS:s[0-9]+]]
     13 ; CIVI: s_load_dword [[RHS:s[0-9]+]]
     14 
     15 ; CIVI-DAG: s_ashr_i32
     16 ; CIVI-DAG: s_ashr_i32
     17 ; CIVI-DAG: s_sext_i32_i16
     18 ; CIVI-DAG: s_sext_i32_i16
     19 ; CIVI-DAG: s_ashr_i32
     20 ; CIVI-DAG: s_ashr_i32
     21 ; CIVI-DAG: s_lshl_b32
     22 ; CIVI: s_and_b32
     23 ; CIVI: s_or_b32
     24 
     25 define amdgpu_kernel void @s_ashr_v2i16(<2 x i16> addrspace(1)* %out, i32, <2 x i16> %lhs, i32, <2 x i16> %rhs) #0 {
     26   %result = ashr <2 x i16> %lhs, %rhs
     27   store <2 x i16> %result, <2 x i16> addrspace(1)* %out
     28   ret void
     29 }
     30 
     31 ; GCN-LABEL: {{^}}v_ashr_v2i16:
     32 ; GCN: {{buffer|flat|global}}_load_dword [[LHS:v[0-9]+]]
     33 ; GCN: {{buffer|flat|global}}_load_dword [[RHS:v[0-9]+]]
     34 ; GFX9: v_pk_ashrrev_i16 [[RESULT:v[0-9]+]], [[RHS]], [[LHS]]
     35 
     36 ; VI: v_ashrrev_i16_e32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}}
     37 ; VI: v_ashrrev_i16_sdwa v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}} dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1 src1_sel:WORD_1
     38 ; VI: v_or_b32_e32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}}
     39 
     40 ; CI: s_mov_b32 [[MASK:s[0-9]+]], 0xffff{{$}}
     41 ; CI-DAG: v_and_b32_e32 v{{[0-9]+}}, [[MASK]], [[RHS]]
     42 ; CI-DAG: v_bfe_i32 v{{[0-9]+}}, v{{[0-9]+}}, 0, 16
     43 ; CI: v_ashrrev_i32_e32 v{{[0-9]+}}, 16, [[LHS]]
     44 ; CI: v_lshrrev_b32_e32 v{{[0-9]+}}, 16, v{{[0-9]+}}
     45 ; CI: v_ashrrev_i32_e32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}}
     46 ; CI: v_ashrrev_i32_e32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}}
     47 ; CI: v_lshlrev_b32_e32 v{{[0-9]+}}, 16, v{{[0-9]+}}
     48 ; CI: v_and_b32_e32 v{{[0-9]+}}, [[MASK]], v{{[0-9]+}}
     49 ; CI: v_or_b32_e32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}}
     50 define amdgpu_kernel void @v_ashr_v2i16(<2 x i16> addrspace(1)* %out, <2 x i16> addrspace(1)* %in) #0 {
     51   %tid = call i32 @llvm.amdgcn.workitem.id.x()
     52   %tid.ext = sext i32 %tid to i64
     53   %in.gep = getelementptr inbounds <2 x i16>, <2 x i16> addrspace(1)* %in, i64 %tid.ext
     54   %out.gep = getelementptr inbounds <2 x i16>, <2 x i16> addrspace(1)* %out, i64 %tid.ext
     55   %b_ptr = getelementptr <2 x i16>, <2 x i16> addrspace(1)* %in.gep, i32 1
     56   %a = load <2 x i16>, <2 x i16> addrspace(1)* %in.gep
     57   %b = load <2 x i16>, <2 x i16> addrspace(1)* %b_ptr
     58   %result = ashr <2 x i16> %a, %b
     59   store <2 x i16> %result, <2 x i16> addrspace(1)* %out.gep
     60   ret void
     61 }
     62 
     63 ; GCN-LABEL: {{^}}ashr_v_s_v2i16:
     64 ; GFX9: s_load_dword [[RHS:s[0-9]+]]
     65 ; GFX9: {{buffer|flat|global}}_load_dword [[LHS:v[0-9]+]]
     66 ; GFX9: v_pk_ashrrev_i16 [[RESULT:v[0-9]+]], [[RHS]], [[LHS]]
     67 define amdgpu_kernel void @ashr_v_s_v2i16(<2 x i16> addrspace(1)* %out, <2 x i16> addrspace(1)* %in, <2 x i16> %sgpr) #0 {
     68   %tid = call i32 @llvm.amdgcn.workitem.id.x()
     69   %tid.ext = sext i32 %tid to i64
     70   %in.gep = getelementptr inbounds <2 x i16>, <2 x i16> addrspace(1)* %in, i64 %tid.ext
     71   %out.gep = getelementptr inbounds <2 x i16>, <2 x i16> addrspace(1)* %out, i64 %tid.ext
     72   %vgpr = load <2 x i16>, <2 x i16> addrspace(1)* %in.gep
     73   %result = ashr <2 x i16> %vgpr, %sgpr
     74   store <2 x i16> %result, <2 x i16> addrspace(1)* %out.gep
     75   ret void
     76 }
     77 
     78 ; GCN-LABEL: {{^}}ashr_s_v_v2i16:
     79 ; GFX9: s_load_dword [[LHS:s[0-9]+]]
     80 ; GFX9: {{buffer|flat|global}}_load_dword [[RHS:v[0-9]+]]
     81 ; GFX9: v_pk_ashrrev_i16 [[RESULT:v[0-9]+]], [[RHS]], [[LHS]]
     82 define amdgpu_kernel void @ashr_s_v_v2i16(<2 x i16> addrspace(1)* %out, <2 x i16> addrspace(1)* %in, <2 x i16> %sgpr) #0 {
     83   %tid = call i32 @llvm.amdgcn.workitem.id.x()
     84   %tid.ext = sext i32 %tid to i64
     85   %in.gep = getelementptr inbounds <2 x i16>, <2 x i16> addrspace(1)* %in, i64 %tid.ext
     86   %out.gep = getelementptr inbounds <2 x i16>, <2 x i16> addrspace(1)* %out, i64 %tid.ext
     87   %vgpr = load <2 x i16>, <2 x i16> addrspace(1)* %in.gep
     88   %result = ashr <2 x i16> %sgpr, %vgpr
     89   store <2 x i16> %result, <2 x i16> addrspace(1)* %out.gep
     90   ret void
     91 }
     92 
     93 ; GCN-LABEL: {{^}}ashr_imm_v_v2i16:
     94 ; GCN: {{buffer|flat|global}}_load_dword [[RHS:v[0-9]+]]
     95 ; GFX9: v_pk_ashrrev_i16 [[RESULT:v[0-9]+]], [[RHS]], -4
     96 define amdgpu_kernel void @ashr_imm_v_v2i16(<2 x i16> addrspace(1)* %out, <2 x i16> addrspace(1)* %in) #0 {
     97   %tid = call i32 @llvm.amdgcn.workitem.id.x()
     98   %tid.ext = sext i32 %tid to i64
     99   %in.gep = getelementptr inbounds <2 x i16>, <2 x i16> addrspace(1)* %in, i64 %tid.ext
    100   %out.gep = getelementptr inbounds <2 x i16>, <2 x i16> addrspace(1)* %out, i64 %tid.ext
    101   %vgpr = load <2 x i16>, <2 x i16> addrspace(1)* %in.gep
    102   %result = ashr <2 x i16> <i16 -4, i16 -4>, %vgpr
    103   store <2 x i16> %result, <2 x i16> addrspace(1)* %out.gep
    104   ret void
    105 }
    106 
    107 ; GCN-LABEL: {{^}}ashr_v_imm_v2i16:
    108 ; GCN: {{buffer|flat|global}}_load_dword [[LHS:v[0-9]+]]
    109 ; GFX9: v_pk_ashrrev_i16 [[RESULT:v[0-9]+]], 8, [[LHS]]
    110 define amdgpu_kernel void @ashr_v_imm_v2i16(<2 x i16> addrspace(1)* %out, <2 x i16> addrspace(1)* %in) #0 {
    111   %tid = call i32 @llvm.amdgcn.workitem.id.x()
    112   %tid.ext = sext i32 %tid to i64
    113   %in.gep = getelementptr inbounds <2 x i16>, <2 x i16> addrspace(1)* %in, i64 %tid.ext
    114   %out.gep = getelementptr inbounds <2 x i16>, <2 x i16> addrspace(1)* %out, i64 %tid.ext
    115   %vgpr = load <2 x i16>, <2 x i16> addrspace(1)* %in.gep
    116   %result = ashr <2 x i16> %vgpr, <i16 8, i16 8>
    117   store <2 x i16> %result, <2 x i16> addrspace(1)* %out.gep
    118   ret void
    119 }
    120 
    121 ; GCN-LABEL: {{^}}v_ashr_v4i16:
    122 ; GCN: {{buffer|flat|global}}_load_dwordx2
    123 ; GCN: {{buffer|flat|global}}_load_dwordx2
    124 ; GFX9: v_pk_ashrrev_i16 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}}
    125 ; GFX9: v_pk_ashrrev_i16 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}}
    126 
    127 ; VI: v_ashrrev_i16_e32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}}
    128 ; VI: v_ashrrev_i16_sdwa v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}} dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1 src1_sel:WORD_1
    129 ; VI: v_ashrrev_i16_e32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}}
    130 ; VI: v_ashrrev_i16_sdwa v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}} dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:WORD_1 src1_sel:WORD_1
    131 ; VI: v_or_b32_e32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}}
    132 ; VI: v_or_b32_e32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}}
    133 
    134 ; GCN: {{buffer|flat|global}}_store_dwordx2
    135 define amdgpu_kernel void @v_ashr_v4i16(<4 x i16> addrspace(1)* %out, <4 x i16> addrspace(1)* %in) #0 {
    136   %tid = call i32 @llvm.amdgcn.workitem.id.x()
    137   %tid.ext = sext i32 %tid to i64
    138   %in.gep = getelementptr inbounds <4 x i16>, <4 x i16> addrspace(1)* %in, i64 %tid.ext
    139   %out.gep = getelementptr inbounds <4 x i16>, <4 x i16> addrspace(1)* %out, i64 %tid.ext
    140   %b_ptr = getelementptr <4 x i16>, <4 x i16> addrspace(1)* %in.gep, i32 1
    141   %a = load <4 x i16>, <4 x i16> addrspace(1)* %in.gep
    142   %b = load <4 x i16>, <4 x i16> addrspace(1)* %b_ptr
    143   %result = ashr <4 x i16> %a, %b
    144   store <4 x i16> %result, <4 x i16> addrspace(1)* %out.gep
    145   ret void
    146 }
    147 
    148 ; GCN-LABEL: {{^}}ashr_v_imm_v4i16:
    149 ; GCN: {{buffer|flat|global}}_load_dwordx2
    150 ; GFX9: v_pk_ashrrev_i16 v{{[0-9]+}}, 8, v{{[0-9]+}}
    151 ; GFX9: v_pk_ashrrev_i16 v{{[0-9]+}}, 8, v{{[0-9]+}}
    152 ; GCN: {{buffer|flat|global}}_store_dwordx2
    153 define amdgpu_kernel void @ashr_v_imm_v4i16(<4 x i16> addrspace(1)* %out, <4 x i16> addrspace(1)* %in) #0 {
    154   %tid = call i32 @llvm.amdgcn.workitem.id.x()
    155   %tid.ext = sext i32 %tid to i64
    156   %in.gep = getelementptr inbounds <4 x i16>, <4 x i16> addrspace(1)* %in, i64 %tid.ext
    157   %out.gep = getelementptr inbounds <4 x i16>, <4 x i16> addrspace(1)* %out, i64 %tid.ext
    158   %vgpr = load <4 x i16>, <4 x i16> addrspace(1)* %in.gep
    159   %result = ashr <4 x i16> %vgpr, <i16 8, i16 8, i16 8, i16 8>
    160   store <4 x i16> %result, <4 x i16> addrspace(1)* %out.gep
    161   ret void
    162 }
    163 
    164 declare i32 @llvm.amdgcn.workitem.id.x() #1
    165 
    166 attributes #0 = { nounwind }
    167 attributes #1 = { nounwind readnone }
    168