Home | History | Annotate | Download | only in AMDGPU
      1 ; RUN: llc -march=amdgcn -verify-machineinstrs < %s | FileCheck -check-prefix=SI %s
      2 ;
      3 ;
      4 ; Most SALU instructions ignore control flow, so we need to make sure
      5 ; they don't overwrite values from other blocks.
      6 
      7 ; If the branch decision is made based on a value in an SGPR then all
      8 ; threads will execute the same code paths, so we don't need to worry
      9 ; about instructions in different blocks overwriting each other.
     10 ; SI-LABEL: {{^}}sgpr_if_else_salu_br:
     11 ; SI: s_add
     12 ; SI: s_branch
     13 
     14 ; SI: s_sub
     15 
     16 define amdgpu_kernel void @sgpr_if_else_salu_br(i32 addrspace(1)* %out, i32 %a, i32 %b, i32 %c, i32 %d, i32 %e) {
     17 entry:
     18   %0 = icmp eq i32 %a, 0
     19   br i1 %0, label %if, label %else
     20 
     21 if:
     22   %1 = sub i32 %b, %c
     23   br label %endif
     24 
     25 else:
     26   %2 = add i32 %d, %e
     27   br label %endif
     28 
     29 endif:
     30   %3 = phi i32 [%1, %if], [%2, %else]
     31   %4 = add i32 %3, %a
     32   store i32 %4, i32 addrspace(1)* %out
     33   ret void
     34 }
     35 
     36 ; SI-LABEL: {{^}}sgpr_if_else_salu_br_opt:
     37 ; SI: s_cmp_lg_u32
     38 ; SI: s_cbranch_scc0 [[IF:BB[0-9]+_[0-9]+]]
     39 
     40 ; SI: ; %bb.1: ; %else
     41 ; SI: s_load_dword [[LOAD0:s[0-9]+]], s{{\[[0-9]+:[0-9]+\]}}, 0x2e
     42 ; SI: s_load_dword [[LOAD1:s[0-9]+]], s{{\[[0-9]+:[0-9]+\]}}, 0x37
     43 ; SI-NOT: add
     44 ; SI: s_branch [[ENDIF:BB[0-9]+_[0-9]+]]
     45 
     46 ; SI: [[IF]]: ; %if
     47 ; SI: s_load_dword [[LOAD0]], s{{\[[0-9]+:[0-9]+\]}}, 0x1c
     48 ; SI: s_load_dword [[LOAD1]], s{{\[[0-9]+:[0-9]+\]}}, 0x25
     49 ; SI-NOT: add
     50 
     51 ; SI: [[ENDIF]]: ; %endif
     52 ; SI: s_add_i32 s{{[0-9]+}}, [[LOAD0]], [[LOAD1]]
     53 ; SI: buffer_store_dword
     54 ; SI-NEXT: s_endpgm
     55 define amdgpu_kernel void @sgpr_if_else_salu_br_opt(i32 addrspace(1)* %out, [8 x i32], i32 %a, [8 x i32], i32 %b, [8 x i32], i32 %c, [8 x i32], i32 %d, [8 x i32], i32 %e) {
     56 entry:
     57   %cmp0 = icmp eq i32 %a, 0
     58   br i1 %cmp0, label %if, label %else
     59 
     60 if:
     61   %add0 = add i32 %b, %c
     62   br label %endif
     63 
     64 else:
     65   %add1 = add i32 %d, %e
     66   br label %endif
     67 
     68 endif:
     69   %phi = phi i32 [%add0, %if], [%add1, %else]
     70   %add2 = add i32 %phi, %a
     71   store i32 %add2, i32 addrspace(1)* %out
     72   ret void
     73 }
     74 
     75 ; The two S_ADD instructions should write to different registers, since
     76 ; different threads will take different control flow paths.
     77 
     78 ; SI-LABEL: {{^}}sgpr_if_else_valu_br:
     79 ; SI: s_add_i32 [[SGPR:s[0-9]+]]
     80 ; SI-NOT: s_add_i32 [[SGPR]]
     81 
     82 define amdgpu_kernel void @sgpr_if_else_valu_br(i32 addrspace(1)* %out, float %a, i32 %b, i32 %c, i32 %d, i32 %e) {
     83 entry:
     84   %tid = call i32 @llvm.amdgcn.workitem.id.x() #0
     85   %tid_f = uitofp i32 %tid to float
     86   %tmp1 = fcmp ueq float %tid_f, 0.0
     87   br i1 %tmp1, label %if, label %else
     88 
     89 if:
     90   %tmp2 = add i32 %b, %c
     91   br label %endif
     92 
     93 else:
     94   %tmp3 = add i32 %d, %e
     95   br label %endif
     96 
     97 endif:
     98   %tmp4 = phi i32 [%tmp2, %if], [%tmp3, %else]
     99   store i32 %tmp4, i32 addrspace(1)* %out
    100   ret void
    101 }
    102 
    103 ; FIXME: Should write to different SGPR pairs instead of copying to
    104 ; VALU for i1 phi.
    105 
    106 ; SI-LABEL: {{^}}sgpr_if_else_valu_cmp_phi_br:
    107 ; SI: buffer_load_dword [[AVAL:v[0-9]+]]
    108 ; SI: v_cmp_gt_i32_e32 [[CMP_IF:vcc]], 0, [[AVAL]]
    109 ; SI: v_cndmask_b32_e64 [[V_CMP:v[0-9]+]], 0, -1, [[CMP_IF]]
    110 
    111 ; SI: BB{{[0-9]+}}_2:
    112 ; SI: buffer_load_dword [[AVAL:v[0-9]+]]
    113 ; SI: v_cmp_eq_u32_e32 [[CMP_ELSE:vcc]], 0, [[AVAL]]
    114 ; SI: v_cndmask_b32_e64 [[V_CMP]], 0, -1, [[CMP_ELSE]]
    115 
    116 ; SI: v_cmp_ne_u32_e32 [[CMP_CMP:vcc]], 0, [[V_CMP]]
    117 ; SI: v_cndmask_b32_e64 [[RESULT:v[0-9]+]], 0, -1, [[CMP_CMP]]
    118 ; SI: buffer_store_dword [[RESULT]]
    119 define amdgpu_kernel void @sgpr_if_else_valu_cmp_phi_br(i32 addrspace(1)* %out, i32 addrspace(1)* %a, i32 addrspace(1)* %b) {
    120 entry:
    121   %tid = call i32 @llvm.amdgcn.workitem.id.x() #0
    122   %tmp1 = icmp eq i32 %tid, 0
    123   br i1 %tmp1, label %if, label %else
    124 
    125 if:
    126   %gep.if = getelementptr i32, i32 addrspace(1)* %a, i32 %tid
    127   %a.val = load i32, i32 addrspace(1)* %gep.if
    128   %cmp.if = icmp eq i32 %a.val, 0
    129   br label %endif
    130 
    131 else:
    132   %gep.else = getelementptr i32, i32 addrspace(1)* %b, i32 %tid
    133   %b.val = load i32, i32 addrspace(1)* %gep.else
    134   %cmp.else = icmp slt i32 %b.val, 0
    135   br label %endif
    136 
    137 endif:
    138   %tmp4 = phi i1 [%cmp.if, %if], [%cmp.else, %else]
    139   %ext = sext i1 %tmp4 to i32
    140   store i32 %ext, i32 addrspace(1)* %out
    141   ret void
    142 }
    143 
    144 declare i32 @llvm.amdgcn.workitem.id.x() #0
    145 
    146 attributes #0 = { readnone }
    147