Home | History | Annotate | Download | only in LoopSimplify
      1 ; RUN: opt -basicaa -loop-rotate -licm -instcombine -indvars -loop-unroll -S %s | FileCheck %s
      2 ;
      3 ; PR18361: ScalarEvolution::getAddRecExpr():
      4 ;          Assertion `isLoopInvariant(Operands[i],...
      5 ;
      6 ; After a series of loop optimizations, SCEV's LoopDispositions grow stale.
      7 ; In particular, LoopSimplify hoists %cmp4, resulting in this SCEV for %add:
      8 ; {(zext i1 %cmp4 to i32),+,1}<nw><%for.cond1.preheader>
      9 ;
     10 ; When recomputing the SCEV for %ashr, we truncate the operands to get:
     11 ; (zext i1 %cmp4 to i16)
     12 ;
     13 ; This SCEV was never mapped to a value so never invalidated. It's
     14 ; loop disposition is still marked as non-loop-invariant, which is
     15 ; inconsistent with the AddRec.
     16 
     17 target datalayout = "e-i64:64-f80:128-n8:16:32:64-S128"
     18 target triple = "x86_64-apple-macosx"
     19 
     20 @d = common global i32 0, align 4
     21 @a = common global i32 0, align 4
     22 @c = common global i32 0, align 4
     23 @b = common global i32 0, align 4
     24 
     25 ; Check that the def-use chain that leads to the bad SCEV is still
     26 ; there.
     27 ;
     28 ; CHECK-LABEL: @foo
     29 ; CHECK-LABEL: entry:
     30 ; CHECK-LABEL: for.cond1.preheader:
     31 ; CHECK-LABEL: for.body3:
     32 ; CHECK: %cmp4.le.le
     33 ; CHECK: %conv.le.le = zext i1 %cmp4.le.le to i32
     34 ; CHECK: %xor.le.le = xor i32 %conv6.le.le, 1
     35 define void @foo() {
     36 entry:
     37   br label %for.cond
     38 
     39 for.cond:                                         ; preds = %for.inc7, %entry
     40   %storemerge = phi i32 [ 0, %entry ], [ %inc8, %for.inc7 ]
     41   %f.0 = phi i32 [ undef, %entry ], [ %f.1, %for.inc7 ]
     42   store i32 %storemerge, i32* @d, align 4
     43   %cmp = icmp slt i32 %storemerge, 1
     44   br i1 %cmp, label %for.cond1, label %for.end9
     45 
     46 for.cond1:                                        ; preds = %for.cond, %for.body3
     47   %storemerge1 = phi i32 [ %inc, %for.body3 ], [ 0, %for.cond ]
     48   %f.1 = phi i32 [ %xor, %for.body3 ], [ %f.0, %for.cond ]
     49   store i32 %storemerge1, i32* @a, align 4
     50   %cmp2 = icmp slt i32 %storemerge1, 1
     51   br i1 %cmp2, label %for.body3, label %for.inc7
     52 
     53 for.body3:                                        ; preds = %for.cond1
     54   %0 = load i32, i32* @c, align 4
     55   %cmp4 = icmp sge i32 %storemerge1, %0
     56   %conv = zext i1 %cmp4 to i32
     57   %1 = load i32, i32* @d, align 4
     58   %add = add nsw i32 %conv, %1
     59   %sext = shl i32 %add, 16
     60   %conv6 = ashr exact i32 %sext, 16
     61   %xor = xor i32 %conv6, 1
     62   %inc = add nsw i32 %storemerge1, 1
     63   br label %for.cond1
     64 
     65 for.inc7:                                         ; preds = %for.cond1
     66   %2 = load i32, i32* @d, align 4
     67   %inc8 = add nsw i32 %2, 1
     68   br label %for.cond
     69 
     70 for.end9:                                         ; preds = %for.cond
     71   %cmp10 = icmp sgt i32 %f.0, 0
     72   br i1 %cmp10, label %if.then, label %if.end
     73 
     74 if.then:                                          ; preds = %for.end9
     75   store i32 0, i32* @b, align 4
     76   br label %if.end
     77 
     78 if.end:                                           ; preds = %if.then, %for.end9
     79   ret void
     80 }
     81