Home | History | Annotate | Download | only in X86
      1 ;; X's live range extends beyond the shift, so the register allocator
      2 ;; cannot coalesce it with Y.  Because of this, a copy needs to be
      3 ;; emitted before the shift to save the register value before it is
      4 ;; clobbered.  However, this copy is not needed if the register
      5 ;; allocator turns the shift into an LEA.  This also occurs for ADD.
      6 
      7 ; Check that the shift gets turned into an LEA.
      8 ; RUN: llc < %s -mcpu=generic -mtriple=x86_64-apple-darwin | FileCheck %s
      9 
     10 @G = external global i32
     11 
     12 define i32 @test1(i32 %X) nounwind {
     13 ; CHECK-LABEL: test1:
     14 ; CHECK-NOT: mov
     15 ; CHECK: leal 1(%rdi)
     16         %Z = add i32 %X, 1
     17         store volatile i32 %Z, i32* @G
     18         ret i32 %X
     19 }
     20 
     21 ; rdar://8977508
     22 ; The second add should not be transformed to leal nor should it be
     23 ; commutted (which would require inserting a copy).
     24 define i32 @test2(i32 inreg %a, i32 inreg %b, i32 %c, i32 %d) nounwind {
     25 entry:
     26 ; CHECK-LABEL: test2:
     27 ; CHECK: leal
     28 ; CHECK-NEXT: addl
     29 ; CHECK-NEXT: addl
     30 ; CHECK-NEXT: ret
     31  %add = add i32 %b, %a
     32  %add3 = add i32 %add, %c
     33  %add5 = add i32 %add3, %d
     34  ret i32 %add5
     35 }
     36 
     37 ; rdar://9002648
     38 define i64 @test3(i64 %x) nounwind readnone ssp {
     39 entry:
     40 ; CHECK-LABEL: test3:
     41 ; CHECK: leaq (%rdi,%rdi), %rax
     42 ; CHECK-NOT: addq
     43 ; CHECK-NEXT: ret
     44   %0 = shl i64 %x, 1
     45   ret i64 %0
     46 }
     47 
     48 @global = external global i32, align 4
     49 @global2 = external global i64, align 8
     50 
     51 ; Test that liveness is properly updated and we do not encounter the
     52 ; assert/crash from http://llvm.org/PR28301
     53 ; CHECK-LABEL: ham
     54 define void @ham() {
     55 bb:
     56   br label %bb1
     57 
     58 bb1:
     59   %tmp = phi i64 [ %tmp40, %bb9 ], [ 0, %bb ]
     60   %tmp2 = phi i32 [ %tmp39, %bb9 ], [ 0, %bb ]
     61   %tmp3 = icmp sgt i32 undef, 10
     62   br i1 %tmp3, label %bb2, label %bb3
     63 
     64 bb2:
     65   %tmp6 = load i32, i32* @global, align 4
     66   %tmp8 = add nsw i32 %tmp6, %tmp2
     67   %tmp9 = sext i32 %tmp8 to i64
     68   br label %bb6
     69 
     70 bb3:
     71 ; CHECK: subl %e[[REG0:[a-z0-9]+]],
     72 ; CHECK: leaq 4({{%[a-z0-9]+}}), %r[[REG0]]
     73   %tmp14 = phi i64 [ %tmp15, %bb5 ], [ 0, %bb1 ]
     74   %tmp15 = add nuw i64 %tmp14, 4
     75   %tmp16 = trunc i64 %tmp14 to i32
     76   %tmp17 = sub i32 %tmp2, %tmp16
     77   br label %bb4
     78 
     79 bb4:
     80   %tmp20 = phi i64 [ %tmp14, %bb3 ], [ %tmp34, %bb5 ]
     81   %tmp28 = icmp eq i32 %tmp17, 0
     82   br i1 %tmp28, label %bb5, label %bb8
     83 
     84 bb5:
     85   %tmp34 = add nuw nsw i64 %tmp20, 1
     86   %tmp35 = icmp slt i64 %tmp34, %tmp15
     87   br i1 %tmp35, label %bb4, label %bb3
     88 
     89 bb6:
     90   store volatile i64 %tmp, i64* @global2, align 8
     91   store volatile i64 %tmp9, i64* @global2, align 8
     92   store volatile i32 %tmp6, i32* @global, align 4
     93   %tmp45 = icmp slt i32 undef, undef
     94   br i1 %tmp45, label %bb6, label %bb9
     95 
     96 bb8:
     97   unreachable
     98 
     99 bb9:
    100   %tmp39 = add nuw nsw i32 %tmp2, 4
    101   %tmp40 = add nuw i64 %tmp, 4
    102   br label %bb1
    103 }
    104