Home | History | Annotate | Download | only in X86
      1 ; RUN: llc < %s -mtriple=x86_64-linux -mattr=+sse2 -mcpu=nehalem | FileCheck %s --check-prefix=SSE
      2 ; RUN: llc < %s -mtriple=x86_64-win32 -mattr=+sse2 -mcpu=nehalem | FileCheck %s --check-prefix=SSE
      3 ; RUN: llc < %s -mtriple=x86_64-win32 -mattr=+avx -mcpu=corei7-avx | FileCheck %s --check-prefix=AVX
      4 
      5 define double @t1(float* nocapture %x) nounwind readonly ssp {
      6 entry:
      7 ; SSE-LABEL: t1:
      8 ; SSE: movss ([[A0:%rdi|%rcx]]), %xmm0
      9 ; SSE: cvtss2sd %xmm0, %xmm0
     10 
     11   %0 = load float, float* %x, align 4
     12   %1 = fpext float %0 to double
     13   ret double %1
     14 }
     15 
     16 define float @t2(double* nocapture %x) nounwind readonly ssp optsize {
     17 entry:
     18 ; SSE-LABEL: t2:
     19 ; SSE: cvtsd2ss ([[A0]]), %xmm0
     20   %0 = load double, double* %x, align 8
     21   %1 = fptrunc double %0 to float
     22   ret float %1
     23 }
     24 
     25 define float @squirtf(float* %x) nounwind {
     26 entry:
     27 ; SSE-LABEL: squirtf:
     28 ; SSE: movss ([[A0]]), %xmm0
     29 ; SSE: sqrtss %xmm0, %xmm0
     30   %z = load float, float* %x
     31   %t = call float @llvm.sqrt.f32(float %z)
     32   ret float %t
     33 }
     34 
     35 define double @squirt(double* %x) nounwind {
     36 entry:
     37 ; SSE-LABEL: squirt:
     38 ; SSE: movsd ([[A0]]), %xmm0
     39 ; SSE: sqrtsd %xmm0, %xmm0
     40   %z = load double, double* %x
     41   %t = call double @llvm.sqrt.f64(double %z)
     42   ret double %t
     43 }
     44 
     45 define float @squirtf_size(float* %x) nounwind optsize {
     46 entry:
     47 ; SSE-LABEL: squirtf_size:
     48 ; SSE: sqrtss ([[A0]]), %xmm0
     49   %z = load float, float* %x
     50   %t = call float @llvm.sqrt.f32(float %z)
     51   ret float %t
     52 }
     53 
     54 define double @squirt_size(double* %x) nounwind optsize {
     55 entry:
     56 ; SSE-LABEL: squirt_size:
     57 ; SSE: sqrtsd ([[A0]]), %xmm0
     58   %z = load double, double* %x
     59   %t = call double @llvm.sqrt.f64(double %z)
     60   ret double %t
     61 }
     62 
     63 declare float @llvm.sqrt.f32(float)
     64 declare double @llvm.sqrt.f64(double)
     65 
     66 ; SSE-LABEL: loopdep1
     67 ; SSE: for.body{{$}}
     68 ;
     69 ; This loop contains two cvtsi2ss instructions that update the same xmm
     70 ; register.  Verify that the execution dependency fix pass breaks those
     71 ; dependencies by inserting xorps instructions.
     72 ;
     73 ; If the register allocator chooses different registers for the two cvtsi2ss
     74 ; instructions, they are still dependent on themselves.
     75 ; SSE: xorps [[XMM1:%xmm[0-9]+]]
     76 ; SSE: , [[XMM1]]
     77 ; SSE: cvtsi2ssl %{{.*}}, [[XMM1]]
     78 ; SSE: xorps [[XMM2:%xmm[0-9]+]]
     79 ; SSE: , [[XMM2]]
     80 ; SSE: cvtsi2ssl %{{.*}}, [[XMM2]]
     81 ;
     82 define float @loopdep1(i32 %m) nounwind uwtable readnone ssp {
     83 entry:
     84   %tobool3 = icmp eq i32 %m, 0
     85   br i1 %tobool3, label %for.end, label %for.body
     86 
     87 for.body:                                         ; preds = %entry, %for.body
     88   %m.addr.07 = phi i32 [ %dec, %for.body ], [ %m, %entry ]
     89   %s1.06 = phi float [ %add, %for.body ], [ 0.000000e+00, %entry ]
     90   %s2.05 = phi float [ %add2, %for.body ], [ 0.000000e+00, %entry ]
     91   %n.04 = phi i32 [ %inc, %for.body ], [ 1, %entry ]
     92   %conv = sitofp i32 %n.04 to float
     93   %add = fadd float %s1.06, %conv
     94   %conv1 = sitofp i32 %m.addr.07 to float
     95   %add2 = fadd float %s2.05, %conv1
     96   %inc = add nsw i32 %n.04, 1
     97   %dec = add nsw i32 %m.addr.07, -1
     98   %tobool = icmp eq i32 %dec, 0
     99   br i1 %tobool, label %for.end, label %for.body
    100 
    101 for.end:                                          ; preds = %for.body, %entry
    102   %s1.0.lcssa = phi float [ 0.000000e+00, %entry ], [ %add, %for.body ]
    103   %s2.0.lcssa = phi float [ 0.000000e+00, %entry ], [ %add2, %for.body ]
    104   %sub = fsub float %s1.0.lcssa, %s2.0.lcssa
    105   ret float %sub
    106 }
    107 
    108 ; rdar:15221834 False AVX register dependencies cause 5x slowdown on
    109 ; flops-6. Make sure the unused register read by vcvtsi2sdq is zeroed
    110 ; to avoid cyclic dependence on a write to the same register in a
    111 ; previous iteration.
    112 
    113 ; AVX-LABEL: loopdep2:
    114 ; AVX-LABEL: %loop
    115 ; AVX: vxorps %[[REG:xmm.]], %{{xmm.}}, %{{xmm.}}
    116 ; AVX: vcvtsi2sdq %{{r[0-9a-x]+}}, %[[REG]], %{{xmm.}}
    117 ; SSE-LABEL: loopdep2:
    118 ; SSE-LABEL: %loop
    119 ; SSE: xorps %[[REG:xmm.]], %[[REG]]
    120 ; SSE: cvtsi2sdq %{{r[0-9a-x]+}}, %[[REG]]
    121 define i64 @loopdep2(i64* nocapture %x, double* nocapture %y) nounwind {
    122 entry:
    123   %vx = load i64, i64* %x
    124   br label %loop
    125 loop:
    126   %i = phi i64 [ 1, %entry ], [ %inc, %loop ]
    127   %s1 = phi i64 [ %vx, %entry ], [ %s2, %loop ]
    128   %fi = sitofp i64 %i to double
    129   %vy = load double, double* %y
    130   %fipy = fadd double %fi, %vy
    131   %iipy = fptosi double %fipy to i64
    132   %s2 = add i64 %s1, %iipy
    133   %inc = add nsw i64 %i, 1
    134   %exitcond = icmp eq i64 %inc, 156250000
    135   br i1 %exitcond, label %ret, label %loop
    136 ret:
    137   ret i64 %s2
    138 }
    139 
    140 ; This loop contains a cvtsi2sd instruction that has a loop-carried
    141 ; false dependency on an xmm that is modified by other scalar instructions
    142 ; that follow it in the loop. Additionally, the source of convert is a
    143 ; memory operand. Verify the execution dependency fix pass breaks this
    144 ; dependency by inserting a xor before the convert.
    145 @x = common global [1024 x double] zeroinitializer, align 16
    146 @y = common global [1024 x double] zeroinitializer, align 16
    147 @z = common global [1024 x double] zeroinitializer, align 16
    148 @w = common global [1024 x double] zeroinitializer, align 16
    149 @v = common global [1024 x i32] zeroinitializer, align 16
    150 
    151 define void @loopdep3() {
    152 entry:
    153   br label %for.cond1.preheader
    154 
    155 for.cond1.preheader:                              ; preds = %for.inc14, %entry
    156   %i.025 = phi i32 [ 0, %entry ], [ %inc15, %for.inc14 ]
    157   br label %for.body3
    158 
    159 for.body3:
    160   %indvars.iv = phi i64 [ 0, %for.cond1.preheader ], [ %indvars.iv.next, %for.body3 ]
    161   %arrayidx = getelementptr inbounds [1024 x i32], [1024 x i32]* @v, i64 0, i64 %indvars.iv
    162   %0 = load i32, i32* %arrayidx, align 4
    163   %conv = sitofp i32 %0 to double
    164   %arrayidx5 = getelementptr inbounds [1024 x double], [1024 x double]* @x, i64 0, i64 %indvars.iv
    165   %1 = load double, double* %arrayidx5, align 8
    166   %mul = fmul double %conv, %1
    167   %arrayidx7 = getelementptr inbounds [1024 x double], [1024 x double]* @y, i64 0, i64 %indvars.iv
    168   %2 = load double, double* %arrayidx7, align 8
    169   %mul8 = fmul double %mul, %2
    170   %arrayidx10 = getelementptr inbounds [1024 x double], [1024 x double]* @z, i64 0, i64 %indvars.iv
    171   %3 = load double, double* %arrayidx10, align 8
    172   %mul11 = fmul double %mul8, %3
    173   %arrayidx13 = getelementptr inbounds [1024 x double], [1024 x double]* @w, i64 0, i64 %indvars.iv
    174   store double %mul11, double* %arrayidx13, align 8
    175   %indvars.iv.next = add nuw nsw i64 %indvars.iv, 1
    176   %exitcond = icmp eq i64 %indvars.iv.next, 1024
    177   br i1 %exitcond, label %for.inc14, label %for.body3
    178 
    179 for.inc14:                                        ; preds = %for.body3
    180   %inc15 = add nsw i32 %i.025, 1
    181   %exitcond26 = icmp eq i32 %inc15, 100000
    182   br i1 %exitcond26, label %for.end16, label %for.cond1.preheader
    183 
    184 for.end16:                                        ; preds = %for.inc14
    185   ret void
    186 
    187 ;SSE-LABEL:@loopdep3
    188 ;SSE: xorps [[XMM0:%xmm[0-9]+]], [[XMM0]]
    189 ;SSE-NEXT: cvtsi2sdl {{.*}}, [[XMM0]]
    190 ;SSE-NEXT: mulsd {{.*}}, [[XMM0]]
    191 ;SSE-NEXT: mulsd {{.*}}, [[XMM0]]
    192 ;SSE-NEXT: mulsd {{.*}}, [[XMM0]]
    193 ;SSE-NEXT: movsd [[XMM0]],
    194 ;AVX-LABEL:@loopdep3
    195 ;AVX: vxorps [[XMM0:%xmm[0-9]+]], [[XMM0]]
    196 ;AVX-NEXT: vcvtsi2sdl {{.*}}, [[XMM0]], [[XMM0]]
    197 ;AVX-NEXT: vmulsd {{.*}}, [[XMM0]], [[XMM0]]
    198 ;AVX-NEXT: vmulsd {{.*}}, [[XMM0]], [[XMM0]]
    199 ;AVX-NEXT: vmulsd {{.*}}, [[XMM0]], [[XMM0]]
    200 ;AVX-NEXT: vmovsd [[XMM0]],
    201 }
    202