Home | History | Annotate | Download | only in X86
      1 ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
      2 ; RUN: llc < %s -mtriple=i386-unknown-unknown -mattr=+sse -O3 | FileCheck %s --check-prefixes=CHECK,X86
      3 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=-sse2,+sse -O3 | FileCheck %s --check-prefixes=CHECK,X64
      4 
      5 ; Tests for SSE1 and below, without SSE2+.
      6 
      7 ; PR7993
      8 ;define <4 x i32> @test3(<4 x i16> %a) nounwind {
      9 ;  %c = sext <4 x i16> %a to <4 x i32>             ; <<4 x i32>> [#uses=1]
     10 ;  ret <4 x i32> %c
     11 ;}
     12 
     13 ; This should not emit shuffles to populate the top 2 elements of the 4-element
     14 ; vector that this ends up returning.
     15 ; rdar://8368414
     16 define <2 x float> @test4(<2 x float> %A, <2 x float> %B) nounwind {
     17 ; CHECK-LABEL: test4:
     18 ; CHECK:       # %bb.0: # %entry
     19 ; CHECK-NEXT:    movaps %xmm0, %xmm2
     20 ; CHECK-NEXT:    shufps {{.*#+}} xmm2 = xmm2[1,1],xmm0[2,3]
     21 ; CHECK-NEXT:    addss %xmm1, %xmm0
     22 ; CHECK-NEXT:    shufps {{.*#+}} xmm1 = xmm1[1,1,2,3]
     23 ; CHECK-NEXT:    subss %xmm1, %xmm2
     24 ; CHECK-NEXT:    unpcklps {{.*#+}} xmm0 = xmm0[0],xmm2[0],xmm0[1],xmm2[1]
     25 ; CHECK-NEXT:    ret{{[l|q]}}
     26 entry:
     27   %tmp7 = extractelement <2 x float> %A, i32 0
     28   %tmp5 = extractelement <2 x float> %A, i32 1
     29   %tmp3 = extractelement <2 x float> %B, i32 0
     30   %tmp1 = extractelement <2 x float> %B, i32 1
     31   %add.r = fadd float %tmp7, %tmp3
     32   %add.i = fsub float %tmp5, %tmp1
     33   %tmp11 = insertelement <2 x float> undef, float %add.r, i32 0
     34   %tmp9 = insertelement <2 x float> %tmp11, float %add.i, i32 1
     35   ret <2 x float> %tmp9
     36 }
     37 
     38 ; We used to get stuck in type legalization for this example when lowering the
     39 ; vselect. With SSE1 v4f32 is a legal type but v4i1 (or any vector integer type)
     40 ; is not. We used to ping pong between splitting the vselect for the v4i
     41 ; condition operand and widening the resulting vselect for the v4f32 result.
     42 ; PR18036
     43 
     44 define <4 x float> @vselect(<4 x float>*%p, <4 x i32> %q) {
     45 ; X86-LABEL: vselect:
     46 ; X86:       # %bb.0: # %entry
     47 ; X86-NEXT:    cmpl $0, {{[0-9]+}}(%esp)
     48 ; X86-NEXT:    xorps %xmm0, %xmm0
     49 ; X86-NEXT:    je .LBB1_1
     50 ; X86-NEXT:  # %bb.2: # %entry
     51 ; X86-NEXT:    xorps %xmm1, %xmm1
     52 ; X86-NEXT:    cmpl $0, {{[0-9]+}}(%esp)
     53 ; X86-NEXT:    jne .LBB1_5
     54 ; X86-NEXT:  .LBB1_4:
     55 ; X86-NEXT:    movss {{.*#+}} xmm2 = mem[0],zero,zero,zero
     56 ; X86-NEXT:    cmpl $0, {{[0-9]+}}(%esp)
     57 ; X86-NEXT:    jne .LBB1_8
     58 ; X86-NEXT:  .LBB1_7:
     59 ; X86-NEXT:    movss {{.*#+}} xmm3 = mem[0],zero,zero,zero
     60 ; X86-NEXT:    unpcklps {{.*#+}} xmm2 = xmm2[0],xmm3[0],xmm2[1],xmm3[1]
     61 ; X86-NEXT:    cmpl $0, {{[0-9]+}}(%esp)
     62 ; X86-NEXT:    je .LBB1_10
     63 ; X86-NEXT:    jmp .LBB1_11
     64 ; X86-NEXT:  .LBB1_1:
     65 ; X86-NEXT:    movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
     66 ; X86-NEXT:    cmpl $0, {{[0-9]+}}(%esp)
     67 ; X86-NEXT:    je .LBB1_4
     68 ; X86-NEXT:  .LBB1_5: # %entry
     69 ; X86-NEXT:    xorps %xmm2, %xmm2
     70 ; X86-NEXT:    cmpl $0, {{[0-9]+}}(%esp)
     71 ; X86-NEXT:    je .LBB1_7
     72 ; X86-NEXT:  .LBB1_8: # %entry
     73 ; X86-NEXT:    xorps %xmm3, %xmm3
     74 ; X86-NEXT:    unpcklps {{.*#+}} xmm2 = xmm2[0],xmm3[0],xmm2[1],xmm3[1]
     75 ; X86-NEXT:    cmpl $0, {{[0-9]+}}(%esp)
     76 ; X86-NEXT:    jne .LBB1_11
     77 ; X86-NEXT:  .LBB1_10:
     78 ; X86-NEXT:    movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
     79 ; X86-NEXT:  .LBB1_11: # %entry
     80 ; X86-NEXT:    unpcklps {{.*#+}} xmm0 = xmm0[0],xmm1[0],xmm0[1],xmm1[1]
     81 ; X86-NEXT:    movlhps {{.*#+}} xmm0 = xmm0[0],xmm2[0]
     82 ; X86-NEXT:    retl
     83 ;
     84 ; X64-LABEL: vselect:
     85 ; X64:       # %bb.0: # %entry
     86 ; X64-NEXT:    testl %edx, %edx
     87 ; X64-NEXT:    xorps %xmm0, %xmm0
     88 ; X64-NEXT:    je .LBB1_1
     89 ; X64-NEXT:  # %bb.2: # %entry
     90 ; X64-NEXT:    xorps %xmm1, %xmm1
     91 ; X64-NEXT:    testl %ecx, %ecx
     92 ; X64-NEXT:    jne .LBB1_5
     93 ; X64-NEXT:  .LBB1_4:
     94 ; X64-NEXT:    movss {{.*#+}} xmm2 = mem[0],zero,zero,zero
     95 ; X64-NEXT:    testl %r8d, %r8d
     96 ; X64-NEXT:    jne .LBB1_8
     97 ; X64-NEXT:  .LBB1_7:
     98 ; X64-NEXT:    movss {{.*#+}} xmm3 = mem[0],zero,zero,zero
     99 ; X64-NEXT:    unpcklps {{.*#+}} xmm2 = xmm2[0],xmm3[0],xmm2[1],xmm3[1]
    100 ; X64-NEXT:    testl %esi, %esi
    101 ; X64-NEXT:    je .LBB1_10
    102 ; X64-NEXT:    jmp .LBB1_11
    103 ; X64-NEXT:  .LBB1_1:
    104 ; X64-NEXT:    movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
    105 ; X64-NEXT:    testl %ecx, %ecx
    106 ; X64-NEXT:    je .LBB1_4
    107 ; X64-NEXT:  .LBB1_5: # %entry
    108 ; X64-NEXT:    xorps %xmm2, %xmm2
    109 ; X64-NEXT:    testl %r8d, %r8d
    110 ; X64-NEXT:    je .LBB1_7
    111 ; X64-NEXT:  .LBB1_8: # %entry
    112 ; X64-NEXT:    xorps %xmm3, %xmm3
    113 ; X64-NEXT:    unpcklps {{.*#+}} xmm2 = xmm2[0],xmm3[0],xmm2[1],xmm3[1]
    114 ; X64-NEXT:    testl %esi, %esi
    115 ; X64-NEXT:    jne .LBB1_11
    116 ; X64-NEXT:  .LBB1_10:
    117 ; X64-NEXT:    movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
    118 ; X64-NEXT:  .LBB1_11: # %entry
    119 ; X64-NEXT:    unpcklps {{.*#+}} xmm0 = xmm0[0],xmm1[0],xmm0[1],xmm1[1]
    120 ; X64-NEXT:    movlhps {{.*#+}} xmm0 = xmm0[0],xmm2[0]
    121 ; X64-NEXT:    retq
    122 entry:
    123   %a1 = icmp eq <4 x i32> %q, zeroinitializer
    124   %a14 = select <4 x i1> %a1, <4 x float> <float 1.000000e+00, float 2.000000e+00, float 3.000000e+00, float 4.000000e+0> , <4 x float> zeroinitializer
    125   ret <4 x float> %a14
    126 }
    127 
    128 ; v4i32 isn't legal for SSE1, but this should be cmpps.
    129 
    130 define <4 x float> @PR28044(<4 x float> %a0, <4 x float> %a1) nounwind {
    131 ; CHECK-LABEL: PR28044:
    132 ; CHECK:       # %bb.0:
    133 ; CHECK-NEXT:    cmpeqps %xmm1, %xmm0
    134 ; CHECK-NEXT:    ret{{[l|q]}}
    135   %cmp = fcmp oeq <4 x float> %a0, %a1
    136   %sext = sext <4 x i1> %cmp to <4 x i32>
    137   %res = bitcast <4 x i32> %sext to <4 x float>
    138   ret <4 x float> %res
    139 }
    140 
    141 ; Don't crash trying to do the impossible: an integer vector comparison doesn't exist, so we must scalarize.
    142 ; https://llvm.org/bugs/show_bug.cgi?id=30512
    143 
    144 define <4 x i32> @PR30512(<4 x i32> %x, <4 x i32> %y) nounwind {
    145 ; X86-LABEL: PR30512:
    146 ; X86:       # %bb.0:
    147 ; X86-NEXT:    pushl %ebx
    148 ; X86-NEXT:    pushl %edi
    149 ; X86-NEXT:    pushl %esi
    150 ; X86-NEXT:    subl $16, %esp
    151 ; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax
    152 ; X86-NEXT:    movl {{[0-9]+}}(%esp), %ecx
    153 ; X86-NEXT:    movl {{[0-9]+}}(%esp), %edx
    154 ; X86-NEXT:    movl {{[0-9]+}}(%esp), %esi
    155 ; X86-NEXT:    movl {{[0-9]+}}(%esp), %edi
    156 ; X86-NEXT:    xorl %ebx, %ebx
    157 ; X86-NEXT:    cmpl {{[0-9]+}}(%esp), %edi
    158 ; X86-NEXT:    sete %bl
    159 ; X86-NEXT:    negl %ebx
    160 ; X86-NEXT:    movl %ebx, {{[0-9]+}}(%esp)
    161 ; X86-NEXT:    xorl %ebx, %ebx
    162 ; X86-NEXT:    cmpl {{[0-9]+}}(%esp), %esi
    163 ; X86-NEXT:    sete %bl
    164 ; X86-NEXT:    negl %ebx
    165 ; X86-NEXT:    movl %ebx, {{[0-9]+}}(%esp)
    166 ; X86-NEXT:    xorl %ebx, %ebx
    167 ; X86-NEXT:    cmpl {{[0-9]+}}(%esp), %edx
    168 ; X86-NEXT:    sete %bl
    169 ; X86-NEXT:    negl %ebx
    170 ; X86-NEXT:    movl %ebx, {{[0-9]+}}(%esp)
    171 ; X86-NEXT:    xorl %edx, %edx
    172 ; X86-NEXT:    cmpl {{[0-9]+}}(%esp), %ecx
    173 ; X86-NEXT:    sete %dl
    174 ; X86-NEXT:    negl %edx
    175 ; X86-NEXT:    movl %edx, (%esp)
    176 ; X86-NEXT:    movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
    177 ; X86-NEXT:    movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
    178 ; X86-NEXT:    unpcklps {{.*#+}} xmm1 = xmm1[0],xmm0[0],xmm1[1],xmm0[1]
    179 ; X86-NEXT:    movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
    180 ; X86-NEXT:    movss {{.*#+}} xmm2 = mem[0],zero,zero,zero
    181 ; X86-NEXT:    unpcklps {{.*#+}} xmm2 = xmm2[0],xmm0[0],xmm2[1],xmm0[1]
    182 ; X86-NEXT:    movlhps {{.*#+}} xmm2 = xmm2[0],xmm1[0]
    183 ; X86-NEXT:    andps {{\.LCPI.*}}, %xmm2
    184 ; X86-NEXT:    movaps %xmm2, (%eax)
    185 ; X86-NEXT:    addl $16, %esp
    186 ; X86-NEXT:    popl %esi
    187 ; X86-NEXT:    popl %edi
    188 ; X86-NEXT:    popl %ebx
    189 ; X86-NEXT:    retl $4
    190 ;
    191 ; X64-LABEL: PR30512:
    192 ; X64:       # %bb.0:
    193 ; X64-NEXT:    xorl %eax, %eax
    194 ; X64-NEXT:    cmpl {{[0-9]+}}(%rsp), %r8d
    195 ; X64-NEXT:    sete %al
    196 ; X64-NEXT:    negl %eax
    197 ; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp)
    198 ; X64-NEXT:    xorl %eax, %eax
    199 ; X64-NEXT:    cmpl {{[0-9]+}}(%rsp), %ecx
    200 ; X64-NEXT:    sete %al
    201 ; X64-NEXT:    negl %eax
    202 ; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp)
    203 ; X64-NEXT:    xorl %eax, %eax
    204 ; X64-NEXT:    cmpl {{[0-9]+}}(%rsp), %edx
    205 ; X64-NEXT:    sete %al
    206 ; X64-NEXT:    negl %eax
    207 ; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp)
    208 ; X64-NEXT:    xorl %eax, %eax
    209 ; X64-NEXT:    cmpl %r9d, %esi
    210 ; X64-NEXT:    sete %al
    211 ; X64-NEXT:    negl %eax
    212 ; X64-NEXT:    movl %eax, -{{[0-9]+}}(%rsp)
    213 ; X64-NEXT:    movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
    214 ; X64-NEXT:    movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
    215 ; X64-NEXT:    unpcklps {{.*#+}} xmm1 = xmm1[0],xmm0[0],xmm1[1],xmm0[1]
    216 ; X64-NEXT:    movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
    217 ; X64-NEXT:    movss {{.*#+}} xmm2 = mem[0],zero,zero,zero
    218 ; X64-NEXT:    unpcklps {{.*#+}} xmm2 = xmm2[0],xmm0[0],xmm2[1],xmm0[1]
    219 ; X64-NEXT:    movlhps {{.*#+}} xmm2 = xmm2[0],xmm1[0]
    220 ; X64-NEXT:    andps {{.*}}(%rip), %xmm2
    221 ; X64-NEXT:    movaps %xmm2, (%rdi)
    222 ; X64-NEXT:    movq %rdi, %rax
    223 ; X64-NEXT:    retq
    224   %cmp = icmp eq <4 x i32> %x, %y
    225   %zext = zext <4 x i1> %cmp to <4 x i32>
    226   ret <4 x i32> %zext
    227 }
    228 
    229 ; Fragile test warning - we need to induce the generation of a vselect
    230 ; post-legalization to cause the crash seen in:
    231 ; https://llvm.org/bugs/show_bug.cgi?id=31672
    232 ; Is there a way to do that without an unsafe/fast sqrt intrinsic call?
    233 ;
    234 ; We now no longer try to lower sqrt using rsqrt with SSE1 only as the
    235 ; v4i32 vselect mentioned above should never have been created. We ended up
    236 ; scalarizing it anyway.
    237 
    238 define <2 x float> @PR31672() #0 {
    239 ; X86-LABEL: PR31672:
    240 ; X86:       # %bb.0:
    241 ; X86-NEXT:    sqrtps {{\.LCPI.*}}, %xmm0
    242 ; X86-NEXT:    retl
    243 ;
    244 ; X64-LABEL: PR31672:
    245 ; X64:       # %bb.0:
    246 ; X64-NEXT:    sqrtps {{.*}}(%rip), %xmm0
    247 ; X64-NEXT:    retq
    248   %t0 = call fast <2 x float> @llvm.sqrt.v2f32(<2 x float> <float 42.0, float 3.0>)
    249   ret <2 x float> %t0
    250 }
    251 
    252 declare <2 x float> @llvm.sqrt.v2f32(<2 x float>) #1
    253 
    254 attributes #0 = { nounwind "unsafe-fp-math"="true" }
    255 
    256