Home | History | Annotate | Download | only in X86
      1 ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
      2 ; RUN: llc < %s -mtriple=i386-unknown-unknown -mattr=+sse4a -show-mc-encoding | FileCheck %s --check-prefixes=CHECK,X86,SSE,X86-SSE
      3 ; RUN: llc < %s -mtriple=i386-unknown-unknown -mattr=+sse4a,+avx -show-mc-encoding | FileCheck %s --check-prefixes=CHECK,X86,AVX,X86-AVX
      4 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+sse4a -show-mc-encoding | FileCheck %s --check-prefixes=CHECK,X64,SSE,X64-SSE
      5 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+sse4a,+avx -show-mc-encoding | FileCheck %s --check-prefixes=CHECK,X64,AVX,X64-AVX
      6 
      7 define <2 x i64> @test_extrqi(<2 x i64> %x) nounwind uwtable ssp {
      8 ; CHECK-LABEL: test_extrqi:
      9 ; CHECK:       # %bb.0:
     10 ; CHECK-NEXT:    extrq $2, $3, %xmm0 # encoding: [0x66,0x0f,0x78,0xc0,0x03,0x02]
     11 ; CHECK-NEXT:    ret{{[l|q]}} # encoding: [0xc3]
     12   %1 = tail call <2 x i64> @llvm.x86.sse4a.extrqi(<2 x i64> %x, i8 3, i8 2)
     13   ret <2 x i64> %1
     14 }
     15 
     16 define <2 x i64> @test_extrqi_domain(<2 x i64> *%p) nounwind uwtable ssp {
     17 ; X86-SSE-LABEL: test_extrqi_domain:
     18 ; X86-SSE:       # %bb.0:
     19 ; X86-SSE-NEXT:    movl {{[0-9]+}}(%esp), %eax # encoding: [0x8b,0x44,0x24,0x04]
     20 ; X86-SSE-NEXT:    movdqa (%eax), %xmm0 # encoding: [0x66,0x0f,0x6f,0x00]
     21 ; X86-SSE-NEXT:    extrq $2, $3, %xmm0 # encoding: [0x66,0x0f,0x78,0xc0,0x03,0x02]
     22 ; X86-SSE-NEXT:    retl # encoding: [0xc3]
     23 ;
     24 ; X86-AVX-LABEL: test_extrqi_domain:
     25 ; X86-AVX:       # %bb.0:
     26 ; X86-AVX-NEXT:    movl {{[0-9]+}}(%esp), %eax # encoding: [0x8b,0x44,0x24,0x04]
     27 ; X86-AVX-NEXT:    vmovdqa (%eax), %xmm0 # encoding: [0xc5,0xf9,0x6f,0x00]
     28 ; X86-AVX-NEXT:    extrq $2, $3, %xmm0 # encoding: [0x66,0x0f,0x78,0xc0,0x03,0x02]
     29 ; X86-AVX-NEXT:    retl # encoding: [0xc3]
     30 ;
     31 ; X64-SSE-LABEL: test_extrqi_domain:
     32 ; X64-SSE:       # %bb.0:
     33 ; X64-SSE-NEXT:    movdqa (%rdi), %xmm0 # encoding: [0x66,0x0f,0x6f,0x07]
     34 ; X64-SSE-NEXT:    extrq $2, $3, %xmm0 # encoding: [0x66,0x0f,0x78,0xc0,0x03,0x02]
     35 ; X64-SSE-NEXT:    retq # encoding: [0xc3]
     36 ;
     37 ; X64-AVX-LABEL: test_extrqi_domain:
     38 ; X64-AVX:       # %bb.0:
     39 ; X64-AVX-NEXT:    vmovdqa (%rdi), %xmm0 # encoding: [0xc5,0xf9,0x6f,0x07]
     40 ; X64-AVX-NEXT:    extrq $2, $3, %xmm0 # encoding: [0x66,0x0f,0x78,0xc0,0x03,0x02]
     41 ; X64-AVX-NEXT:    retq # encoding: [0xc3]
     42   %1 = load <2 x i64>, <2 x i64> *%p
     43   %2 = tail call <2 x i64> @llvm.x86.sse4a.extrqi(<2 x i64> %1, i8 3, i8 2)
     44   ret <2 x i64> %2
     45 }
     46 
     47 declare <2 x i64> @llvm.x86.sse4a.extrqi(<2 x i64>, i8, i8) nounwind
     48 
     49 define <2 x i64> @test_extrq(<2 x i64> %x, <2 x i64> %y) nounwind uwtable ssp {
     50 ; CHECK-LABEL: test_extrq:
     51 ; CHECK:       # %bb.0:
     52 ; CHECK-NEXT:    extrq %xmm1, %xmm0 # encoding: [0x66,0x0f,0x79,0xc1]
     53 ; CHECK-NEXT:    ret{{[l|q]}} # encoding: [0xc3]
     54   %1 = bitcast <2 x i64> %y to <16 x i8>
     55   %2 = tail call <2 x i64> @llvm.x86.sse4a.extrq(<2 x i64> %x, <16 x i8> %1) nounwind
     56   ret <2 x i64> %2
     57 }
     58 
     59 define <2 x i64> @test_extrq_domain(<2 x i64> *%p, <2 x i64> %y) nounwind uwtable ssp {
     60 ; X86-SSE-LABEL: test_extrq_domain:
     61 ; X86-SSE:       # %bb.0:
     62 ; X86-SSE-NEXT:    movl {{[0-9]+}}(%esp), %eax # encoding: [0x8b,0x44,0x24,0x04]
     63 ; X86-SSE-NEXT:    movdqa (%eax), %xmm1 # encoding: [0x66,0x0f,0x6f,0x08]
     64 ; X86-SSE-NEXT:    extrq %xmm0, %xmm1 # encoding: [0x66,0x0f,0x79,0xc8]
     65 ; X86-SSE-NEXT:    movdqa %xmm1, %xmm0 # encoding: [0x66,0x0f,0x6f,0xc1]
     66 ; X86-SSE-NEXT:    retl # encoding: [0xc3]
     67 ;
     68 ; X86-AVX-LABEL: test_extrq_domain:
     69 ; X86-AVX:       # %bb.0:
     70 ; X86-AVX-NEXT:    movl {{[0-9]+}}(%esp), %eax # encoding: [0x8b,0x44,0x24,0x04]
     71 ; X86-AVX-NEXT:    vmovdqa (%eax), %xmm1 # encoding: [0xc5,0xf9,0x6f,0x08]
     72 ; X86-AVX-NEXT:    extrq %xmm0, %xmm1 # encoding: [0x66,0x0f,0x79,0xc8]
     73 ; X86-AVX-NEXT:    vmovdqa %xmm1, %xmm0 # encoding: [0xc5,0xf9,0x6f,0xc1]
     74 ; X86-AVX-NEXT:    retl # encoding: [0xc3]
     75 ;
     76 ; X64-SSE-LABEL: test_extrq_domain:
     77 ; X64-SSE:       # %bb.0:
     78 ; X64-SSE-NEXT:    movdqa (%rdi), %xmm1 # encoding: [0x66,0x0f,0x6f,0x0f]
     79 ; X64-SSE-NEXT:    extrq %xmm0, %xmm1 # encoding: [0x66,0x0f,0x79,0xc8]
     80 ; X64-SSE-NEXT:    movdqa %xmm1, %xmm0 # encoding: [0x66,0x0f,0x6f,0xc1]
     81 ; X64-SSE-NEXT:    retq # encoding: [0xc3]
     82 ;
     83 ; X64-AVX-LABEL: test_extrq_domain:
     84 ; X64-AVX:       # %bb.0:
     85 ; X64-AVX-NEXT:    vmovdqa (%rdi), %xmm1 # encoding: [0xc5,0xf9,0x6f,0x0f]
     86 ; X64-AVX-NEXT:    extrq %xmm0, %xmm1 # encoding: [0x66,0x0f,0x79,0xc8]
     87 ; X64-AVX-NEXT:    vmovdqa %xmm1, %xmm0 # encoding: [0xc5,0xf9,0x6f,0xc1]
     88 ; X64-AVX-NEXT:    retq # encoding: [0xc3]
     89   %1 = load <2 x i64>, <2 x i64> *%p
     90   %2 = bitcast <2 x i64> %y to <16 x i8>
     91   %3 = tail call <2 x i64> @llvm.x86.sse4a.extrq(<2 x i64> %1, <16 x i8> %2) nounwind
     92   ret <2 x i64> %3
     93 }
     94 
     95 declare <2 x i64> @llvm.x86.sse4a.extrq(<2 x i64>, <16 x i8>) nounwind
     96 
     97 define <2 x i64> @test_insertqi(<2 x i64> %x, <2 x i64> %y) nounwind uwtable ssp {
     98 ; CHECK-LABEL: test_insertqi:
     99 ; CHECK:       # %bb.0:
    100 ; CHECK-NEXT:    insertq $6, $5, %xmm1, %xmm0 # encoding: [0xf2,0x0f,0x78,0xc1,0x05,0x06]
    101 ; CHECK-NEXT:    ret{{[l|q]}} # encoding: [0xc3]
    102   %1 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %x, <2 x i64> %y, i8 5, i8 6)
    103   ret <2 x i64> %1
    104 }
    105 
    106 define <2 x i64> @test_insertqi_domain(<2 x i64> *%p, <2 x i64> %y) nounwind uwtable ssp {
    107 ; X86-SSE-LABEL: test_insertqi_domain:
    108 ; X86-SSE:       # %bb.0:
    109 ; X86-SSE-NEXT:    movl {{[0-9]+}}(%esp), %eax # encoding: [0x8b,0x44,0x24,0x04]
    110 ; X86-SSE-NEXT:    movdqa (%eax), %xmm1 # encoding: [0x66,0x0f,0x6f,0x08]
    111 ; X86-SSE-NEXT:    insertq $6, $5, %xmm0, %xmm1 # encoding: [0xf2,0x0f,0x78,0xc8,0x05,0x06]
    112 ; X86-SSE-NEXT:    movdqa %xmm1, %xmm0 # encoding: [0x66,0x0f,0x6f,0xc1]
    113 ; X86-SSE-NEXT:    retl # encoding: [0xc3]
    114 ;
    115 ; X86-AVX-LABEL: test_insertqi_domain:
    116 ; X86-AVX:       # %bb.0:
    117 ; X86-AVX-NEXT:    movl {{[0-9]+}}(%esp), %eax # encoding: [0x8b,0x44,0x24,0x04]
    118 ; X86-AVX-NEXT:    vmovdqa (%eax), %xmm1 # encoding: [0xc5,0xf9,0x6f,0x08]
    119 ; X86-AVX-NEXT:    insertq $6, $5, %xmm0, %xmm1 # encoding: [0xf2,0x0f,0x78,0xc8,0x05,0x06]
    120 ; X86-AVX-NEXT:    vmovdqa %xmm1, %xmm0 # encoding: [0xc5,0xf9,0x6f,0xc1]
    121 ; X86-AVX-NEXT:    retl # encoding: [0xc3]
    122 ;
    123 ; X64-SSE-LABEL: test_insertqi_domain:
    124 ; X64-SSE:       # %bb.0:
    125 ; X64-SSE-NEXT:    movdqa (%rdi), %xmm1 # encoding: [0x66,0x0f,0x6f,0x0f]
    126 ; X64-SSE-NEXT:    insertq $6, $5, %xmm0, %xmm1 # encoding: [0xf2,0x0f,0x78,0xc8,0x05,0x06]
    127 ; X64-SSE-NEXT:    movdqa %xmm1, %xmm0 # encoding: [0x66,0x0f,0x6f,0xc1]
    128 ; X64-SSE-NEXT:    retq # encoding: [0xc3]
    129 ;
    130 ; X64-AVX-LABEL: test_insertqi_domain:
    131 ; X64-AVX:       # %bb.0:
    132 ; X64-AVX-NEXT:    vmovdqa (%rdi), %xmm1 # encoding: [0xc5,0xf9,0x6f,0x0f]
    133 ; X64-AVX-NEXT:    insertq $6, $5, %xmm0, %xmm1 # encoding: [0xf2,0x0f,0x78,0xc8,0x05,0x06]
    134 ; X64-AVX-NEXT:    vmovdqa %xmm1, %xmm0 # encoding: [0xc5,0xf9,0x6f,0xc1]
    135 ; X64-AVX-NEXT:    retq # encoding: [0xc3]
    136   %1 = load <2 x i64>, <2 x i64> *%p
    137   %2 = tail call <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64> %1, <2 x i64> %y, i8 5, i8 6)
    138   ret <2 x i64> %2
    139 }
    140 
    141 declare <2 x i64> @llvm.x86.sse4a.insertqi(<2 x i64>, <2 x i64>, i8, i8) nounwind
    142 
    143 define <2 x i64> @test_insertq(<2 x i64> %x, <2 x i64> %y) nounwind uwtable ssp {
    144 ; CHECK-LABEL: test_insertq:
    145 ; CHECK:       # %bb.0:
    146 ; CHECK-NEXT:    insertq %xmm1, %xmm0 # encoding: [0xf2,0x0f,0x79,0xc1]
    147 ; CHECK-NEXT:    ret{{[l|q]}} # encoding: [0xc3]
    148   %1 = tail call <2 x i64> @llvm.x86.sse4a.insertq(<2 x i64> %x, <2 x i64> %y) nounwind
    149   ret <2 x i64> %1
    150 }
    151 
    152 define <2 x i64> @test_insertq_domain(<2 x i64> *%p, <2 x i64> %y) nounwind uwtable ssp {
    153 ; X86-SSE-LABEL: test_insertq_domain:
    154 ; X86-SSE:       # %bb.0:
    155 ; X86-SSE-NEXT:    movl {{[0-9]+}}(%esp), %eax # encoding: [0x8b,0x44,0x24,0x04]
    156 ; X86-SSE-NEXT:    movdqa (%eax), %xmm1 # encoding: [0x66,0x0f,0x6f,0x08]
    157 ; X86-SSE-NEXT:    insertq %xmm0, %xmm1 # encoding: [0xf2,0x0f,0x79,0xc8]
    158 ; X86-SSE-NEXT:    movdqa %xmm1, %xmm0 # encoding: [0x66,0x0f,0x6f,0xc1]
    159 ; X86-SSE-NEXT:    retl # encoding: [0xc3]
    160 ;
    161 ; X86-AVX-LABEL: test_insertq_domain:
    162 ; X86-AVX:       # %bb.0:
    163 ; X86-AVX-NEXT:    movl {{[0-9]+}}(%esp), %eax # encoding: [0x8b,0x44,0x24,0x04]
    164 ; X86-AVX-NEXT:    vmovdqa (%eax), %xmm1 # encoding: [0xc5,0xf9,0x6f,0x08]
    165 ; X86-AVX-NEXT:    insertq %xmm0, %xmm1 # encoding: [0xf2,0x0f,0x79,0xc8]
    166 ; X86-AVX-NEXT:    vmovdqa %xmm1, %xmm0 # encoding: [0xc5,0xf9,0x6f,0xc1]
    167 ; X86-AVX-NEXT:    retl # encoding: [0xc3]
    168 ;
    169 ; X64-SSE-LABEL: test_insertq_domain:
    170 ; X64-SSE:       # %bb.0:
    171 ; X64-SSE-NEXT:    movdqa (%rdi), %xmm1 # encoding: [0x66,0x0f,0x6f,0x0f]
    172 ; X64-SSE-NEXT:    insertq %xmm0, %xmm1 # encoding: [0xf2,0x0f,0x79,0xc8]
    173 ; X64-SSE-NEXT:    movdqa %xmm1, %xmm0 # encoding: [0x66,0x0f,0x6f,0xc1]
    174 ; X64-SSE-NEXT:    retq # encoding: [0xc3]
    175 ;
    176 ; X64-AVX-LABEL: test_insertq_domain:
    177 ; X64-AVX:       # %bb.0:
    178 ; X64-AVX-NEXT:    vmovdqa (%rdi), %xmm1 # encoding: [0xc5,0xf9,0x6f,0x0f]
    179 ; X64-AVX-NEXT:    insertq %xmm0, %xmm1 # encoding: [0xf2,0x0f,0x79,0xc8]
    180 ; X64-AVX-NEXT:    vmovdqa %xmm1, %xmm0 # encoding: [0xc5,0xf9,0x6f,0xc1]
    181 ; X64-AVX-NEXT:    retq # encoding: [0xc3]
    182   %1 = load <2 x i64>, <2 x i64> *%p
    183   %2 = tail call <2 x i64> @llvm.x86.sse4a.insertq(<2 x i64> %1, <2 x i64> %y) nounwind
    184   ret <2 x i64> %2
    185 }
    186 
    187 declare <2 x i64> @llvm.x86.sse4a.insertq(<2 x i64>, <2 x i64>) nounwind
    188