Home | History | Annotate | Download | only in X86
      1 ; NOTE: Assertions have been autogenerated by update_llc_test_checks.py
      2 ; RUN: llc < %s -mtriple=x86_64-unknown-linux-gn -mcpu=knl | FileCheck %s --check-prefix=ALL --check-prefix=KNL
      3 ; RUN: llc < %s -mtriple=x86_64-unknown-linux-gn -mcpu=skx | FileCheck %s --check-prefix=ALL --check-prefix=SKX
      4 
      5 
      6 define void @any_extend_load_v8i64(<8 x i8> * %ptr) {
      7 ; ALL-LABEL: any_extend_load_v8i64:
      8 ; ALL:       # BB#0:
      9 ; ALL-NEXT:    vpmovzxbq (%rdi), %zmm0
     10 ; ALL-NEXT:    vpaddq {{.*}}(%rip){1to8}, %zmm0, %zmm0
     11 ; ALL-NEXT:    vpmovqb %zmm0, (%rdi)
     12 ; ALL-NEXT:    retq
     13   %wide.load = load <8 x i8>, <8 x i8>* %ptr, align 1
     14   %1 = zext <8 x i8> %wide.load to <8 x i64>
     15   %2 = add nuw nsw <8 x i64> %1, <i64 4, i64 4, i64 4, i64 4, i64 4, i64 4, i64 4, i64 4>
     16   %3 = xor <8 x i64> %2, zeroinitializer
     17   %4 = trunc <8 x i64> %3 to <8 x i8>
     18   store <8 x i8> %4, <8 x i8>* %ptr, align 1
     19   ret void
     20 }
     21 
     22 define void @any_extend_load_v8i32(<8 x i8> * %ptr) {
     23 ; KNL-LABEL: any_extend_load_v8i32:
     24 ; KNL:       # BB#0:
     25 ; KNL-NEXT:    vpmovzxbd {{.*#+}} ymm0 = mem[0],zero,zero,zero,mem[1],zero,zero,zero,mem[2],zero,zero,zero,mem[3],zero,zero,zero,mem[4],zero,zero,zero,mem[5],zero,zero,zero,mem[6],zero,zero,zero,mem[7],zero,zero,zero
     26 ; KNL-NEXT:    vpbroadcastd {{.*}}(%rip), %ymm1
     27 ; KNL-NEXT:    vpaddd %ymm1, %ymm0, %ymm0
     28 ; KNL-NEXT:    vpmovdw %zmm0, %ymm0
     29 ; KNL-NEXT:    vpshufb {{.*#+}} xmm0 = xmm0[0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u]
     30 ; KNL-NEXT:    vmovq %xmm0, (%rdi)
     31 ; KNL-NEXT:    retq
     32 ;
     33 ; SKX-LABEL: any_extend_load_v8i32:
     34 ; SKX:       # BB#0:
     35 ; SKX-NEXT:    vpmovzxbd (%rdi), %ymm0
     36 ; SKX-NEXT:    vpaddd {{.*}}(%rip){1to8}, %ymm0, %ymm0
     37 ; SKX-NEXT:    vpmovdb %ymm0, (%rdi)
     38 ; SKX-NEXT:    retq
     39   %wide.load = load <8 x i8>, <8 x i8>* %ptr, align 1
     40   %1 = zext <8 x i8> %wide.load to <8 x i32>
     41   %2 = add nuw nsw <8 x i32> %1, <i32 4, i32 4, i32 4, i32 4, i32 4, i32 4, i32 4, i32 4>
     42   %3 = xor <8 x i32> %2, zeroinitializer
     43   %4 = trunc <8 x i32> %3 to <8 x i8>
     44   store <8 x i8> %4, <8 x i8>* %ptr, align 1
     45   ret void
     46 }
     47 
     48 define void @any_extend_load_v8i16(<8 x i8> * %ptr) {
     49 ; KNL-LABEL: any_extend_load_v8i16:
     50 ; KNL:       # BB#0:
     51 ; KNL-NEXT:    vpmovzxbw {{.*#+}} xmm0 = mem[0],zero,mem[1],zero,mem[2],zero,mem[3],zero,mem[4],zero,mem[5],zero,mem[6],zero,mem[7],zero
     52 ; KNL-NEXT:    vpaddb {{.*}}(%rip), %xmm0, %xmm0
     53 ; KNL-NEXT:    vpshufb {{.*#+}} xmm0 = xmm0[0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u]
     54 ; KNL-NEXT:    vmovq %xmm0, (%rdi)
     55 ; KNL-NEXT:    retq
     56 ;
     57 ; SKX-LABEL: any_extend_load_v8i16:
     58 ; SKX:       # BB#0:
     59 ; SKX-NEXT:    vpmovzxbw (%rdi), %xmm0
     60 ; SKX-NEXT:    vpaddw {{.*}}(%rip), %xmm0, %xmm0
     61 ; SKX-NEXT:    vpmovwb %xmm0, (%rdi)
     62 ; SKX-NEXT:    retq
     63   %wide.load = load <8 x i8>, <8 x i8>* %ptr, align 1
     64   %1 = zext <8 x i8> %wide.load to <8 x i16>
     65   %2 = add nuw nsw <8 x i16> %1, <i16 4, i16 4, i16 4, i16 4, i16 4, i16 4, i16 4, i16 4>
     66   %3 = xor <8 x i16> %2, zeroinitializer
     67   %4 = trunc <8 x i16> %3 to <8 x i8>
     68   store <8 x i8> %4, <8 x i8>* %ptr, align 1
     69   ret void
     70 }
     71