1 ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py 2 ; RUN: llc < %s -mtriple=i686-unknown-unknown -mattr=+avx2 | FileCheck %s --check-prefix=X32 3 ; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx2 | FileCheck %s --check-prefix=X64 4 5 declare <4 x float> @llvm.x86.avx2.gather.d.ps(<4 x float>, i8*, 6 <4 x i32>, <4 x float>, i8) nounwind readonly 7 8 define <4 x float> @test_x86_avx2_gather_d_ps(i8* %a1, <4 x i32> %idx, <4 x float> %mask) { 9 ; X32-LABEL: test_x86_avx2_gather_d_ps: 10 ; X32: # %bb.0: 11 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax 12 ; X32-NEXT: vxorps %xmm2, %xmm2, %xmm2 13 ; X32-NEXT: vgatherdps %xmm1, (%eax,%xmm0,2), %xmm2 14 ; X32-NEXT: vmovaps %xmm2, %xmm0 15 ; X32-NEXT: retl 16 ; 17 ; X64-LABEL: test_x86_avx2_gather_d_ps: 18 ; X64: # %bb.0: 19 ; X64-NEXT: vxorps %xmm2, %xmm2, %xmm2 20 ; X64-NEXT: vgatherdps %xmm1, (%rdi,%xmm0,2), %xmm2 21 ; X64-NEXT: vmovaps %xmm2, %xmm0 22 ; X64-NEXT: retq 23 %res = call <4 x float> @llvm.x86.avx2.gather.d.ps(<4 x float> undef, 24 i8* %a1, <4 x i32> %idx, <4 x float> %mask, i8 2) ; 25 ret <4 x float> %res 26 } 27 28 declare <2 x double> @llvm.x86.avx2.gather.d.pd(<2 x double>, i8*, 29 <4 x i32>, <2 x double>, i8) nounwind readonly 30 31 define <2 x double> @test_x86_avx2_gather_d_pd(i8* %a1, <4 x i32> %idx, <2 x double> %mask) { 32 ; X32-LABEL: test_x86_avx2_gather_d_pd: 33 ; X32: # %bb.0: 34 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax 35 ; X32-NEXT: vxorpd %xmm2, %xmm2, %xmm2 36 ; X32-NEXT: vgatherdpd %xmm1, (%eax,%xmm0,2), %xmm2 37 ; X32-NEXT: vmovapd %xmm2, %xmm0 38 ; X32-NEXT: retl 39 ; 40 ; X64-LABEL: test_x86_avx2_gather_d_pd: 41 ; X64: # %bb.0: 42 ; X64-NEXT: vxorpd %xmm2, %xmm2, %xmm2 43 ; X64-NEXT: vgatherdpd %xmm1, (%rdi,%xmm0,2), %xmm2 44 ; X64-NEXT: vmovapd %xmm2, %xmm0 45 ; X64-NEXT: retq 46 %res = call <2 x double> @llvm.x86.avx2.gather.d.pd(<2 x double> undef, 47 i8* %a1, <4 x i32> %idx, <2 x double> %mask, i8 2) ; 48 ret <2 x double> %res 49 } 50 51 declare <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float>, i8*, 52 <8 x i32>, <8 x float>, i8) nounwind readonly 53 54 define <8 x float> @test_x86_avx2_gather_d_ps_256(i8* %a1, <8 x i32> %idx, <8 x float> %mask) { 55 ; X32-LABEL: test_x86_avx2_gather_d_ps_256: 56 ; X32: # %bb.0: 57 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax 58 ; X32-NEXT: vxorps %xmm2, %xmm2, %xmm2 59 ; X32-NEXT: vgatherdps %ymm1, (%eax,%ymm0,4), %ymm2 60 ; X32-NEXT: vmovaps %ymm2, %ymm0 61 ; X32-NEXT: retl 62 ; 63 ; X64-LABEL: test_x86_avx2_gather_d_ps_256: 64 ; X64: # %bb.0: 65 ; X64-NEXT: vxorps %xmm2, %xmm2, %xmm2 66 ; X64-NEXT: vgatherdps %ymm1, (%rdi,%ymm0,4), %ymm2 67 ; X64-NEXT: vmovaps %ymm2, %ymm0 68 ; X64-NEXT: retq 69 %res = call <8 x float> @llvm.x86.avx2.gather.d.ps.256(<8 x float> undef, 70 i8* %a1, <8 x i32> %idx, <8 x float> %mask, i8 4) ; 71 ret <8 x float> %res 72 } 73 74 declare <4 x double> @llvm.x86.avx2.gather.d.pd.256(<4 x double>, i8*, 75 <4 x i32>, <4 x double>, i8) nounwind readonly 76 77 define <4 x double> @test_x86_avx2_gather_d_pd_256(i8* %a1, <4 x i32> %idx, <4 x double> %mask) { 78 ; X32-LABEL: test_x86_avx2_gather_d_pd_256: 79 ; X32: # %bb.0: 80 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax 81 ; X32-NEXT: vxorpd %xmm2, %xmm2, %xmm2 82 ; X32-NEXT: vgatherdpd %ymm1, (%eax,%xmm0,8), %ymm2 83 ; X32-NEXT: vmovapd %ymm2, %ymm0 84 ; X32-NEXT: retl 85 ; 86 ; X64-LABEL: test_x86_avx2_gather_d_pd_256: 87 ; X64: # %bb.0: 88 ; X64-NEXT: vxorpd %xmm2, %xmm2, %xmm2 89 ; X64-NEXT: vgatherdpd %ymm1, (%rdi,%xmm0,8), %ymm2 90 ; X64-NEXT: vmovapd %ymm2, %ymm0 91 ; X64-NEXT: retq 92 %res = call <4 x double> @llvm.x86.avx2.gather.d.pd.256(<4 x double> undef, 93 i8* %a1, <4 x i32> %idx, <4 x double> %mask, i8 8) ; 94 ret <4 x double> %res 95 } 96 97 define <2 x i64> @test_mm_i32gather_epi32(i32 *%a0, <2 x i64> %a1) { 98 ; X32-LABEL: test_mm_i32gather_epi32: 99 ; X32: # %bb.0: 100 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax 101 ; X32-NEXT: vpcmpeqd %xmm2, %xmm2, %xmm2 102 ; X32-NEXT: vpxor %xmm1, %xmm1, %xmm1 103 ; X32-NEXT: vpgatherdd %xmm2, (%eax,%xmm0,2), %xmm1 104 ; X32-NEXT: vmovdqa %xmm1, %xmm0 105 ; X32-NEXT: retl 106 ; 107 ; X64-LABEL: test_mm_i32gather_epi32: 108 ; X64: # %bb.0: 109 ; X64-NEXT: vpcmpeqd %xmm2, %xmm2, %xmm2 110 ; X64-NEXT: vpxor %xmm1, %xmm1, %xmm1 111 ; X64-NEXT: vpgatherdd %xmm2, (%rdi,%xmm0,2), %xmm1 112 ; X64-NEXT: vmovdqa %xmm1, %xmm0 113 ; X64-NEXT: retq 114 %arg0 = bitcast i32 *%a0 to i8* 115 %arg1 = bitcast <2 x i64> %a1 to <4 x i32> 116 %mask = bitcast <2 x i64> <i64 -1, i64 -1> to <4 x i32> 117 %call = call <4 x i32> @llvm.x86.avx2.gather.d.d(<4 x i32> zeroinitializer, i8* %arg0, <4 x i32> %arg1, <4 x i32> %mask, i8 2) 118 %bc = bitcast <4 x i32> %call to <2 x i64> 119 ret <2 x i64> %bc 120 } 121 declare <4 x i32> @llvm.x86.avx2.gather.d.d(<4 x i32>, i8*, <4 x i32>, <4 x i32>, i8) nounwind readonly 122 123 define <2 x double> @test_mm_i32gather_pd(double *%a0, <2 x i64> %a1) { 124 ; X32-LABEL: test_mm_i32gather_pd: 125 ; X32: # %bb.0: 126 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax 127 ; X32-NEXT: vpcmpeqd %xmm2, %xmm2, %xmm2 128 ; X32-NEXT: vxorpd %xmm1, %xmm1, %xmm1 129 ; X32-NEXT: vgatherdpd %xmm2, (%eax,%xmm0,2), %xmm1 130 ; X32-NEXT: vmovapd %xmm1, %xmm0 131 ; X32-NEXT: retl 132 ; 133 ; X64-LABEL: test_mm_i32gather_pd: 134 ; X64: # %bb.0: 135 ; X64-NEXT: vpcmpeqd %xmm2, %xmm2, %xmm2 136 ; X64-NEXT: vxorpd %xmm1, %xmm1, %xmm1 137 ; X64-NEXT: vgatherdpd %xmm2, (%rdi,%xmm0,2), %xmm1 138 ; X64-NEXT: vmovapd %xmm1, %xmm0 139 ; X64-NEXT: retq 140 %arg0 = bitcast double *%a0 to i8* 141 %arg1 = bitcast <2 x i64> %a1 to <4 x i32> 142 %cmp = fcmp oeq <2 x double> zeroinitializer, zeroinitializer 143 %sext = sext <2 x i1> %cmp to <2 x i64> 144 %mask = bitcast <2 x i64> %sext to <2 x double> 145 %res = call <2 x double> @llvm.x86.avx2.gather.d.pd(<2 x double> zeroinitializer, i8* %arg0, <4 x i32> %arg1, <2 x double> %mask, i8 2) 146 ret <2 x double> %res 147 } 148