Lines Matching full:next
13 ; SSE-NEXT: pmullw {{.*}}(%rip), %xmm0
14 ; SSE-NEXT: retq
18 ; AVX2-NEXT: vpmullw {{.*}}(%rip), %xmm0, %xmm0
19 ; AVX2-NEXT: retq
23 ; AVX512-NEXT: vpmullw {{.*}}(%rip), %xmm0, %xmm0
24 ; AVX512-NEXT: retq
32 ; SSE-NEXT: pmullw {{.*}}(%rip), %xmm0
33 ; SSE-NEXT: retq
37 ; AVX2-NEXT: vpmullw {{.*}}(%rip), %xmm0, %xmm0
38 ; AVX2-NEXT: retq
42 ; AVX512-NEXT: vpmullw {{.*}}(%rip), %xmm0, %xmm0
43 ; AVX512-NEXT: retq
55 ; SSE-NEXT: pmulld {{.*}}(%rip), %xmm0
56 ; SSE-NEXT: retq
60 ; AVX2-NEXT: vpsllvd {{.*}}(%rip), %xmm0, %xmm0
61 ; AVX2-NEXT: retq
65 ; AVX512-NEXT: vpsllvd {{.*}}(%rip), %xmm0, %xmm0
66 ; AVX512-NEXT: retq
74 ; SSE-NEXT: pmulld {{.*}}(%rip), %xmm0
75 ; SSE-NEXT: retq
79 ; AVX2-NEXT: vpsllvd {{.*}}(%rip), %xmm0, %xmm0
80 ; AVX2-NEXT: retq
84 ; AVX512-NEXT: vpsllvd {{.*}}(%rip), %xmm0, %xmm0
85 ; AVX512-NEXT: retq
97 ; SSE-NEXT: movdqa {{.*#+}} xmm2 = [2,2,4,8,128,1,512,2048]
98 ; SSE-NEXT: pmullw %xmm2, %xmm0
99 ; SSE-NEXT: pmullw %xmm2, %xmm1
100 ; SSE-NEXT: retq
104 ; AVX2-NEXT: vpmullw {{.*}}(%rip), %ymm0, %ymm0
105 ; AVX2-NEXT: retq
109 ; AVX512-NEXT: vpmullw {{.*}}(%rip), %ymm0, %ymm0
110 ; AVX512-NEXT: retq
122 ; SSE-NEXT: movdqa {{.*#+}} xmm2 = [2,2,4,8]
123 ; SSE-NEXT: pmulld %xmm2, %xmm0
124 ; SSE-NEXT: pmulld %xmm2, %xmm1
125 ; SSE-NEXT: retq
129 ; AVX2-NEXT: vpsllvd {{.*}}(%rip), %ymm0, %ymm0
130 ; AVX2-NEXT: retq
134 ; AVX512-NEXT: vpsllvd {{.*}}(%rip), %ymm0, %ymm0
135 ; AVX512-NEXT: retq
147 ; SSE-NEXT: movdqa {{.*#+}} xmm4 = [2,2,4,8,128,1,512,2048]
148 ; SSE-NEXT: pmullw %xmm4, %xmm0
149 ; SSE-NEXT: pmullw %xmm4, %xmm1
150 ; SSE-NEXT: pmullw %xmm4, %xmm2
151 ; SSE-NEXT: pmullw %xmm4, %xmm3
152 ; SSE-NEXT: retq
156 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = [2,2,4,8,128,1,512,2048,2,2,4,8,128,1,512,2048]
157 ; AVX2-NEXT: vpmullw %ymm2, %ymm0, %ymm0
158 ; AVX2-NEXT: vpmullw %ymm2, %ymm1, %ymm1
159 ; AVX2-NEXT: retq
163 ; AVX512-NEXT: vmovdqa {{.*#+}} ymm2 = [2,2,4,8,128,1,512,2048,2,2,4,8,128,1,512,2048]
164 ; AVX512-NEXT: vpmullw %ymm2, %ymm0, %ymm0
165 ; AVX512-NEXT: vpmullw %ymm2, %ymm1, %ymm1
166 ; AVX512-NEXT: retq
177 ; SSE-NEXT: movdqa {{.*#+}} xmm4 = [2,2,4,8]
178 ; SSE-NEXT: pmulld %xmm4, %xmm0
179 ; SSE-NEXT: pmulld %xmm4, %xmm1
180 ; SSE-NEXT: pmulld %xmm4, %xmm2
181 ; SSE-NEXT: pmulld %xmm4, %xmm3
182 ; SSE-NEXT: retq
186 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = [1,1,2,3,1,1,2,3]
187 ; AVX2-NEXT: vpsllvd %ymm2, %ymm0, %ymm0
188 ; AVX2-NEXT: vpsllvd %ymm2, %ymm1, %ymm1
189 ; AVX2-NEXT: retq
193 ; AVX512-NEXT: vpsllvd {{.*}}(%rip), %zmm0, %zmm0
194 ; AVX512-NEXT: retq
204 ; SSE-NEXT: movdqa %xmm1, %xmm4
205 ; SSE-NEXT: psllq $3, %xmm4
206 ; SSE-NEXT: psllq $2, %xmm1
207 ; SSE-NEXT: pblendw {{.*#+}} xmm1 = xmm1[0,1,2,3],xmm4[4,5,6,7]
208 ; SSE-NEXT: movdqa %xmm3, %xmm4
209 ; SSE-NEXT: psllq $3, %xmm4
210 ; SSE-NEXT: psllq $2, %xmm3
211 ; SSE-NEXT: pblendw {{.*#+}} xmm3 = xmm3[0,1,2,3],xmm4[4,5,6,7]
212 ; SSE-NEXT: paddq %xmm0, %xmm0
213 ; SSE-NEXT: paddq %xmm2, %xmm2
214 ; SSE-NEXT: retq
218 ; AVX2-NEXT: vmovdqa {{.*#+}} ymm2 = [1,1,2,3]
219 ; AVX2-NEXT: vpsllvq %ymm2, %ymm0, %ymm0
220 ; AVX2-NEXT: vpsllvq %ymm2, %ymm1, %ymm1
221 ; AVX2-NEXT: retq
225 ; AVX512-NEXT: vpsllvq {{.*}}(%rip), %zmm0, %zmm0
226 ; AVX512-NEXT: retq