Home | History | Annotate | Download | only in X86

Lines Matching full:next

11 ; SSE-NEXT:    addsubps %xmm1, %xmm0
12 ; SSE-NEXT: retq
16 ; AVX-NEXT: vaddsubps %xmm1, %xmm0, %xmm0
17 ; AVX-NEXT: retq
40 ; SSE-NEXT: addsubps %xmm1, %xmm0
41 ; SSE-NEXT: retq
45 ; AVX-NEXT: vaddsubps %xmm1, %xmm0, %xmm0
46 ; AVX-NEXT: retq
61 ; SSE-NEXT: addsubps %xmm1, %xmm0
62 ; SSE-NEXT: retq
66 ; AVX-NEXT: vaddsubps %xmm1, %xmm0, %xmm0
67 ; AVX-NEXT: retq
82 ; SSE-NEXT: addsubps %xmm1, %xmm0
83 ; SSE-NEXT: retq
87 ; AVX-NEXT: vaddsubps %xmm1, %xmm0, %xmm0
88 ; AVX-NEXT: retq
103 ; SSE-NEXT: addsubps %xmm1, %xmm0
104 ; SSE-NEXT: retq
108 ; AVX-NEXT: vaddsubps %xmm1, %xmm0, %xmm0
109 ; AVX-NEXT: retq
124 ; SSE-NEXT: addsubps %xmm1, %xmm0
125 ; SSE-NEXT: retq
129 ; AVX-NEXT: vaddsubps %xmm1, %xmm0, %xmm0
130 ; AVX-NEXT: retq
153 ; SSE-NEXT: addsubpd %xmm2, %xmm0
154 ; SSE-NEXT: addsubpd %xmm3, %xmm1
155 ; SSE-NEXT: retq
159 ; AVX-NEXT: vaddsubpd %ymm1, %ymm0, %ymm0
160 ; AVX-NEXT: retq
183 ; SSE-NEXT: addsubpd %xmm1, %xmm0
184 ; SSE-NEXT: retq
188 ; AVX-NEXT: vaddsubpd %xmm1, %xmm0, %xmm0
189 ; AVX-NEXT: retq
204 ; SSE-NEXT: addsubps %xmm2, %xmm0
205 ; SSE-NEXT: addsubps %xmm3, %xmm1
206 ; SSE-NEXT: retq
210 ; AVX-NEXT: vaddsubps %ymm1, %ymm0, %ymm0
211 ; AVX-NEXT: retq
253 ; SSE-NEXT: subss %xmm1, %xmm0
254 ; SSE-NEXT: retq
258 ; AVX-NEXT: vsubss %xmm1, %xmm0, %xmm0
259 ; AVX-NEXT: retq
270 ; SSE-NEXT: shufpd {{.*#+}} xmm0 = xmm0[1,0]
271 ; SSE-NEXT: shufpd {{.*#+}} xmm1 = xmm1[1,0]
272 ; SSE-NEXT: subss %xmm1, %xmm0
273 ; SSE-NEXT: movddup {{.*#+}} xmm0 = xmm0[0,0]
274 ; SSE-NEXT: retq
278 ; AVX-NEXT: vpermilpd {{.*#+}} xmm0 = xmm0[1,0]
279 ; AVX-NEXT: vpermilpd {{.*#+}} xmm1 = xmm1[1,0]
280 ; AVX-NEXT: vsubss %xmm1, %xmm0, %xmm0
281 ; AVX-NEXT: vmovddup {{.*#+}} xmm0 = xmm0[0,0]
282 ; AVX-NEXT: retq
293 ; SSE-NEXT: movshdup {{.*#+}} xmm0 = xmm0[1,1,3,3]
294 ; SSE-NEXT: movshdup {{.*#+}} xmm1 = xmm1[1,1,3,3]
295 ; SSE-NEXT: addss %xmm0, %xmm1
296 ; SSE-NEXT: movsldup {{.*#+}} xmm0 = xmm1[0,0,2,2]
297 ; SSE-NEXT: retq
301 ; AVX-NEXT: vmovshdup {{.*#+}} xmm0 = xmm0[1,1,3,3]
302 ; AVX-NEXT: vmovshdup {{.*#+}} xmm1 = xmm1[1,1,3,3]
303 ; AVX-NEXT: vaddss %xmm1, %xmm0, %xmm0
304 ; AVX-NEXT: vmovsldup {{.*#+}} xmm0 = xmm0[0,0,2,2]
305 ; AVX-NEXT: retq
316 ; SSE-NEXT: shufps {{.*#+}} xmm0 = xmm0[3,1,2,3]
317 ; SSE-NEXT: shufps {{.*#+}} xmm1 = xmm1[3,1,2,3]
318 ; SSE-NEXT: addss %xmm0, %xmm1
319 ; SSE-NEXT: shufps {{.*#+}} xmm1 = xmm1[0,1,2,0]
320 ; SSE-NEXT: movaps %xmm1, %xmm0
321 ; SSE-NEXT: retq
325 ; AVX-NEXT: vpermilps {{.*#+}} xmm0 = xmm0[3,1,2,3]
326 ; AVX-NEXT: vpermilps {{.*#+}} xmm1 = xmm1[3,1,2,3]
327 ; AVX-NEXT: vaddss %xmm1, %xmm0, %xmm0
328 ; AVX-NEXT: vpermilps {{.*#+}} xmm0 = xmm0[0,1,2,0]
329 ; AVX-NEXT: retq
340 ; SSE-NEXT: movaps %xmm0, %xmm2
341 ; SSE-NEXT: subss %xmm1, %xmm2
342 ; SSE-NEXT: shufpd {{.*#+}} xmm0 = xmm0[1,0]
343 ; SSE-NEXT: shufpd {{.*#+}} xmm1 = xmm1[1,0]
344 ; SSE-NEXT: subss %xmm1, %xmm0
345 ; SSE-NEXT: unpcklps {{.*#+}} xmm2 = xmm2[0],xmm0[0],xmm2[1],xmm0[1]
346 ; SSE-NEXT: shufps {{.*#+}} xmm2 = xmm2[0,1,1,3]
347 ; SSE-NEXT: movaps %xmm2, %xmm0
348 ; SSE-NEXT: retq
352 ; AVX-NEXT: vsubss %xmm1, %xmm0, %xmm2
353 ; AVX-NEXT: vpermilpd {{.*#+}} xmm0 = xmm0[1,0]
354 ; AVX-NEXT: vpermilpd {{.*#+}} xmm1 = xmm1[1,0]
355 ; AVX-NEXT: vsubss %xmm1, %xmm0, %xmm0
356 ; AVX-NEXT: vinsertps {{.*#+}} xmm0 = xmm2[0,1],xmm0[0],xmm2[3]
357 ; AVX-NEXT: retq
372 ; SSE-NEXT: movshdup {{.*#+}} xmm3 = xmm0[1,1,3,3]
373 ; SSE-NEXT: movshdup {{.*#+}} xmm2 = xmm1[1,1,3,3]
374 ; SSE-NEXT: addss %xmm3, %xmm2
375 ; SSE-NEXT: shufps {{.*#+}} xmm0 = xmm0[3,1,2,3]
376 ; SSE-NEXT: shufps {{.*#+}} xmm1 = xmm1[3,1,2,3]
377 ; SSE-NEXT: addss %xmm0, %xmm1
378 ; SSE-NEXT: unpcklps {{.*#+}} xmm2 = xmm2[0],xmm1[0],xmm2[1],xmm1[1]
379 ; SSE-NEXT: shufps {{.*#+}} xmm2 = xmm2[0,0,2,1]
380 ; SSE-NEXT: movaps %xmm2, %xmm0
381 ; SSE-NEXT: retq
385 ; AVX-NEXT: vmovshdup {{.*#+}} xmm2 = xmm0[1,1,3,3]
386 ; AVX-NEXT: vmovshdup {{.*#+}} xmm3 = xmm1[1,1,3,3]
387 ; AVX-NEXT: vaddss %xmm3, %xmm2, %xmm2
388 ; AVX-NEXT: vpermilps {{.*#+}} xmm0 = xmm0[3,1,2,3]
389 ; AVX-NEXT: vpermilps {{.*#+}} xmm1 = xmm1[3,1,2,3]
390 ; AVX-NEXT: vaddss %xmm1, %xmm0, %xmm0
391 ; AVX-NEXT: vmovsldup {{.*#+}} xmm1 = xmm2[0,0,2,2]
392 ; AVX-NEXT: vinsertps {{.*#+}} xmm0 = xmm1[0,1,2],xmm0[0]
393 ; AVX-NEXT: retq
408 ; SSE-NEXT: movaps %xmm0, %xmm2
409 ; SSE-NEXT: subss %xmm0, %xmm2
410 ; SSE-NEXT: movaps %xmm0, %xmm3
411 ; SSE-NEXT: shufpd {{.*#+}} xmm3 = xmm3[1,0]
412 ; SSE-NEXT: movapd %xmm1, %xmm4
413 ; SSE-NEXT: shufpd {{.*#+}} xmm4 = xmm4[1,0]
414 ; SSE-NEXT: subss %xmm4, %xmm3
415 ; SSE-NEXT: movshdup {{.*#+}} xmm4 = xmm0[1,1,3,3]
416 ; SSE-NEXT: addss %xmm0, %xmm4
417 ; SSE-NEXT: shufps {{.*#+}} xmm0 = xmm0[3,1,2,3]
418 ; SSE-NEXT: shufps {{.*#+}} xmm1 = xmm1[3,1,2,3]
419 ; SSE-NEXT: addss %xmm0, %xmm1
420 ; SSE-NEXT: unpcklps {{.*#+}} xmm4 = xmm4[0],xmm1[0],xmm4[1],xmm1[1]
421 ; SSE-NEXT: unpcklps {{.*#+}} xmm2 = xmm2[0],xmm3[0],xmm2[1],xmm3[1]
422 ; SSE-NEXT: unpcklps {{.*#+}} xmm2 = xmm2[0],xmm4[0],xmm2[1],xmm4[1]
423 ; SSE-NEXT: movaps %xmm2, %xmm0
424 ; SSE-NEXT: retq
428 ; AVX-NEXT: vsubss %xmm0, %xmm0, %xmm2
429 ; AVX-NEXT: vpermilpd {{.*#+}} xmm3 = xmm0[1,0]
430 ; AVX-NEXT: vpermilpd {{.*#+}} xmm4 = xmm1[1,0]
431 ; AVX-NEXT: vsubss %xmm4, %xmm3, %xmm3
432 ; AVX-NEXT: vmovshdup {{.*#+}} xmm4 = xmm0[1,1,3,3]
433 ; AVX-NEXT: vaddss %xmm0, %xmm4, %xmm4
434 ; AVX-NEXT: vpermilps {{.*#+}} xmm0 = xmm0[3,1,2,3]
435 ; AVX-NEXT: vpermilps {{.*#+}} xmm1 = xmm1[3,1,2,3]
436 ; AVX-NEXT: vaddss %xmm1, %xmm0, %xmm0
437 ; AVX-NEXT: vinsertps {{.*#+}} xmm1 = xmm2[0],xmm4[0],xmm2[2,3]
438 ; AVX-NEXT: vinsertps {{.*#+}} xmm1 = xmm1[0,1],xmm3[0],xmm1[3]
439 ; AVX-NEXT: vinsertps {{.*#+}} xmm0 = xmm1[0,1,2],xmm0[0]
440 ; AVX-NEXT: retq
463 ; SSE-NEXT: addsubps %xmm1, %xmm0
464 ; SSE-NEXT: retq
468 ; AVX-NEXT: vaddsubps %xmm1, %xmm0, %xmm0
469 ; AVX-NEXT: retq