Home | History | Annotate | Download | only in X86

Lines Matching full:next

9 ; X32-NEXT:    pinsrd $1, {{[0-9]+}}(%esp), %xmm0
10 ; X32-NEXT: retl
14 ; X64-NEXT: pinsrd $1, %edi, %xmm0
15 ; X64-NEXT: retq
23 ; X32-NEXT: pinsrb $1, {{[0-9]+}}(%esp), %xmm0
24 ; X32-NEXT: retl
28 ; X64-NEXT: pinsrb $1, %edi, %xmm0
29 ; X64-NEXT: retq
37 ; X32-NEXT: movl L_g16$non_lazy_ptr, %eax
38 ; X32-NEXT: pmovzxbq {{.*#+}} xmm0 = mem[0],zero,zero,zero,zero,zero,zero,zero,mem[1],zero,zero,zero,zero,zero,zero,zero
39 ; X32-NEXT: retl
43 ; X64-NEXT: movq _g16@{{.*}}(%rip), %rax
44 ; X64-NEXT: pmovzxbq {{.*#+}} xmm0 = mem[0],zero,zero,zero,zero,zero,zero,zero,mem[1],zero,zero,zero,zero,zero,zero,zero
45 ; X64-NEXT: retq
59 ; X32-NEXT: extractps $3, %xmm0, %eax
60 ; X32-NEXT: retl
64 ; X64-NEXT: extractps $3, %xmm0, %eax
65 ; X64-NEXT: retq
73 ; X32-NEXT: extractps $3, %xmm0, %eax
74 ; X32-NEXT: retl
78 ; X64-NEXT: extractps $3, %xmm0, %eax
79 ; X64-NEXT: retq
93 ; X32-NEXT: pushl %eax
94 ; X32-NEXT: shufps {{.*#+}} xmm0 = xmm0[3,1,2,3]
95 ; X32-NEXT: addss LCPI5_0, %xmm0
96 ; X32-NEXT: movss %xmm0, (%esp)
97 ; X32-NEXT: flds (%esp)
98 ; X32-NEXT: popl %eax
99 ; X32-NEXT: retl
103 ; X64-NEXT: shufps {{.*#+}} xmm0 = xmm0[3,1,2,3]
104 ; X64-NEXT: addss {{.*}}(%rip), %xmm0
105 ; X64-NEXT: retq
113 ; X32-NEXT: pushl %eax
114 ; X32-NEXT: shufps {{.*#+}} xmm0 = xmm0[3,1,2,3]
115 ; X32-NEXT: movss %xmm0, (%esp)
116 ; X32-NEXT: flds (%esp)
117 ; X32-NEXT: popl %eax
118 ; X32-NEXT: retl
122 ; X64-NEXT: shufps {{.*#+}} xmm0 = xmm0[3,1,2,3]
123 ; X64-NEXT: retq
130 ; X32-NEXT: pextrd $3, %xmm0, %eax
131 ; X32-NEXT: retl
135 ; X64-NEXT: pextrd $3, %xmm0, %eax
136 ; X64-NEXT: retq
144 ; X32-NEXT: insertps {{.*#+}} xmm0 = zero,xmm1[0],zero,xmm0[3]
145 ; X32-NEXT: retl
149 ; X64-NEXT: insertps {{.*#+}} xmm0 = zero,xmm1[0],zero,xmm0[3]
150 ; X64-NEXT: retq
162 ; X32-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
163 ; X32-NEXT: blendps {{.*#+}} xmm0 = xmm1[0],xmm0[1,2,3]
164 ; X32-NEXT: retl
168 ; X64-NEXT: blendps {{.*#+}} xmm0 = xmm1[0],xmm0[1,2,3]
169 ; X64-NEXT: retq
180 ; X32-NEXT: insertps {{.*#+}} xmm0 = mem[0],xmm0[1,2,3]
181 ; X32-NEXT: retl
185 ; X64-NEXT: blendps {{.*#+}} xmm0 = xmm1[0],xmm0[1,2,3]
186 ; X64-NEXT: retq
196 ; X32-NEXT: blendps {{.*#+}} xmm0 = xmm1[0],xmm0[1,2,3]
197 ; X32-NEXT: retl
201 ; X64-NEXT: blendps {{.*#+}} xmm0 = xmm1[0],xmm0[1,2,3]
202 ; X64-NEXT: retq
211 ; X32-NEXT: xorl %eax, %eax
212 ; X32-NEXT: ptest %xmm1, %xmm0
213 ; X32-NEXT: sete %al
214 ; X32-NEXT: retl
218 ; X64-NEXT: xorl %eax, %eax
219 ; X64-NEXT: ptest %xmm1, %xmm0
220 ; X64-NEXT: sete %al
221 ; X64-NEXT: retq
229 ; X32-NEXT: ptest %xmm1, %xmm0
230 ; X32-NEXT: sbbl %eax, %eax
231 ; X32-NEXT: andl $1, %eax
232 ; X32-NEXT: retl
236 ; X64-NEXT: ptest %xmm1, %xmm0
237 ; X64-NEXT: sbbl %eax, %eax
238 ; X64-NEXT: andl $1, %eax
239 ; X64-NEXT: retq
247 ; X32-NEXT: xorl %eax, %eax
248 ; X32-NEXT: ptest %xmm1, %xmm0
249 ; X32-NEXT: seta %al
250 ; X32-NEXT: retl
254 ; X64-NEXT: xorl %eax, %eax
255 ; X64-NEXT: ptest %xmm1, %xmm0
256 ; X64-NEXT: seta %al
257 ; X64-NEXT: retq
272 ; X32-NEXT: movshdup {{.*#+}} xmm2 = xmm0[1,1,3,3]
273 ; X32-NEXT: movshdup {{.*#+}} xmm3 = xmm1[1,1,3,3]
274 ; X32-NEXT: addss %xmm1, %xmm0
275 ; X32-NEXT: addss %xmm2, %xmm3
276 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],xmm3[0],xmm0[2,3]
277 ; X32-NEXT: retl
281 ; X64-NEXT: movshdup {{.*#+}} xmm2 = xmm0[1,1,3,3]
282 ; X64-NEXT: movshdup {{.*#+}} xmm3 = xmm1[1,1,3,3]
283 ; X64-NEXT: addss %xmm1, %xmm0
284 ; X64-NEXT: addss %xmm2, %xmm3
285 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],xmm3[0],xmm0[2,3]
286 ; X64-NEXT: retq
302 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
303 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1,2],mem[0]
304 ; X32-NEXT: retl
308 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1,2],mem[0]
309 ; X64-NEXT: retq
319 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1],xmm1[1],xmm0[3]
320 ; X32-NEXT: retl
324 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1],xmm1[1],xmm0[3]
325 ; X64-NEXT: retq
336 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
337 ; X32-NEXT: pshufd {{.*#+}} xmm1 = mem[0,1,2,0]
338 ; X32-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1,2,3,4,5],xmm1[6,7]
339 ; X32-NEXT: retl
343 ; X64-NEXT: pshufd {{.*#+}} xmm1 = mem[0,1,2,0]
344 ; X64-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1,2,3,4,5],xmm1[6,7]
345 ; X64-NEXT: retq
355 ; X32-NEXT: pshufd {{.*#+}} xmm1 = xmm1[2,3,0,1]
356 ; X32-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,3],xmm0[4,5,6,7]
357 ; X32-NEXT: retl
361 ; X64-NEXT: pshufd {{.*#+}} xmm1 = xmm1[2,3,0,1]
362 ; X64-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,3],xmm0[4,5,6,7]
363 ; X64-NEXT: retq
372 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
373 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],mem[0],xmm0[2,3]
374 ; X32-NEXT: retl
378 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],mem[0],xmm0[2,3]
379 ; X64-NEXT: retq
390 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
391 ; X32-NEXT: movd {{.*#+}} xmm1 = mem[0],zero,zero,zero
392 ; X32-NEXT: pshufd {{.*#+}} xmm1 = xmm1[0,1,0,1]
393 ; X32-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1,2,3],xmm1[4,5],xmm0[6,7]
394 ; X32-NEXT: retl
398 ; X64-NEXT: movd {{.*#+}} xmm1 = mem[0],zero,zero,zero
399 ; X64-NEXT: pshufd {{.*#+}} xmm1 = xmm1[0,1,0,1]
400 ; X64-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1,2,3],xmm1[4,5],xmm0[6,7]
401 ; X64-NEXT: retq
412 ; X32-NEXT: xorps %xmm1, %xmm1
413 ; X32-NEXT: blendps {{.*#+}} xmm0 = xmm0[0,1,2],xmm1[3]
414 ; X32-NEXT: retl
418 ; X64-NEXT: xorps %xmm1, %xmm1
419 ; X64-NEXT: blendps {{.*#+}} xmm0 = xmm0[0,1,2],xmm1[3]
420 ; X64-NEXT: retq
434 ; X32-NEXT: movq {{.*#+}} xmm0 = xmm0[0],zero
435 ; X32-NEXT: retl
439 ; X64-NEXT: movq {{.*#+}} xmm0 = xmm0[0],zero
440 ; X64-NEXT: retq
453 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1,1],zero
454 ; X32-NEXT: retl
458 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1,1],zero
459 ; X64-NEXT: retq
472 ; X32-NEXT
473 ; X32-NEXT: retl
477 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1,3],zero
478 ; X64-NEXT: retq
492 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[3],zero,zero,xmm0[3]
493 ; X32-NEXT: retl
497 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[3],zero,zero,xmm0[3]
498 ; X64-NEXT: retq
510 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],zero,zero,xmm1[0]
511 ; X32-NEXT: retl
515 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],zero,zero,xmm1[0]
516 ; X64-NEXT: retq
528 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],zero,zero,xmm0[0]
529 ; X32-NEXT: retl
533 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],zero,zero,xmm0[0]
534 ; X64-NEXT: retq
546 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],zero,xmm0[1],zero
547 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1,2],xmm1[2]
548 ; X32-NEXT: retl
552 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],zero,xmm0[1],zero
553 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1,2],xmm1[2]
554 ; X64-NEXT: retq
566 ; X32-NEXT: pxor %xmm1, %xmm1
567 ; X32-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1,2,3,4,5],xmm1[6,7]
568 ; X32-NEXT: retl
572 ; X64-NEXT: pxor %xmm1, %xmm1
573 ; X64-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1,2,3,4,5],xmm1[6,7]
574 ; X64-NEXT: retq
588 ; X32-NEXT: movq {{.*#+}} xmm0 = xmm0[0],zero
589 ; X32-NEXT: retl
593 ; X64-NEXT: movq {{.*#+}} xmm0 = xmm0[0],zero
594 ; X64-NEXT: retq
607 ; X32-NEXT: pshufd {{.*#+}} xmm1 = xmm0[0,1,1,3]
608 ; X32-NEXT: pxor %xmm0, %xmm0
609 ; X32-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1,2,3,4,5],xmm0[6,7]
610 ; X32-NEXT: retl
614 ; X64-NEXT: pshufd {{.*#+}} xmm1 = xmm0[0,1,1,3]
615 ; X64-NEXT: pxor %xmm0, %xmm0
616 ; X64-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1,2,3,4,5],xmm0[6,7]
617 ; X64-NEXT: retq
630 ; X32-NEXT: pshufd {{.*#+}} xmm1 = xmm0[0,1,3,3]
631 ; X32-NEXT: pxor %xmm0, %xmm0
632 ; X32-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1,2,3,4,5],xmm0[6,7]
633 ; X32-NEXT: retl
637 ; X64-NEXT: pshufd {{.*#+}} xmm1 = xmm0[0,1,3,3]
638 ; X64-NEXT: pxor %xmm0, %xmm0
639 ; X64-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1,2,3,4,5],xmm0[6,7]
640 ; X64-NEXT: retq
654 ; X32-NEXT: pshufd {{.*#+}} xmm1 = xmm0[3,1,2,3]
655 ; X32-NEXT: pxor %xmm0, %xmm0
656 ; X32-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3,4,5],xmm1[6,7]
657 ; X32-NEXT: retl
661 ; X64-NEXT: pshufd {{.*#+}} xmm1 = xmm0[3,1,2,3]
662 ; X64-NEXT: pxor %xmm0, %xmm0
663 ; X64-NEXT: pblendw {{.*#+}} xmm0 = xmm1[0,1],xmm0[2,3,4,5],xmm1[6,7]
664 ; X64-NEXT: retq
676 ; X32-NEXT: pxor %xmm2, %xmm2
677 ; X32-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1],xmm2[2,3,4,5,6,7]
678 ; X32-NEXT: pshufd {{.*#+}} xmm1 = xmm1[0,1,2,0]
679 ; X32-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1,2,3,4,5],xmm1[6,7]
680 ; X32-NEXT: retl
684 ; X64-NEXT: pxor %xmm2, %xmm2
685 ; X64-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1],xmm2[2,3,4,5,6,7]
686 ; X64-NEXT: pshufd {{.*#+}} xmm1 = xmm1[0,1,2,0]
687 ; X64-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1,2,3,4,5],xmm1[6,7]
688 ; X64-NEXT: retq
700 ; X32-NEXT: pxor %xmm1, %xmm1
701 ; X32-NEXT: pshufd {{.*#+}} xmm0 = xmm0[0,1,2,0]
702 ; X32-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,3,4,5],xmm0[6,7]
703 ; X32-NEXT: retl
707 ; X64-NEXT: pxor %xmm1, %xmm1
708 ; X64-NEXT: pshufd {{.*#+}} xmm0 = xmm0[0,1,2,0]
709 ; X64-NEXT: pblendw {{.*#+}} xmm0 = xmm0[0,1],xmm1[2,3,4,5],xmm0[6,7]
710 ; X64-NEXT: retq
722 ; X32-NEXT: pmovzxdq {{.*#+}} xmm2 = xmm0[0],zero,xmm0[1],zero
723 ; X32-NEXT: pshufd {{.*#+}} xmm0 = xmm1[0,1,2,2]
724 ; X32-NEXT: pblendw {{.*#+}} xmm0 = xmm2[0,1,2,3,4,5],xmm0[6,7]
725 ; X32-NEXT: retl
729 ; X64-NEXT: pmovzxdq {{.*#+}} xmm2 = xmm0[0],zero,xmm0[1],zero
730 ; X64-NEXT: pshufd {{.*#+}} xmm0 = xmm1[0,1,2,2]
731 ; X64-NEXT: pblendw {{.*#+}} xmm0 = xmm2[0,1,2,3,4,5],xmm0[6,7]
732 ; X64-NEXT: retq
745 ; X32-NEXT: xorps %xmm1, %xmm1
746 ; X32-NEXT: blendps {{.*#+}} xmm1 = xmm0[0,1,2],xmm1[3]
747 ; X32-NEXT: maxps %xmm1, %xmm0
748 ; X32-NEXT: retl
752 ; X64-NEXT: xorps %xmm1, %xmm1
753 ; X64-NEXT: blendps {{.*#+}} xmm1 = xmm0[0,1,2],xmm1[3]
754 ; X64-NEXT: maxps %xmm1, %xmm0
755 ; X64-NEXT: retq
771 ; X32-NEXT: psllw $15, %xmm0
772 ; X32-NEXT: psraw $15, %xmm0
773 ; X32-NEXT: pblendvb %xmm1, %xmm2
774 ; X32-NEXT: movdqa %xmm2, %xmm0
775 ; X32-NEXT: retl
779 ; X64-NEXT: psllw $15, %xmm0
780 ; X64-NEXT: psraw $15, %xmm0
781 ; X64-NEXT: pblendvb %xmm1, %xmm2
782 ; X64-NEXT: movdqa %xmm2, %xmm0
783 ; X64-NEXT: retq
792 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
793 ; X32-NEXT: insertps $48, (%{{...}}), {{.*#+}} xmm0 = xmm0[0,1,2],mem[0]
794 ; X32-NEXT: retl
798 ; X64-NEXT: insertps $48, (%{{...}}), {{.*#+}} xmm0 = xmm0[0,1,2],mem[0]
799 ; X64-NEXT: retq
810 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
811 ; X32-NEXT: insertps $32, 4(%{{...}}), {{.*#+}} xmm0 = xmm0[0,1],mem[0],xmm0[3]
812 ; X32-NEXT: retl
816 ; X64-NEXT: insertps $32, 4(%{{...}}), {{.*#+}} xmm0 = xmm0[0,1],mem[0],xmm0[3]
817 ; X64-NEXT: retq
827 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
828 ; X32-NEXT: movl {{[0-9]+}}(%esp), %ecx
829 ; X32-NEXT: shll $4, %ecx
830 ; X32-NEXT: insertps $0, 12(%{{...}},%{{...}}), {{.*#+}} xmm0 = mem[0],xmm0[1,2,3]
831 ; X32-NEXT: retl
835 ; X64-NEXT: shlq $4, %rsi
836 ; X64-NEXT: insertps $0, 12(%{{...}},%{{...}}), {{.*#+}} xmm0 = mem[0],xmm0[1,2,3]
837 ; X64-NEXT: retq
847 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
848 ; X32-NEXT: movl {{[0-9]+}}(%esp), %ecx
849 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1,2],mem[0]
850 ; X32-NEXT: retl
854 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1,2],mem[0]
855 ; X64-NEXT: retq
869 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
870 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1,2],mem[0]
871 ; X32-NEXT: retl
875 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1,2],mem[0]
876 ; X64-NEXT: retq
890 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
891 ; X32-NEXT: movl {{[0-9]+}}(%esp), %ecx
892 ; X32-NEXT: movss {{.*#+}} xmm4 = mem[0],zero,zero,zero
893 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1,2],xmm4[0]
894 ; X32-NEXT: insertps {{.*#+}} xmm1 = xmm1[0,1,2],xmm4[0]
895 ; X32-NEXT: insertps {{.*#+}} xmm2 = xmm2[0,1,2],xmm4[0]
896 ; X32-NEXT: insertps {{.*#+}} xmm3 = xmm3[0,1,2],xmm4[0]
897 ; X32-NEXT: addps %xmm1, %xmm0
898 ; X32-NEXT: addps %xmm2, %xmm3
899 ; X32-NEXT: addps %xmm3, %xmm0
900 ; X32-NEXT: retl
904 ; X64-NEXT: movss {{.*#+}} xmm4 = mem[0],zero,zero,zero
905 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0,1,2],xmm4[0]
906 ; X64-NEXT: insertps {{.*#+}} xmm1 = xmm1[0,1,2],xmm4[0]
907 ; X64-NEXT: insertps {{.*#+}} xmm2 = xmm2[0,1,2],xmm4[0]
908 ; X64-NEXT: insertps {{.*#+}} xmm3 = xmm3[0,1,2],xmm4[0]
909 ; X64-NEXT: addps %xmm1, %xmm0
910 ; X64-NEXT: addps %xmm2, %xmm3
911 ; X64-NEXT: addps %xmm3, %xmm0
912 ; X64-NEXT: retq
932 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
933 ; X32-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
934 ; X32-NEXT: unpcklpd {{.*#+}} xmm1 = xmm1[0],xmm0[0]
935 ; X32-NEXT: movapd %xmm1, %xmm0
936 ; X32-NEXT: retl
940 ; X64-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
941 ; X64-NEXT: unpcklpd {{.*#+}} xmm1 = xmm1[0],xmm0[0]
942 ; X64-NEXT: movapd %xmm1, %xmm0
943 ; X64-NEXT: retq
955 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
956 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],zero,xmm0[2],mem[0]
957 ; X32-NEXT: retl
961 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],zero,xmm0[2],mem[0]
962 ; X64-NEXT: retq
972 ; X32-NEXT: movl {{[0-9]+}}(%esp), %eax
973 ; X32-NEXT: pshufd {{.*#+}} xmm1 = xmm1[2,3,0,1]
974 ; X32-NEXT: pblendw {{.*#+}} xmm1 = xmm0[0,1],xmm1[2,3],xmm0[4,5,6,7]
975 ; X32-NEXT: movdqu %xmm1, (%eax)
976 ; X32-NEXT: retl
980 ; X64-NEXT: pshufd {{.*#+}} xmm1 = xmm1[2,3,0,1]
981 ; X64-NEXT: pblendw {{.*#+}} xmm1 = xmm0[0,1],xmm1[2,3],xmm0[4,5,6,7]
982 ; X64-NEXT: movdqu %xmm1, (%rdi)
983 ; X64-NEXT: retq
993 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],zero,xmm1[2],zero
994 ; X32-NEXT: retl
998 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],zero,xmm1[2],zero
999 ; X64-NEXT: retq
1013 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],xmm1[1],zero,zero
1014 ; X32-NEXT: retl
1018 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],xmm1[1],zero,zero
1019 ; X64-NEXT: retq
1033 ; X32-NEXT: insertps {{.*#+}} xmm0 = zero,xmm0[1],xmm1[2],zero
1034 ; X32-NEXT: retl
1038 ; X64-NEXT: insertps {{.*#+}} xmm0 = zero,xmm0[1],xmm1[2],zero
1039 ; X64-NEXT: retq
1052 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],zero,xmm1[1],zero
1053 ; X32-NEXT: retl
1057 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],zero,xmm1[1],zero
1058 ; X64-NEXT: retq
1072 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],xmm1[0],zero,zero
1073 ; X32-NEXT: retl
1077 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],xmm1[0],zero,zero
1078 ; X64-NEXT: retq
1092 ; X32-NEXT: insertps {{.*#+}} xmm1 = zero,xmm0[0],xmm1[2],zero
1093 ; X32-NEXT: movaps %xmm1, %xmm0
1094 ; X32-NEXT: retl
1098 ; X64-NEXT: insertps {{.*#+}} xmm1 = zero,xmm0[0],xmm1[2],zero
1099 ; X64-NEXT: movaps %xmm1, %xmm0
1100 ; X64-NEXT: retq
1113 ; X32-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],zero,xmm0[0],zero
1114 ; X32-NEXT: retl
1118 ; X64-NEXT: insertps {{.*#+}} xmm0 = xmm0[0],zero,xmm0[0],zero
1119 ; X64-NEXT: retq
1130 ; X32-NEXT: xorps %xmm1, %xmm1
1131 ; X32-NEXT: blendps {{.*#+}} xmm0 = xmm1[0],xmm0[1],xmm1[2],xmm0[3]
1132 ; X32-NEXT: retl
1136 ; X64-NEXT: xorps %xmm1, %xmm1
1137 ; X64-NEXT: blendps {{.*#+}} xmm0 = xmm1[0],xmm0[1],xmm1[2],xmm0[3]
1138 ; X64-NEXT: retq
1150 ; X32-NEXT: xorps %xmm1, %xmm1
1151 ; X32-NEXT: blendps {{.*#+}} xmm0 = xmm1[0],xmm0[1],xmm1[2,3]
1152 ; X32-NEXT: retl
1156 ; X64-NEXT: xorps %xmm1, %xmm1
1157 ; X64-NEXT: blendps {{.*#+}} xmm0 = xmm1[0],xmm0[1],xmm1[2,3]
1158 ; X64-NEXT: retq