Lines Matching full:ymm0
12 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
17 ; AVX2-NEXT: vbroadcastsd %xmm0, %ymm0
22 ; AVX512VL-NEXT: vbroadcastsd %xmm0, %ymm0
32 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
37 ; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,0,0,1]
42 ; AVX512VL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,0,0,1]
51 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm1
54 ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
59 ; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,0,2,0]
64 ; AVX512VL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,0,2,0]
73 ; AVX1-NEXT: vperm2f128 {{.*#+}} ymm1 = ymm0[2,3,0,1]
75 ; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3]
80 ; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,3,0,0]
85 ; AVX512VL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,3,0,0]
96 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
101 ; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[1,0,0,0]
106 ; AVX512VL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[1,0,0,0]
115 ; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3,0,1]
116 ; AVX1-NEXT: vmovddup {{.*#+}} ymm0 = ymm0[0,0,2,2]
121 ; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[2,2,0,0]
126 ; AVX512VL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[2,2,0,0]
135 ; AVX1-NEXT: vperm2f128 {{.*#+}} ymm1 = ymm0[2,3,0,1]
136 ; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm1[0,1,2],ymm0[3]
137 ; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,1,3,2]
142 ; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[3,3,3,0]
147 ; AVX512VL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[3,3,3,0]
156 ; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3,0,1]
157 ; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
162 ; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[3,2,1,0]
167 ; AVX512VL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[3,2,1,0]
176 ; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[0,0,2,3]
186 ; ALL-NEXT: vmovddup {{.*#+}} ymm0 = ymm0[0,0,2,2]
195 ; ALL-NEXT: vmovddup {{.*#+}} ymm0 = mem[0,0,2,2]
205 ; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
214 ; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,1,3,3]
223 ; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,2,3]
232 ; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,2,2]
242 ; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2,3]
248 ; AVX2-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2,3]
254 ; AVX512VL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2,3]
264 ; ALL-NEXT: vmovddup {{.*#+}} ymm0 = ymm0[0,0,2,2]
265 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1,2],ymm0[3]
274 ; ALL-NEXT: vunpcklpd {{.*#+}} ymm0 = ymm0[0],ymm1[0],ymm0[2],ymm1[2]
283 ; ALL-NEXT: vunpckhpd {{.*#+}} ymm0 = ymm0[1],ymm1[1],ymm0[3],ymm1[3]
292 ; ALL-NEXT: vunpcklpd {{.*#+}} ymm0 = ymm1[0],ymm0[0],ymm1[2],ymm0[2]
301 ; ALL-NEXT: vunpckhpd {{.*#+}} ymm0 = ymm1[1],ymm0[1],ymm1[3],ymm0[3]
310 ; ALL-NEXT: vshufpd {{.*#+}} ymm0 = ymm1[1],ymm0[1],ymm1[2],ymm0[3]
319 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2],ymm1[3]
328 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm1[0],ymm0[1],ymm1[2],ymm0[3]
337 ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
342 ; AVX2-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
347 ; AVX512VL-NEXT: vinsertf32x4 $1, %xmm1, %ymm0, %ymm0
356 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
361 ; AVX2-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
366 ; AVX512VL-NEXT: vinsertf32x4 $1, %xmm0, %ymm1, %ymm0
375 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3]
384 ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
385 ; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
390 ; AVX2-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
391 ; AVX2-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
396 ; AVX512VL-NEXT: vinsertf32x4 $1, %xmm1, %ymm0, %ymm0
397 ; AVX512VL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
406 ; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[0,1]
407 ; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
412 ; AVX2-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[0,1]
413 ; AVX2-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
418 ; AVX512VL-NEXT: vperm2i128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[0,1]
419 ; AVX512VL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
428 ; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[2,3]
429 ; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
434 ; AVX2-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[2,3]
435 ; AVX2-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
440 ; AVX512VL-NEXT: vperm2i128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[2,3]
441 ; AVX512VL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
450 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3]
451 ymm0 = ymm0[1,0,3,2]
462 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
468 ; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,1,1,3]
469 ; AVX2-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2],ymm1[3]
475 ; AVX512VL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,1,1,3]
476 ; AVX512VL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2],ymm1[3]
485 ; ALL-NEXT: vunpcklpd {{.*#+}} ymm0 = ymm1[0],ymm0[0],ymm1[2],ymm0[2]
494 ; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,1,2,2]
503 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm0
509 ; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[2,2,2,3]
514 ; AVX512VL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[2,2,2,3]
523 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm0
525 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
530 ; AVX2-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[3,3,3,3]
535 ; AVX512VL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[3,3,3,3]
545 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
550 ; AVX2-NEXT: vbroadcastsd %xmm0, %ymm0
555 ; AVX512VL-NEXT: vpbroadcastq %xmm0, %ymm0
565 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
570 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,0,1]
575 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,0,1]
584 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm1
587 ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
592 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,2,0]
597 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,2,0]
606 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm1
608 ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
613 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,1,2]
618 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,1,2]
627 ; AVX1-NEXT: vperm2f128 {{.*#+}} ymm1 = ymm0[2,3,0,1]
629 ; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3]
634 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,3,0,0]
639 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,3,0,0]
650 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
655 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[1,0,0,0]
660 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[1,0,0,0]
669 ; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3,0,1]
670 ; AVX1-NEXT: vmovddup {{.*#+}} ymm0 = ymm0[0,0,2,2]
675 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,2,0,0]
680 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,2,0,0]
689 ; AVX1-NEXT: vperm2f128 {{.*#+}} ymm1 = ymm0[2,3,0,1]
690 ; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm1[0,1,2],ymm0[3]
691 ; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,1,3,2]
696 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[3,3,3,0]
701 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[3,3,3,0]
710 ; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3,0,1]
711 ; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
716 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[3,2,1,0]
721 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[3,2,1,0]
731 ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm1
732 ; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0,1,2],ymm1[3]
738 ; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3,4,5],ymm1[6,7]
744 ; AVX512VL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3,4,5],ymm1[6,7]
754 ; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[0,1,2,2]
755 ; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2],ymm0[3]
761 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,2,2]
762 ; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3],ymm1[4,5],ymm0[6,7]
768 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,2,2]
769 ; AVX512VL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3],ymm1[4,5],ymm0[6,7]
778 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
780 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
782 ; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2,3]
787 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,1,2]
789 ; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3],ymm0[4,5,6,7]
794 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,1,2]
796 ; AVX512VL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3],ymm0[4,5,6,7]
805 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
808 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
809 ; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm1[0],ymm0[1,2,3]
814 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,1,2]
815 ; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm1[0,1],ymm0[2,3,4,5,6,7]
820 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,1,2]
821 ; AVX512VL-NEXT: vpblendd {{.*#+}} ymm0 = ymm1[0,1],ymm0[2,3,4,5,6,7]
830 ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
835 ; AVX2-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
840 ; AVX512VL-NEXT: vinserti32x4 $1, %xmm1, %ymm0, %ymm0
851 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
856 ; AVX2-NEXT: vinserti128 $1, %xmm0, %ymm0, %ymm0
858 ; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3,4,5],ymm0[6,7]
863 ; AVX512VL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0
865 ; AVX512VL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3,4,5],ymm0[6,7]
874 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
879 ; AVX2-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
884 ; AVX512VL-NEXT: vinserti32x4 $1, %xmm0, %ymm1, %ymm0
895 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
901 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,1,3]
902 ; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm1[0,1],ymm0[2,3,4,5],ymm1[6,7]
908 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,1,3]
909 ; AVX512VL-NEXT: vpblendd {{.*#+}} ymm0 = ymm1[0,1],ymm0[2,3,4,5],ymm1[6,7]
918 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm0
920 ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
925 ; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3],ymm0[4,5,6,7]
926 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,1,3,1]
931 ; AVX512VL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3],ymm0[4,5,6,7]
932 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,1,3,1]
941 ; AVX1-NEXT: vperm2f128 {{.*#+}} ymm2 = ymm0[2,3,0,1]
942 ; AVX1-NEXT: vshufpd {{.*#+}} ymm0 = ymm0[1],ymm2[0],ymm0[2],ymm2[3]
944 ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm1
945 ; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2],ymm0[3]
951 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[1,2,2,1]
952 ; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3],ymm1[4,5],ymm0[6,7]
958 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[1,2,2,1]
959 ; AVX512VL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3],ymm1[4,5],ymm0[6,7]
968 ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
969 ; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
974 ; AVX2-NEXT: vinserti128 $1, %xmm1, %ymm0, %ymm0
975 ; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
980 ; AVX512VL-NEXT: vinserti32x4 $1, %xmm1, %ymm0, %ymm0
981 ; AVX512VL-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
990 ; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[0,1]
991 ; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
996 ; AVX2-NEXT: vperm2i128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[0,1]
997 ; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
1002 ; AVX512VL-NEXT: vperm2i128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[0,1]
1003 ; AVX512VL-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
1012 ; AVX1-NEXT: vperm2f128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[2,3]
1013 ; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
1018 ; AVX2-NEXT: vperm2i128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[2,3]
1019 ; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
1024 ; AVX512VL-NEXT: vperm2i128 {{.*#+}} ymm0 = ymm0[2,3],ymm1[2,3]
1025 ; AVX512VL-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
1034 ; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3]
1035 ; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
1040 ; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3],ymm1[4,5,6,7]
1041 ; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
1046 ; AVX512VL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3],ymm1[4,5,6,7]
1047 ; AVX512VL-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
1058 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
1064 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,1,3]
1065 ; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3],ymm0[4,5],ymm1[6,7]
1071 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,1,3]
1072 ; AVX512VL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3],ymm0[4,5],ymm1[6,7]
1082 ; AVX1-NEXT: vunpcklpd {{.*#+}} ymm0 = ymm1[0],ymm0[0],ymm1[2],ymm0[2]
1087 ; AVX2-NEXT: vpslldq {{.*#+}} ymm0 = zero,zero,zero,zero,zero,zero,zero,zero,ymm0[0,1,2,3,4,5,6,7],zero,zero,zero,zero,zero,zero,zero,zero,ymm0[16,17,18,19,20,21,22,23]
1092 ; AVX512VL-NEXT: vpslldq {{.*#+}} ymm0 = zero,zero,zero,zero,zero,zero,zero,zero,ymm0[0,1,2,3,4,5,6,7],zero,zero,zero,zero,zero,zero,zero,zero,ymm0[16,17,18,19,20,21,22,23]
1102 ; AVX1-NEXT: vunpckhpd {{.*#+}} ymm0 = ymm0[1],ymm1[1],ymm0[3],ymm1[3]
1107 ; AVX2-NEXT: vpsrldq {{.*#+}} ymm0 = ymm0[8,9,10,11,12,13,14,15],zero,zero,zero,zero,zero,zero,zero,zero,ymm0[24,25,26,27,28,29,30,31],zero,zero,zero,zero,zero,zero,zero,zero
1112 ; AVX512VL-NEXT: vpsrldq {{.*#+}} ymm0 = ymm0[8,9,10,11,12,13,14,15],zero,zero,zero,zero,zero,zero,zero,zero,ymm0[24,25,26,27,28,29,30,31],zero,zero,zero,zero,zero,zero,zero,zero
1121 ; AVX1-NEXT: vunpcklpd {{.*#+}} ymm0 = ymm1[0],ymm0[0],ymm1[2],ymm0[2]
1126 ; AVX2-NEXT: vpunpcklqdq {{.*#+}} ymm0 = ymm1[0],ymm0[0],ymm1[2],ymm0[2]
1131 ; AVX512VL-NEXT: vpunpcklqdq {{.*#+}} ymm0 = ymm1[0],ymm0[0],ymm1[2],ymm0[2]
1140 ; AVX1-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,1,2,2]
1145 ; AVX2-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,2,3,6,7,6,7]
1150 ; AVX512VL-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,2,3,6,7,6,7]
1159 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm0
1165 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,2,2,3]
1170 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,2,2,3]
1179 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm0
1181 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
1186 ; AVX2-NEXT: vpermq {{.*#+}} ymm0 = ymm0[3,3,3,3]
1191 ; AVX512VL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[3,3,3,3]
1243 ; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3]
1249 ; AVX2-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3]
1286 ; ALL-NEXT: vbroadcastsd (%rdi), %ymm0
1297 ; AVX1-NEXT: vbroadcastsd (%rdi), %ymm0
1302 ; AVX2-NEXT: vbroadcastsd (%rdi), %ymm0
1307 ; AVX512VL-NEXT: vpbroadcastq (%rdi), %ymm0
1318 ; ALL-NEXT: vbroadcastsd (%rdi), %ymm0
1330 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
1335 ; AVX2-NEXT: vbroadcastsd %xmm0, %ymm0
1340 ; AVX512VL-NEXT: vbroadcastsd %xmm0, %ymm0
1350 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
1355 ; AVX2-NEXT: vbroadcastsd (%rdi), %ymm0
1360 ; AVX512VL-NEXT: vpbroadcastq (%rdi), %ymm0
1370 ; ALL-NEXT: vbroadcastsd (%rdi), %ymm0
1381 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
1387 ; AVX2-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
1393 ; AVX512VL-NEXT: vinserti32x4 $1, %xmm0, %ymm0, %ymm0
1404 ; AVX1-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
1410 ; AVX2-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
1416 ; AVX512VL-NEXT: vinsertf32x4 $1, %xmm0, %ymm0, %ymm0
1426 ; AVX1-NEXT: vunpcklpd {{.*#+}} ymm0 = ymm0[0],ymm1[0],ymm0[2],ymm1[2]
1431 ; AVX2-NEXT: vpunpcklqdq {{.*#+}} ymm0 = ymm0[0],ymm1[0],ymm0[2],ymm1[2]
1436 ; AVX512VL-NEXT: vpunpcklqdq {{.*#+}} ymm0 = ymm0[0],ymm1[0],ymm0[2],ymm1[2]
1450 ; AVX1-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3]
1455 ; AVX2-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3],ymm1[4,5,6,7]
1460 ; AVX512VL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3],ymm1[4,5,6,7]
1471 ; AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
1476 ; AVX2-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
1481 ; AVX512VL-NEXT: vinserti32x4 $1, %xmm1, %ymm0, %ymm0
1495 ; AVX1-NEXT: vbroadcastsd (%rdi), %ymm0
1500 ; AVX2-NEXT: vbroadcastsd (%rdi), %ymm0
1505 ; AVX512VL-NEXT: vpbroadcastq (%rdi), %ymm0