Lines Matching full:ymm0
19 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,0,1,0]
20 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
30 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,2,0,0]
31 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
41 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[3,0,0,0]
42 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
53 ; ALL-NEXT: vbroadcastsd %xmm0, %ymm0
54 ; ALL-NEXT: vblendpd {{.*#+}} ymm1 = ymm0[0,1,2],ymm1[3]
55 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
65 ; ALL-NEXT: vblendpd {{.*#+}} ymm1 = ymm0[0],ymm1[1],ymm0[2,3]
67 ; ALL-NEXT: vbroadcastsd %xmm0, %ymm0
68 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
78 ; ALL-NEXT: vblendpd {{.*#+}} ymm1 = ymm0[0,1],ymm1[2],ymm0[3]
80 ; ALL-NEXT: vbroadcastsd %xmm0, %ymm0
81 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
91 ; ALL-NEXT: vblendpd {{.*#+}} ymm1 = ymm0[0,1,2],ymm1[3]
93 ; ALL-NEXT: vbroadcastsd %xmm0, %ymm0
94 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
105 ; ALL-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
115 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[0,0,1,1]
116 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[2,2,3,3]
117 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
127 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[1,1,1,1]
128 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
140 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm1[0],ymm0[1],ymm1[2],ymm0[3]
150 ; ALL-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
152 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2],ymm1[3]
153 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm0, %zmm0
167 ; ALL-NEXT: vinsertf128 $1, %xmm0, %ymm0, %ymm0
169 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2],ymm1[3]
184 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm1[0,1],ymm0[2,3]
194 ; ALL-NEXT: vblendpd {{.*#+}} ymm2 = ymm1[0,1],ymm0[2,3]
196 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
198 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm1[0,1],ymm0[2,3]
199 ; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
200 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm2, %zmm0
212 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
213 ; ALL-NEXT: vinsertf128 $1, %xmm0, %ymm1, %ymm0
214 ; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
215 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm2, %zmm0
230 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,1,1,3]
231 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2],ymm1[3]
247 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[2,1,3,3]
248 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2],ymm1[3]
259 ; ALL-NEXT: vpermpd {{.*#+}} ymm3 = ymm0[2,1,3,3]
262 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,1,1,3]
263 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2],ymm1[3]
274 ; ALL-NEXT: vblendpd {{.*#+}} ymm2 = ymm0[0],ymm2[1,2,3]
275 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm1[0],ymm0[1],ymm1[2,3]
276 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[1,2,3,3]
277 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm2, %zmm0
287 ; ALL-NEXT: vpermpd {{.*#+}} ymm3 = ymm0[2,1,3,3]
289 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,1,1,3]
290 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1],ymm0[2],ymm1[3]
301 ; ALL-NEXT: vpermilpd {{.*#+}} ymm3 = ymm0[1,0,2,2]
303 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm1[1,2,3]
313 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[0,0,0,1]
314 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
315 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,0,0,1]
316 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
325 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[0,0,2,0]
326 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
327 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,0,2,0]
328 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
337 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[0,3,0,0]
338 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
339 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,3,0,0]
340 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
349 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[1,0,0,0]
350 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
351 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[1,0,0,0]
352 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
361 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[2,2,0,0]
362 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
363 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[2,2,0,0]
364 ; ALL-NEXT: vinsertf64x4 $1, %ymm0
373 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[3,3,3,0]
374 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
375 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[3,3,3,0]
376 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
385 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[3,2,1,0]
386 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
387 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[3,2,1,0]
388 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
397 ; ALL-NEXT: vpermilpd {{.*#+}} ymm1 = ymm0[0,0,2,3]
398 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
399 ; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[0,0,2,3]
400 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
409 ; ALL-NEXT: vmovddup {{.*#+}} ymm1 = ymm0[0,0,2,2]
410 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
411 ; ALL-NEXT: vmovddup {{.*#+}} ymm0 = ymm0[0,0,2,2]
412 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
421 ; ALL-NEXT: vpermilpd {{.*#+}} ymm1 = ymm0[1,0,3,2]
422 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
423 ; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,3,2]
424 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
433 ; ALL-NEXT: vpermilpd {{.*#+}} ymm1 = ymm0[1,1,3,3]
434 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
435 ; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,1,3,3]
436 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
445 ; ALL-NEXT: vpermilpd {{.*#+}} ymm1 = ymm0[1,0,2,3]
446 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
447 ; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,2,3]
448 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
457 ; ALL-NEXT: vpermilpd {{.*#+}} ymm1 = ymm0[1,0,2,2]
458 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
459 ; ALL-NEXT: vpermilpd {{.*#+}} ymm0 = ymm0[1,0,2,2]
460 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
469 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[0,0,0,1]
470 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
471 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[1,0,0,0]
472 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
481 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[0,0,2,0]
482 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
483 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,2,0,0]
484 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
493 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[0,3,0,0]
494 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
495 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,0,3,0]
496 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
505 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[1,0,0,0]
506 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
507 ; ALL-NEXT: vbroadcastsd %xmm0, %ymm0
508 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
517 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[2,2,0,0]
518 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
519 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[2,0,0,2]
520 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
529 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[3,3,3,0]
530 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
531 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[3,0,3,0]
532 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
541 ; ALL-NEXT: vpermpd {{.*#+}} ymm1 = ymm0[3,2,1,0]
542 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
543 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
552 ; ALL-NEXT: vpermilpd {{.*#+}} ymm1 = ymm0[0,0,2,3]
553 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
554 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[2,3,0,0]
555 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
564 ; ALL-NEXT: vmovddup {{.*#+}} ymm1 = ymm0[0,0,2,2]
565 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
566 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[2,2,0,0]
567 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
576 ; ALL-NEXT: vpermilpd {{.*#+}} ymm1 = ymm0[1,0,3,2]
577 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
578 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
587 ; ALL-NEXT: vpermilpd {{.*#+}} ymm1 = ymm0[1,1,3,3]
588 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
589 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
620 ; ALL-NEXT: vmovddup {{.*#+}} ymm1 = ymm0[0,0,2,2]
621 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
622 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[2,1,0,0]
623 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
633 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
634 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[2,2,2,3]
635 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
644 ; ALL-NEXT: vpermilpd {{.*#+}} ymm1 = ymm0[1,0,3,2]
645 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
646 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
655 ; ALL-NEXT: vpermilpd {{.*#+}} ymm1 = ymm0[1,1,3,3]
656 ; ALL-NEXT: vextractf64x4 $1, %zmm0, %ymm0
657 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm1, %zmm0
689 ; ALL-NEXT: vperm2f128 {{.*#+}} ymm2 = ymm0[2,3],ymm2[0,1]
695 ; ALL-NEXT: vbroadcastsd %xmm0, %ymm0
696 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm1[0,1,2],ymm0[3]
697 ; ALL-NEXT: vinsertf64x4 $1, %ymm0, %zmm2, %zmm0
707 ; ALL-NEXT: vblendpd {{.*#+}} ymm3 = ymm0[0],ymm2[1],ymm0[2,3]
711 ; ALL-NEXT: vpermpd {{.*#+}} ymm0 = ymm0[0,1,1,1]
712 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm0[0],ymm2[1],ymm0[2,3]
715 ; ALL-NEXT: vblendpd {{.*#+}} ymm0 = ymm1[0],ymm0[1,2,3]
735 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,1,0]
736 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
746 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,2,0,0]
747 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
757 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[3,0,0,0]
758 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
769 ; ALL-NEXT: vpbroadcastq %xmm0, %ymm0
770 ; ALL-NEXT: vpblendd {{.*#+}} ymm1 = ymm0[0,1,2,3,4,5],ymm1[6,7]
771 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
781 ; ALL-NEXT: vpblendd {{.*#+}} ymm1 = ymm0[0,1],ymm1[2,3],ymm0[4,5,6,7]
783 ; ALL-NEXT: vpbroadcastq %xmm0, %ymm0
784 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
794 ; ALL-NEXT: vpblendd {{.*#+}} ymm1 = ymm0[0,1,2,3],ymm1[4,5],ymm0[6,7]
796 ; ALL-NEXT: vpbroadcastq %xmm0, %ymm0
797 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
807 ; ALL-NEXT: vpblendd {{.*#+}} ymm1 = ymm0[0,1,2,3,4,5],ymm1[6,7]
809 ; ALL-NEXT: vpbroadcastq %xmm0, %ymm0
810 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
821 ; ALL-NEXT: vinserti128 $1, %xmm0, %ymm0, %ymm0
831 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[0,0,1,1]
832 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,2,3,3]
833 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
843 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[1,1,1,1]
844 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
856 ; ALL-NEXT: vpblendd {{.*#+}} ymm0 = ymm1[0,1],ymm0[2,3],ymm1[4,5],ymm0[6,7]
866 ; ALL-NEXT: vinserti128 $1, %xmm0, %ymm0, %ymm0
868 ; ALL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3],ymm0[4,5],ymm1[6,7]
869 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm0, %zmm0
883 ; ALL-NEXT: vinserti128 $1, %xmm0, %ymm0, %ymm0
885 ; ALL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3],ymm0[4,5],ymm1[6,7]
900 ; ALL-NEXT: vpblendd {{.*#+}} ymm0 = ymm1[0,1,2,3],ymm0[4,5,6,7]
910 ; ALL-NEXT: vpblendd {{.*#+}} ymm2 = ymm1[0,1,2,3],ymm0[4,5,6,7]
912 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
914 ; ALL-NEXT: vpblendd {{.*#+}} ymm0 = ymm1[0,1,2,3],ymm0[4,5,6,7]
915 ; ALL-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
916 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm2, %zmm0
928 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
929 ; ALL-NEXT: vinserti128 $1, %xmm0, %ymm1, %ymm0
930 ; ALL-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
931 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm2, %zmm0
946 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,1,3]
947 ; ALL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3],ymm0[4,5],ymm1[6,7]
963 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,1,3,3]
964 ; ALL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3],ymm0[4,5],ymm1[6,7]
975 ; ALL-NEXT: vpermq {{.*#+}} ymm3 = ymm0[2,1,3,3]
978 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,1,3]
979 ; ALL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3],ymm0[4,5],ymm1[6,7]
990 ; ALL-NEXT: vpblendd {{.*#+}} ymm2 = ymm0[0,1],ymm2[2,3,4,5,6,7]
991 ; ALL-NEXT: vpblendd {{.*#+}} ymm0 = ymm1[0,1],ymm0[2,3],ymm1[4,5,6,7]
992 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[1,2,3,3]
993 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm2, %zmm0
1003 ; ALL-NEXT: vpermq {{.*#+}} ymm3 = ymm0[2,1,3,3]
1005 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,1,1,3]
1006 ; ALL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3],ymm0[4,5],ymm1[6,7]
1017 ; ALL-NEXT: vpshufd {{.*#+}} ymm3 = ymm0[2,3,2,3,6,7,6,7]
1019 ; ALL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3,4,5,6,7]
1029 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[0,0,0,1]
1030 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1031 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,0,1]
1032 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1041 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[0,0,2,0]
1042 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1043 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,2,0]
1044 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1053 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[0,3,0,0]
1054 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1055 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,3,0,0]
1056 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1065 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[1,0,0,0]
1066 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1067 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[1,0,0,0]
1068 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1077 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[2,2,0,0]
1078 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1079 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,2,0,0]
1080 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1089 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[3,3,3,0]
1090 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1091 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[3,3,3,0]
1092 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1101 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[3,2,1,0]
1102 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1103 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[3,2,1,0]
1104 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1113 ymm0[0,0,2,3]
1114 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1115 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,2,3]
1116 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1125 ; ALL-NEXT: vpshufd {{.*#+}} ymm1 = ymm0[0,1,0,1,4,5,4,5]
1126 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1127 ; ALL-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[0,1,0,1,4,5,4,5]
1128 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1137 ; ALL-NEXT: vpshufd {{.*#+}} ymm1 = ymm0[2,3,0,1,6,7,4,5]
1138 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1139 ; ALL-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,0,1,6,7,4,5]
1140 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1149 ; ALL-NEXT: vpshufd {{.*#+}} ymm1 = ymm0[2,3,2,3,6,7,6,7]
1150 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1151 ; ALL-NEXT: vpshufd {{.*#+}} ymm0 = ymm0[2,3,2,3,6,7,6,7]
1152 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1161 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[1,0,2,3]
1162 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1163 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[1,0,2,3]
1164 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1173 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[1,0,2,2]
1174 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1175 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[1,0,2,2]
1176 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1185 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[0,0,0,1]
1186 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1187 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[1,0,0,0]
1188 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1197 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[0,0,2,0]
1198 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1199 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,2,0,0]
1200 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1209 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[0,3,0,0]
1210 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1211 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[0,0,3,0]
1212 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1221 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[1,0,0,0]
1222 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1223 ; ALL-NEXT: vpbroadcastq %xmm0, %ymm0
1224 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1233 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[2,2,0,0]
1234 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1235 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,0,0,2]
1236 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1245 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[3,3,3,0]
1246 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1247 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[3,0,3,0]
1248 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1257 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[3,2,1,0]
1258 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1259 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1268 ; ALL-NEXT: vpermq {{.*#+}} ymm1 = ymm0[0,0,2,3]
1269 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1270 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,3,0,0]
1271 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1280 ; ALL-NEXT: vpshufd {{.*#+}} ymm1 = ymm0[0,1,0,1,4,5,4,5]
1281 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1282 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,2,0,0]
1283 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1292 ; ALL-NEXT: vpshufd {{.*#+}} ymm1 = ymm0[2,3,0,1,6,7,4,5]
1293 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1294 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1303 ; ALL-NEXT: vpshufd {{.*#+}} ymm1 = ymm0[2,3,2,3,6,7,6,7]
1304 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1305 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1336 ; ALL-NEXT: vpshufd {{.*#+}} ymm1 = ymm0[0,1,0,1,4,5,4,5]
1337 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1338 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,1,0,0]
1339 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1349 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1350 ; ALL-NEXT: vpermq {{.*#+}} ymm0 = ymm0[2,2,2,3]
1351 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1360 ; ALL-NEXT: vpshufd {{.*#+}} ymm1 = ymm0[2,3,0,1,6,7,4,5]
1361 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1362 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1371 ; ALL-NEXT: vpshufd {{.*#+}} ymm1 = ymm0[2,3,2,3,6,7,6,7]
1372 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1373 ; ALL-NEXT: vinserti64x4 $1, %ymm0, %zmm1, %zmm0
1404 ; ALL-NEXT: vextracti64x4 $1, %zmm0, %ymm0
1405 ; ALL-NEXT: vperm2i128 {{.*#+}} ymm0 = ymm0[2,3,0,1]
1408 ; ALL-NEXT: vpblendd {{.*#+}} ymm3 = ymm3[0,1],ymm0[2,3],ymm3[4,5],ymm0[6,7]
1411 ; ALL-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1],ymm1[2,3,4,5,6,7]