Home | History | Annotate | Download | only in X86

Lines Matching full:xmm2

294 ; AVX1-NEXT:    vpxor %xmm2, %xmm2, %xmm2
295 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
316 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
317 ; AVX1-NEXT: vpunpcklbw {{.*#+}} xmm0 = xmm0[0],xmm2[0],xmm0[1],xmm2[1],xmm0[2],xmm2[2],xmm0[3],xmm2[3],xmm0[4],xmm2[4],xmm0[5],xmm2[5],xmm0[6],xmm2[6],xmm0[7],xmm2[7]
340 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
341 ; AVX1-NEXT: vpunpcklbw {{.*#+}} xmm0 = xmm0[0],xmm2[0],xmm0[1],xmm2[1],xmm0[2],xmm2[2],xmm0[3],xmm2[3],xmm0[4],xmm2[4],xmm0[5],xmm2[5],xmm0[6],xmm2[6],xmm0[7],xmm2[7]
362 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
363 ; AVX1-NEXT: vpunpcklbw {{.*#+}} xmm0 = xmm0[0],xmm2[0],xmm0[1],xmm2[1],xmm0[2],xmm2[2],xmm0[3],xmm2[3],xmm0[4],xmm2[4],xmm0[5],xmm2[5],xmm0[6],xmm2[6],xmm0[7],xmm2[7]
384 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
385 ; AVX1-NEXT: vpunpcklbw {{.*#+}} xmm0 = xmm0[0],xmm2[0],xmm0[1],xmm2[1],xmm0[2],xmm2[2],xmm0[3],xmm2[3],xmm0[4],xmm2[4],xmm0[5],xmm2[5],xmm0[6],xmm2[6],xmm0[7],xmm2[7]
406 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
407 ; AVX1-NEXT: vpunpcklbw {{.*#+}} xmm0 = xmm0[0],xmm2[0],xmm0[1],xmm2[1],xmm0[2],xmm2[2],xmm0[3],xmm2[3],xmm0[4],xmm2[4],xmm0[5],xmm2[5],xmm0[6],xmm2[6],xmm0[7],xmm2[7]
427 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
428 ; AVX1-NEXT: vpunpcklbw {{.*#+}} xmm0 = xmm0[0],xmm2[0],xmm0[1],xmm2[1],xmm0[2],xmm2[2],xmm0[3],xmm2[3],xmm0[4],xmm2[4],xmm0[5],xmm2[5],xmm0[6],xmm2[6],xmm0[7],xmm2[7]
448 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
449 ; AVX1-NEXT: vpunpcklbw {{.*#+}} xmm0 = xmm0[0],xmm2[0],xmm0[1],xmm2[1],xmm0[2],xmm2[2],xmm0[3],xmm2[3],xmm0[4],xmm2[4],xmm0[5],xmm2[5],xmm0[6],xmm2[6],xmm0[7],xmm2[7]
469 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
470 ; AVX1-NEXT: vpunpcklbw {{.*#+}} xmm0 = xmm0[0],xmm2[0],xmm0[1],xmm2[1],xmm0[2],xmm2[2],xmm0[3],xmm2[3],xmm0[4],xmm2[4],xmm0[5],xmm2[5],xmm0[6],xmm2[6],xmm0[7],xmm2[7]
490 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
491 ; AVX1-NEXT: vpshufb {{.*#+}} xmm2 = zero,zero,zero,zero,zero,zero,zero,xmm2[8],zero,zero,zero,zero,zero,zero,zero,zero
493 ; AVX1-NEXT: vpor %xmm2, %xmm0, %xmm0
512 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
513 ; AVX1-NEXT: vpshufb {{.*#+}} xmm2 = zero,zero,zero,zero,zero,zero,xmm2[9],zero,zero,zero,zero,zero,zero,zero,zero,zero
515 ; AVX1-NEXT: vpor %xmm2, %xmm0, %xmm0
534 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
535 ; AVX1-NEXT: vpshufb {{.*#+}} xmm2 = zero,zero,zero,zero,zero,xmm2[10],zero,zero,zero,zero,zero,zero,zero,zero,zero,zero
537 ; AVX1-NEXT: vpor %xmm2, %xmm0, %xmm0
556 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
557 ; AVX1-NEXT: vpshufb {{.*#+}} xmm2 = zero,zero,zero,zero,xmm2[11],zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero
559 ; AVX1-NEXT: vpor %xmm2, %xmm0, %xmm0
578 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
579 ; AVX1-NEXT: vpshufb {{.*#+}} xmm2 = zero,zero,zero,xmm2[12],zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero
581 ; AVX1-NEXT: vpor %xmm2, %xmm0, %xmm0
600 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
601 ; AVX1-NEXT: vpshufb {{.*#+}} xmm2 = zero,zero,xmm2[13],zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero
603 ; AVX1-NEXT: vpor %xmm2, %xmm0, %xmm0
622 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
623 ; AVX1-NEXT: vpshufb {{.*#+}} xmm2 = zero,xmm2[14],zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero
625 ; AVX1-NEXT: vpor %xmm2, %xmm0, %xmm0
645 ; AVX1-NEXT: vmovd %eax, %xmm2
646 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm2
649 ; AVX1-NEXT: vpor %xmm0, %xmm2, %xmm0
669 ; AVX1-NEXT: vpxor %xmm2, %xmm2, %xmm2
670 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
671 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
688 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [15,15,15,15,15,15,15,15,15,15,15,15,15,15,15,15]
689 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
690 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
706 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,0,0,0,0,0,0,0,8,8,8,8,8,8,8,8]
707 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
708 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
724 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [7,7,7,7,7,7,7,7,15,15,15,15,15,15,15,15]
725 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
726 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
742 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2
743 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
744 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
760 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [3,3,3,3,7,7,7,7,11,11,11,11,15,15,15,15]
761 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
762 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
778 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,0,2,2,4,4,6,6,8,8,10,10,12,12,14,14]
779 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
780 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
796 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [1,1,3,3,5,5,7,7,9,9,11,11,13,13,15,15]
797 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
798 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
913 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
916 ; AVX1-NEXT: vpblendvb %xmm4, %xmm2, %xmm3, %xmm2
918 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
933 ; AVX1-NEXT: vextractf128 $1, %ymm1, %xmm2
936 ; AVX1-NEXT: vpblendvb %xmm4, %xmm2, %xmm3, %xmm2
938 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
972 ; AVX1-NEXT: vextractf128 $1, %ymm1, %xmm2
974 ; AVX1-NEXT: vpunpcklbw {{.*#+}} xmm2 = xmm3[0],xmm2[0],xmm3[1],xmm2[1],xmm3[2],xmm2[2],xmm3[3],xmm2[3],xmm3[4],xmm2[4],xmm3[5],xmm2[5],xmm3[6],xmm2[6],xmm3[7],xmm2[7]
976 ; AVX1-NEXT: vpshufb %xmm3, %xmm2, %xmm2
979 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
997 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
998 ; AVX1-NEXT: vpshufd {{.*#+}} xmm2 = xmm2[2,3,0,1]
1002 ; AVX1-NEXT: vpunpcklqdq {{.*#+}} xmm2 = xmm3[0],xmm2[0]
1007 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
1023 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
1025 ; AVX1-NEXT: vpshufb %xmm3, %xmm2, %xmm2
1029 ; AVX1-NEXT: vpunpcklqdq {{.*#+}} xmm2 = xmm4[0],xmm2[0]
1033 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
1048 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
1050 ; AVX1-NEXT: vpunpcklbw {{.*#+}} xmm2 = xmm3[0],xmm2[0],xmm3[1],xmm2[1],xmm3[2],xmm2[2],xmm3[3],xmm2[3],xmm3[4],xmm2[4],xmm3[5],xmm2[5],xmm3[6],xmm2[6],xmm3[7],xmm2[7]
1052 ; AVX1-NEXT: vpshufb %xmm3, %xmm2, %xmm2
1055 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
1072 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0]
1073 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
1074 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
1090 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,0,0,0,0,0,0,0,0,0,0,0,0,2,0,0]
1091 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
1092 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
1108 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,0,0,0,0,0,0,0,7,0,0,0,0,0,0,0]
1109 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
1110 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
1126 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,0,0,0,0,0,0,8,0,0,0,0,0,0,0,0]
1127 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
1128 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
1144 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,14,0,0,0,0,0,0,0,0,0,0,0,0,0,0]
1145 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
1146 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
1163 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
1164 ; AVX1-NEXT: vpshufb %xmm1, %xmm2, %xmm2
1166 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
1180 ; AVX1-NEXT: vextractf128 $1, %ymm1, %xmm2
1182 ; AVX1-NEXT: vpunpcklbw {{.*#+}} xmm2 = xmm3[0],xmm2[0],xmm3[1],xmm2[1],xmm3[2],xmm2[2],xmm3[3],xmm2[3],xmm3[4],xmm2[4],xmm3[5],xmm2[5],xmm3[6],xmm2[6],xmm3[7],xmm2[7]
1184 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
1198 ; AVX1-NEXT: vextractf128 $1, %ymm1, %xmm2
1200 ; AVX1-NEXT: vpunpckhbw {{.*#+}} xmm2 = xmm3[8],xmm2[8],xmm3[9],xmm2[9],xmm3[10],xmm2[10],xmm3[11],xmm2[11],xmm3[12],xmm2[12],xmm3[13],xmm2[13],xmm3[14],xmm2[14],xmm3[15],xmm2[15]
1202 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
1216 ; AVX1-NEXT: vextractf128 $1, %ymm1, %xmm2
1218 ; AVX1-NEXT: vpunpckhbw {{.*#+}} xmm2 = xmm3[8],xmm2[8],xmm3[9],xmm2[9],xmm3[10],xmm2[10],xmm3[11],xmm2[11],xmm3[12],xmm2[12],xmm3[13],xmm2[13],xmm3[14],xmm2[14],xmm3[15],xmm2[15]
1220 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
1237 ; AVX1-NEXT: vextractf128 $1, %ymm1, %xmm2
1239 ; AVX1-NEXT: vpunpcklbw {{.*#+}} xmm2 = xmm3[0],xmm2[0],xmm3[1],xmm2[1],xmm3[2],xmm2[2],xmm3[3],xmm2[3],xmm3[4],xmm2[4],xmm3[5],xmm2[5],xmm3[6],xmm2[6],xmm3[7],xmm2[7]
1241 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
1464 ; AVX1-NEXT: vextractf128 $1, %ymm1, %xmm2
1465 ; AVX1-NEXT: vpshufb {{.*#+}} xmm3 = xmm2[u,u,4,u,1,6],zero,zero,xmm2[0],zero,xmm2[11,u],zero,zero,zero,zero
1473 ; AVX1-NEXT: vpshufb {{.*#+}} xmm2 = zero,zero,xmm2[u,u],zero,zero,xmm2[12],zero,xmm2[u,u,u],zero,zero,xmm2[u,0,3]
1475 ; AVX1-NEXT: vpor %xmm2, %xmm1, %xmm1
1476 ; AVX1-NEXT: vpshufb {{.*#+}} xmm2 = xmm4[u,u],zero,zero,xmm4[u,u,u,u,1,6,13,u,u],zero,xmm4[u,u]
1478 ; AVX1-NEXT: vpor %xmm2, %xmm0, %xmm0
1479 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [255,255,0,0,255,255,255,255,0,0,0,255,255,0,255,255]
1480 ; AVX1-NEXT: vpblendvb %xmm2, %xmm1, %xmm0, %xmm0
1505 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,0,0,0,0,0,0,0,8,8,8,8,8,8,8,8]
1506 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
1507 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
1523 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,0,0,0,0,0,0,0,8,8,8,8,8,8,8,8]
1524 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
1526 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
1543 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,0,0,0,0,0,0,0,8,8,8,8,8,8,8,8]
1544 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
1546 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
1563 ; AVX1-NEXT: vmovdqa {{.*#+}} xmm2 = [0,0,0,0,0,0,0,0,8,8,8,8,8,8,8,8]
1564 ; AVX1-NEXT: vpshufb %xmm2, %xmm1, %xmm1
1565 ; AVX1-NEXT: vpshufb %xmm2, %xmm0, %xmm0
1581 ; AVX1-NEXT: vpunpckhbw {{.*#+}} xmm2 = xmm0[8],xmm1[8],xmm0[9],xmm1[9],xmm0[10],xmm1[10],xmm0[11],xmm1[11],xmm0[12],xmm1[12],xmm0[13],xmm1[13],xmm0[14],xmm1[14],xmm0[15],xmm1[15]
1583 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
1588 ; AVX2-NEXT: vpunpckhbw {{.*#+}} xmm2 = xmm0[8],xmm1[8],xmm0[9],xmm1[9],xmm0[10],xmm1[10],xmm0[11],xmm1[11],xmm0[12],xmm1[12],xmm0[13],xmm1[13],xmm0[14],xmm1[14],xmm0[15],xmm1[15]
1590 ; AVX2-NEXT: vinserti128 $1, %xmm2, %ymm0, %ymm0
1788 ; AVX1-NEXT: vextractf128 $1, %ymm1, %xmm2
1790 ; AVX1-NEXT: vpalignr {{.*#+}} xmm2 = xmm2[15],xmm3[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14]
1792 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
1807 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
1808 ; AVX1-NEXT: vpalignr {{.*#+}} xmm1 = xmm1[15],xmm2[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14]
1875 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
1877 ; AVX1-NEXT: vpalignr {{.*#+}} xmm2 = xmm2[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15],xmm3[0]
1879 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
1893 ; AVX1-NEXT: vextractf128 $1, %ymm1, %xmm2
1895 ; AVX1-NEXT: vpalignr {{.*#+}} xmm2 = xmm2[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15],xmm3[0]
1897 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0
1911 ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm2
1913 ; AVX1-NEXT: vpalignr {{.*#+}} xmm2 = xmm2[15],xmm3[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14]
1915 ; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0