Lines Matching full:xmm8
84 : /*TRASH*/"xmm0","xmm7","xmm8","xmm6","xmm9","r14","memory","cc" \
104 "xmm0","xmm8","xmm7","xmm9","r14","rax","memory","cc" \
118 "vpor %%xmm6, %%xmm8, %%xmm7",
119 "vpor (%%rax), %%xmm8, %%xmm7")
122 "vpxor %%xmm6, %%xmm8, %%xmm7",
123 "vpxor (%%rax), %%xmm8, %%xmm7")
126 "vpsubb %%xmm6, %%xmm8, %%xmm7",
127 "vpsubb (%%rax), %%xmm8, %%xmm7")
130 "vpsubd %%xmm6, %%xmm8, %%xmm7",
131 "vpsubd (%%rax), %%xmm8, %%xmm7")
134 "vpaddd %%xmm6, %%xmm8, %%xmm7",
135 "vpaddd (%%rax), %%xmm8, %%xmm7")
138 "vpmovzxwd %%xmm6, %%xmm8",
139 "vpmovzxwd (%%rax), %%xmm8")
142 "vpmovzxbw %%xmm6, %%xmm8",
143 "vpmovzxbw (%%rax), %%xmm8")
146 "vpblendvb %%xmm9, %%xmm6, %%xmm8, %%xmm7",
147 "vpblendvb %%xmm9, (%%rax), %%xmm8, %%xmm7")
150 "vpminsd %%xmm6, %%xmm8, %%xmm7",
151 "vpminsd (%%rax), %%xmm8, %%xmm7")
154 "vpmaxsd %%xmm6, %%xmm8, %%xmm7",
155 "vpmaxsd (%%rax), %%xmm8, %%xmm7")
158 "vandpd %%xmm6, %%xmm8, %%xmm7",
159 "vandpd (%%rax), %%xmm8, %%xmm7")
162 "vcvtsi2sdl %%r14d, %%xmm8, %%xmm7",
163 "vcvtsi2sdl (%%rax), %%xmm8, %%xmm7")
166 "vcvtsi2sdq %%r14, %%xmm8, %%xmm7",
167 "vcvtsi2sdq (%%rax), %%xmm8, %%xmm7")
170 "vcvtsi2ssq %%r14, %%xmm8, %%xmm7",
171 "vcvtsi2ssq (%%rax), %%xmm8, %%xmm7")
174 "vcvttsd2si %%xmm8, %%r14d",
178 "vcvttsd2si %%xmm8, %%r14",
182 "vcvtsd2si %%xmm8, %%r14d",
186 "vcvtsd2si %%xmm8, %%r14",
190 "vpshufb %%xmm6, %%xmm8, %%xmm7",
191 "vpshufb (%%rax), %%xmm8, %%xmm7")
194 "vcmpsd $0, %%xmm6, %%xmm8, %%xmm7",
195 "vcmpsd $0, (%%rax), %%xmm8, %%xmm7")
197 "vcmpsd $1, %%xmm6, %%xmm8, %%xmm7",
198 "vcmpsd $1, (%%rax), %%xmm8, %%xmm7")
200 "vcmpsd $2, %%xmm6, %%xmm8, %%xmm7",
201 "vcmpsd $2, (%%rax), %%xmm8, %%xmm7")
203 "vcmpsd $3, %%xmm6, %%xmm8, %%xmm7",
204 "vcmpsd $3, (%%rax), %%xmm8, %%xmm7")
206 "vcmpsd $4, %%xmm6, %%xmm8, %%xmm7",
207 "vcmpsd $4, (%%rax), %%xmm8, %%xmm7")
209 "vcmpsd $5, %%xmm6, %%xmm8, %%xmm7",
210 "vcmpsd $5, (%%rax), %%xmm8, %%xmm7")
212 "vcmpsd $6, %%xmm6, %%xmm8, %%xmm7",
213 "vcmpsd $6, (%%rax), %%xmm8, %%xmm7")
215 "vcmpsd $7, %%xmm6, %%xmm8, %%xmm7",
216 "vcmpsd $7, (%%rax), %%xmm8, %%xmm7")
218 "vcmpsd $8, %%xmm6, %%xmm8, %%xmm7",
219 "vcmpsd $8, (%%rax), %%xmm8, %%xmm7")
221 "vcmpsd $0xA, %%xmm6, %%xmm8, %%xmm7",
222 "vcmpsd $0xA, (%%rax), %%xmm8, %%xmm7")
224 "vcmpsd $0xC, %%xmm6, %%xmm8, %%xmm7",
225 "vcmpsd $0xC, (%%rax), %%xmm8, %%xmm7")
227 "vcmpsd $0xD, %%xmm6, %%xmm8, %%xmm7",
228 "vcmpsd $0xD, (%%rax), %%xmm8, %%xmm7")
230 "vcmpsd $0xE, %%xmm6, %%xmm8, %%xmm7",
231 "vcmpsd $0xE, (%%rax), %%xmm8, %%xmm7")
233 "vcmpsd $0x11, %%xmm6, %%xmm8, %%xmm7",
234 "vcmpsd $0x11, (%%rax), %%xmm8, %%xmm7")
236 "vcmpsd $0x12, %%xmm6, %%xmm8, %%xmm7",
237 "vcmpsd $0x12, (%%rax), %%xmm8, %%xmm7")
239 "vcmpsd $0x16, %%xmm6, %%xmm8, %%xmm7",
240 "vcmpsd $0x16, (%%rax), %%xmm8, %%xmm7")
242 "vcmpsd $0x1E, %%xmm6, %%xmm8, %%xmm7",
243 "vcmpsd $0x1E, (%%rax), %%xmm8, %%xmm7")
246 "vsqrtsd %%xmm6, %%xmm8, %%xmm7",
247 "vsqrtsd (%%rax), %%xmm8, %%xmm7")
250 "vorps %%xmm6, %%xmm8, %%xmm7",
251 "vorps (%%rax), %%xmm8, %%xmm7")
254 "vandnps %%xmm6, %%xmm8, %%xmm7",
255 "vandnps (%%rax), %%xmm8, %%xmm7")
258 "vmaxss %%xmm6, %%xmm8, %%xmm7",
259 "vmaxss (%%rax), %%xmm8, %%xmm7")
262 "vminss %%xmm6, %%xmm8, %%xmm7",
263 "vminss (%%rax), %%xmm8, %%xmm7")
266 "vandps %%xmm6, %%xmm8, %%xmm7",
267 "vandps (%%rax), %%xmm8, %%xmm7")
270 "vcvtsi2ssl %%r14d, %%xmm8, %%xmm7",
271 "vcvtsi2ssl (%%rax), %%xmm8, %%xmm7")
274 "vunpcklps %%xmm6, %%xmm8, %%xmm7",
275 "vunpcklps (%%rax), %%xmm8, %%xmm7")
278 "vdivss %%xmm6, %%xmm8, %%xmm7",
279 "vdivss (%%rax), %%xmm8, %%xmm7")
282 "vaddss %%xmm6, %%xmm8, %%xmm7",
283 "vaddss (%%rax), %%xmm8, %%xmm7")
286 "vsubss %%xmm6, %%xmm8, %%xmm7",
287 "vsubss (%%rax), %%xmm8, %%xmm7")
290 "vmulss %%xmm6, %%xmm8, %%xmm7",
291 "vmulss (%%rax), %%xmm8, %%xmm7")
294 "vpunpcklbw %%xmm6, %%xmm8, %%xmm7",
295 "vpunpcklbw (%%rax), %%xmm8, %%xmm7")
298 "vpunpckhbw %%xmm6, %%xmm8, %%xmm7",
299 "vpunpckhbw (%%rax), %%xmm8, %%xmm7")
302 "vcvttss2si %%xmm8, %%r14d",
306 "vcvtss2si %%xmm8, %%r14d",
310 "vmovq %%xmm7, %%xmm8",
311 "vmovq (%%rax), %%xmm8")
319 "vmovhlps %%xmm6, %%xmm8, %%xmm7")
322 "vpabsd %%xmm6, %%xmm8",
323 "vpabsd (%%rax), %%xmm8")
327 "vmovlhps %%xmm6, %%xmm8, %%xmm7")
330 "vmovntdq %%xmm8, (%%rax)")
336 "vmovups %%xmm8, %%xmm7",
344 "vpcmpestrm $0x45, %%xmm7, %%xmm8; movapd %%xmm0, %%xmm9",
345 "vpcmpestrm $0x45, (%%rax), %%xmm8; movapd %%xmm0, %%xmm9")
352 "vcvtsd2ss %%xmm9, %%xmm8, %%xmm7",
353 "vcvtsd2ss (%%rax), %%xmm8, %%xmm7")
356 "vcvtss2sd %%xmm9, %%xmm8, %%xmm7",
357 "vcvtss2sd (%%rax), %%xmm8, %%xmm7")
360 "vpackuswb %%xmm9, %%xmm8, %%xmm7",
361 "vpackuswb (%%rax), %%xmm8, %%xmm7")
364 "vcvttss2si %%xmm8, %%r14",
368 "vcvtss2si %%xmm8, %%r14",
372 "vpmovmskb %%xmm8, %%r14")
375 "vpand %%xmm9, %%xmm8, %%xmm7",
376 "vpand (%%rax), %%xmm8, %%xmm7")
379 "vmovhpd %%xmm8, (%%rax)")
382 "vmovhps %%xmm8, (%%rax)")
385 "vpcmpeqb %%xmm9, %%xmm8, %%xmm7",
386 "vpcmpeqb (%%rax), %%xmm8, %%xmm7")
389 "vshufps $0x39, %%xmm9, %%xmm8, %%xmm7",
390 "vshufps $0xC6, (%%rax), %%xmm8, %%xmm7")
393 "vmulps %%xmm9, %%xmm8, %%xmm7",
394 "vmulps (%%rax), %%xmm8, %%xmm7")
397 "vsubps %%xmm9, %%xmm8, %%xmm7",
398 "vsubps (%%rax), %%xmm8, %%xmm7")
401 "vaddps %%xmm9, %%xmm8, %%xmm7",
402 "vaddps (%%rax), %%xmm8, %%xmm7")
405 "vmaxps %%xmm9, %%xmm8, %%xmm7",
406 "vmaxps (%%rax), %%xmm8, %%xmm7")
413 "vmaxpd %%xmm9, %%xmm8, %%xmm7",
414 "vmaxpd (%%rax), %%xmm8, %%xmm7")
421 "vminps %%xmm9, %%xmm8, %%xmm7",
422 "vminps (%%rax), %%xmm8, %%xmm7")
429 "vminpd %%xmm9, %%xmm8, %%xmm7",
430 "vminpd (%%rax), %%xmm8, %%xmm7")
437 "vcvtps2dq %%xmm8, %%xmm7",
438 "vcvtps2dq (%%rax), %%xmm8")
442 "vpshuflw $0xC6, (%%rax), %%xmm8")
446 "vpshufhw $0xC6, (%%rax), %%xmm8")
449 "vpmullw %%xmm9, %%xmm8, %%xmm7",
450 "vpmullw (%%rax), %%xmm8, %%xmm7")
453 "vpaddusw %%xmm9, %%xmm8, %%xmm7",
454 "vpaddusw (%%rax), %%xmm8, %%xmm7")
457 "vpmulhuw %%xmm9, %%xmm8, %%xmm7",
458 "vpmulhuw (%%rax), %%xmm8, %%xmm7")
461 "vpaddusb %%xmm9, %%xmm8, %%xmm7",
462 "vpaddusb (%%rax), %%xmm8, %%xmm7")
465 "vpunpcklwd %%xmm6, %%xmm8, %%xmm7",
466 "vpunpcklwd (%%rax), %%xmm8, %%xmm7")
469 "vpunpckhwd %%xmm6, %%xmm8, %%xmm7",
470 "vpunpckhwd (%%rax), %%xmm8, %%xmm7")
482 "vpsubusb %%xmm9, %%xmm8, %%xmm7",
483 "vpsubusb (%%rax), %%xmm8, %%xmm7")
486 "vpsubsb %%xmm9, %%xmm8, %%xmm7",
487 "vpsubsb (%%rax), %%xmm8, %%xmm7")
496 "vpandn %%xmm9, %%xmm8, %%xmm7",
497 "vpandn (%%rax), %%xmm8, %%xmm7")
504 "vpinsrd $0, %%r14d, %%xmm8, %%xmm7",
505 "vpinsrd $3, (%%rax), %%xmm8, %%xmm7")
508 "vpunpcklqdq %%xmm6, %%xmm8, %%xmm7",
509 "vpunpcklqdq (%%rax), %%xmm8, %%xmm7")
518 "vpaddw %%xmm6, %%xmm8, %%xmm7",
519 "vpaddw (%%rax), %%xmm8, %%xmm7")
522 "vpackssdw %%xmm9, %%xmm8, %%xmm7",
523 "vpackssdw (%%rax), %%xmm8, %%xmm7")
526 "vpunpckldq %%xmm6, %%xmm8, %%xmm7",
527 "vpunpckldq (%%rax), %%xmm8, %%xmm7")
530 "vinsertps $0x39, %%xmm6, %%xmm8, %%xmm7",
531 "vinsertps $0xC6, (%%rax), %%xmm8, %%xmm7")
533 GEN_test_Monly(VMOVSD_M64_XMM, "vmovsd (%%rax), %%xmm8")
535 GEN_test_Monly(VMOVSS_M64_XMM, "vmovss (%%rax), %%xmm8")
537 GEN_test_Monly(VMOVSD_XMM_M64, "vmovsd %%xmm8, (%%rax)")
539 GEN_test_Monly(VMOVSS_XMM_M32, "vmovss %%xmm8, (%%rax)")
546 "vmovapd %%xmm6, %%xmm8",
554 "vmovaps %%xmm6, %%xmm8",
574 "vmovdqu %%xmm6, %%xmm8",
578 "vmovdqa %%xmm6, %%xmm8",
597 GEN_test_Monly(VMOVQ_XMM_MEM64, "vmovq %%xmm8, (%%rax)")
604 "vmovddup %%xmm8, %%xmm7",
608 "vcmpss $0, %%xmm6, %%xmm8, %%xmm7",
609 "vcmpss $0, (%%rax), %%xmm8, %%xmm7")
611 "vcmpss $1, %%xmm6, %%xmm8, %%xmm7",
612 "vcmpss $1, (%%rax), %%xmm8, %%xmm7")
614 "vcmpss $2, %%xmm6, %%xmm8, %%xmm7",
615 "vcmpss $2, (%%rax), %%xmm8, %%xmm7")
617 "vcmpss $3, %%xmm6, %%xmm8, %%xmm7",
618 "vcmpss $3, (%%rax), %%xmm8, %%xmm7")
620 "vcmpss $4, %%xmm6, %%xmm8, %%xmm7",
621 "vcmpss $4, (%%rax), %%xmm8, %%xmm7")
623 "vcmpss $5, %%xmm6, %%xmm8, %%xmm7",
624 "vcmpss $5, (%%rax), %%xmm8, %%xmm7")
626 "vcmpss $6, %%xmm6, %%xmm8, %%xmm7",
627 "vcmpss $6, (%%rax), %%xmm8, %%xmm7")
629 "vcmpss $7, %%xmm6, %%xmm8, %%xmm7",
630 "vcmpss $7, (%%rax), %%xmm8, %%xmm7")
632 "vcmpss $8, %%xmm6, %%xmm8, %%xmm7",
633 "vcmpss $8, (%%rax), %%xmm8, %%xmm7")
635 "vcmpss $0xA, %%xmm6, %%xmm8, %%xmm7",
636 "vcmpss $0xA, (%%rax), %%xmm8, %%xmm7")
638 "vcmpss $0xC, %%xmm6, %%xmm8, %%xmm7",
639 "vcmpss $0xC, (%%rax), %%xmm8, %%xmm7")
641 "vcmpss $0xD, %%xmm6, %%xmm8, %%xmm7",
642 "vcmpss $0xD, (%%rax), %%xmm8, %%xmm7")
644 "vcmpss $0xE, %%xmm6, %%xmm8, %%xmm7",
645 "vcmpss $0xE, (%%rax), %%xmm8, %%xmm7")
647 "vcmpss $0x11, %%xmm6, %%xmm8, %%xmm7",
648 "vcmpss $0x11, (%%rax), %%xmm8, %%xmm7")
650 "vcmpss $0x12, %%xmm6, %%xmm8, %%xmm7",
651 "vcmpss $0x12, (%%rax), %%xmm8, %%xmm7")
653 "vcmpss $0x16, %%xmm6, %%xmm8, %%xmm7",
654 "vcmpss $0x16, (%%rax), %%xmm8, %%xmm7")
656 "vcmpss $0x1E, %%xmm6, %%xmm8, %%xmm7",
657 "vcmpss $0x1E, (%%rax), %%xmm8, %%xmm7")
661 "vcvtpd2psx %%xmm8, %%xmm7",
689 "vpcmpeqd %%xmm6, %%xmm8, %%xmm7",
690 "vpcmpeqd (%%rax), %%xmm8, %%xmm7")
693 "vpshufd $0x39, %%xmm9, %%xmm8",
697 "vmaxsd %%xmm6, %%xmm8, %%xmm7",
698 "vmaxsd (%%rax), %%xmm8, %%xmm7")
701 "vdivsd %%xmm6, %%xmm8, %%xmm7",
702 "vdivsd (%%rax), %%xmm8, %%xmm7")
705 "vminsd %%xmm6, %%xmm8, %%xmm7",
706 "vminsd (%%rax), %%xmm8, %%xmm7")
709 "vsubsd %%xmm6, %%xmm8, %%xmm7",
710 "vsubsd (%%rax), %%xmm8, %%xmm7")
713 "vaddsd %%xmm6, %%xmm8, %%xmm7",
714 "vaddsd (%%rax), %%xmm8, %%xmm7")
717 "vmulsd %%xmm6, %%xmm8, %%xmm7",
718 "vmulsd (%%rax), %%xmm8, %%xmm7")
721 "vxorps %%xmm6, %%xmm8, %%xmm7",
722 "vxorps (%%rax), %%xmm8, %%xmm7")
725 "vxorpd %%xmm6, %%xmm8, %%xmm7",
726 "vxorpd (%%rax), %%xmm8, %%xmm7")
729 "vorpd %%xmm6, %%xmm8, %%xmm7",
730 "vorpd (%%rax), %%xmm8, %%xmm7")
733 "vandnpd %%xmm6, %%xmm8, %%xmm7",
734 "vandnpd (%%rax), %%xmm8, %%xmm7")
737 "vcvtps2pd %%xmm6, %%xmm8",
738 "vcvtps2pd (%%rax), %%xmm8")
741 "vucomisd %%xmm6, %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14",
742 "vucomisd (%%rax), %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14")
745 "vucomiss %%xmm6, %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14",
746 "vucomiss (%%rax), %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14")
749 "vpinsrq $0, %%r14, %%xmm8, %%xmm7",
750 "vpinsrq $1, (%%rax), %%xmm8, %%xmm7")
753 "vpaddq %%xmm6, %%xmm8, %%xmm7",
754 "vpaddq (%%rax), %%xmm8, %%xmm7")
757 "vpsubq %%xmm6, %%xmm8, %%xmm7",
758 "vpsubq (%%rax), %%xmm8, %%xmm7")
761 "vpsubw %%xmm6, %%xmm8, %%xmm7",
762 "vpsubw (%%rax), %%xmm8, %%xmm7")
793 "vpcmpeqq %%xmm6, %%xmm8, %%xmm7",
794 "vpcmpeqq (%%rax), %%xmm8, %%xmm7")
797 "vsubpd %%xmm6, %%xmm8, %%xmm7",
798 "vsubpd (%%rax), %%xmm8, %%xmm7")
801 "vaddpd %%xmm6, %%xmm8, %%xmm7",
802 "vaddpd (%%rax), %%xmm8, %%xmm7")
805 "vunpcklpd %%xmm6, %%xmm8, %%xmm7",
806 "vunpcklpd (%%rax), %%xmm8, %%xmm7")
809 "vunpckhpd %%xmm6, %%xmm8, %%xmm7",
810 "vunpckhpd (%%rax), %%xmm8, %%xmm7")
813 "vunpckhps %%xmm6, %%xmm8, %%xmm7",
814 "vunpckhps (%%rax), %%xmm8, %%xmm7")
817 "vmovups %%xmm6, %%xmm8",
837 "vpcmpgtq %%xmm6, %%xmm8, %%xmm7",
838 "vpcmpgtq (%%rax), %%xmm8, %%xmm7")
852 "vpmuludq %%xmm6, %%xmm8, %%xmm7",
853 "vpmuludq (%%rax), %%xmm8, %%xmm7")
856 "vpmuldq %%xmm6, %%xmm8, %%xmm7",
857 "vpmuldq (%%rax), %%xmm8, %%xmm7")
863 "vpmaxud %%xmm6, %%xmm8, %%xmm7",
864 "vpmaxud (%%rax), %%xmm8, %%xmm7")
867 "vpminud %%xmm6, %%xmm8, %%xmm7",
868 "vpminud (%%rax), %%xmm8, %%xmm7")
871 "vpmulld %%xmm6, %%xmm8, %%xmm7",
872 "vpmulld (%%rax), %%xmm8, %%xmm7")
875 "vpmaxuw %%xmm6, %%xmm8, %%xmm7",
876 "vpmaxuw (%%rax), %%xmm8, %%xmm7")
885 "vpminuw %%xmm6, %%xmm8, %%xmm7",
886 "vpminuw (%%rax), %%xmm8, %%xmm7")
889 "vphminposuw %%xmm6, %%xmm8",
893 "vpmaxsw %%xmm6, %%xmm8, %%xmm7",
894 "vpmaxsw (%%rax), %%xmm8, %%xmm7")
897 "vpminsw %%xmm6, %%xmm8, %%xmm7",
898 "vpminsw (%%rax), %%xmm8, %%xmm7")
901 "vpmaxub %%xmm6, %%xmm8, %%xmm7",
902 "vpmaxub (%%rax), %%xmm8, %%xmm7")
905 "vpextrb $0x0, %%xmm8, %%r14",
906 "vpextrb $0x0, %%xmm8, (%%rax)")
909 "vpextrb $0x1, %%xmm8, %%r14",
910 "vpextrb $0x1, %%xmm8, (%%rax)")
913 "vpextrb $0x2, %%xmm8, %%r14",
914 "vpextrb $0x2, %%xmm8, (%%rax)")
917 "vpextrb $0x3, %%xmm8, %%r14",
918 "vpextrb $0x3, %%xmm8, (%%rax)")
921 "vpextrb $0x4, %%xmm8, %%r14",
922 "vpextrb $0x4, %%xmm8, (%%rax)")
925 "vpextrb $0x9, %%xmm8, %%r14",
926 "vpextrb $0x9, %%xmm8, (%%rax)")
929 "vpextrb $0xE, %%xmm8, %%r14",
930 "vpextrb $0xE, %%xmm8, (%%rax)")
933 "vpextrb $0xF, %%xmm8, %%r14",
934 "vpextrb $0xF, %%xmm8, (%%rax)")
937 "vpminub %%xmm6, %%xmm8, %%xmm7",
938 "vpminub (%%rax), %%xmm8, %%xmm7")
941 "vpmaxsb %%xmm6, %%xmm8, %%xmm7",
942 "vpmaxsb (%%rax), %%xmm8, %%xmm7")
945 "vpminsb %%xmm6, %%xmm8, %%xmm7",
946 "vpminsb (%%rax), %%xmm8, %%xmm7")
987 "vpermilpd $0x0, %%xmm6, %%xmm8",
988 "vpermilpd $0x1, (%%rax), %%xmm8")
990 "vpermilpd $0x3, %%xmm6, %%xmm8",
991 "vpermilpd $0x2, (%%rax), %%xmm8")
1021 "vcmppd $4, %%xmm6, %%xmm8, %%xmm7",
1022 "vcmppd $4, (%%rax), %%xmm8, %%xmm7")
1029 "vcmpps $4, %%xmm6, %%xmm8, %%xmm7",
1030 "vcmpps $4, (%%rax), %%xmm8, %%xmm7")
1037 "vcvtdq2pd %%xmm6, %%xmm8",
1038 "vcvtdq2pd (%%rax), %%xmm8")
1041 "vdivpd %%xmm6, %%xmm8, %%xmm7",
1042 "vdivpd (%%rax), %%xmm8, %%xmm7")
1049 "vpmovsxbw %%xmm6, %%xmm8",
1050 "vpmovsxbw (%%rax), %%xmm8")
1053 "vpsubusw %%xmm9, %%xmm8, %%xmm7",
1054 "vpsubusw (%%rax), %%xmm8, %%xmm7")
1057 "vpsubsw %%xmm9, %%xmm8, %%xmm7",
1058 "vpsubsw (%%rax), %%xmm8, %%xmm7")
1061 "vpcmpeqw %%xmm6, %%xmm8, %%xmm7",
1062 "vpcmpeqw (%%rax), %%xmm8, %%xmm7")
1065 "vpaddb %%xmm6, %%xmm8, %%xmm7",
1066 "vpaddb (%%rax), %%xmm8, %%xmm7")
1077 "vmovhpd (%%rax), %%xmm8, %%xmm7")
1080 "vmovhps (%%rax), %%xmm8, %%xmm7")
1088 "vpunpckhdq %%xmm6, %%xmm8, %%xmm7",
1089 "vpunpckhdq (%%rax), %%xmm8, %%xmm7")
1092 "vbroadcastss (%%rax), %%xmm8")
1095 "vpmovsxdq %%xmm6, %%xmm8",
1096 "vpmovsxdq (%%rax), %%xmm8")
1099 "vpmovsxwd %%xmm6, %%xmm8",
1100 "vpmovsxwd (%%rax), %%xmm8")
1103 "vdivps %%xmm9, %%xmm8, %%xmm7",
1104 "vdivps (%%rax), %%xmm8, %%xmm7")
1144 "vpmulhw %%xmm9, %%xmm8, %%xmm7",
1145 "vpmulhw (%%rax), %%xmm8, %%xmm7")
1148 "vpunpckhqdq %%xmm6, %%xmm8, %%xmm7",
1149 "vpunpckhqdq (%%rax), %%xmm8, %%xmm7")
1155 "vpcmpgtb %%xmm6, %%xmm8, %%xmm7",
1156 "vpcmpgtb (%%rax), %%xmm8, %%xmm7")
1159 "vpcmpgtw %%xmm6, %%xmm8, %%xmm7",
1160 "vpcmpgtw (%%rax), %%xmm8, %%xmm7")
1163 "vpcmpgtd %%xmm6, %%xmm8, %%xmm7",
1164 "vpcmpgtd (%%rax), %%xmm8, %%xmm7")
1167 "vpmovzxbd %%xmm6, %%xmm8",
1168 "vpmovzxbd (%%rax), %%xmm8")
1171 "vpmovsxbd %%xmm6, %%xmm8",
1172 "vpmovsxbd (%%rax), %%xmm8")
1175 "vpinsrb $0, %%r14d, %%xmm8, %%xmm7",
1176 "vpinsrb $3, (%%rax), %%xmm8, %%xmm7")
1178 "vpinsrb $6, %%r14d, %%xmm8, %%xmm7",
1179 "vpinsrb $9, (%%rax), %%xmm8, %%xmm7")
1181 "vpinsrb $12, %%r14d, %%xmm8, %%xmm7",
1182 "vpinsrb $15, (%%rax), %%xmm8, %%xmm7")
1185 "vpinsrw $0, %%r14d, %%xmm8, %%xmm7",
1186 "vpinsrw $3, (%%rax), %%xmm8, %%xmm7")
1188 "vpinsrw $2, %%r14d, %%xmm8, %%xmm7",
1189 "vpinsrw $3, (%%rax), %%xmm8, %%xmm7")
1191 "vpinsrw $4, %%r14d, %%xmm8, %%xmm7",
1192 "vpinsrw $5, (%%rax), %%xmm8, %%xmm7")
1194 "vpinsrw $6, %%r14d, %%xmm8, %%xmm7",
1195 "vpinsrw $7, (%%rax), %%xmm8, %%xmm7")
1198 "vcomisd %%xmm6, %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14",
1199 "vcomisd (%%rax), %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14")
1202 "vcomiss %%xmm6, %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14",
1203 "vcomiss (%%rax), %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14")
1210 "vdppd $0x00, %%xmm6, %%xmm8, %%xmm7",
1213 "vdppd $0x5A, %%xmm6, %%xmm8, %%xmm7",
1216 "vdppd $0x0F, %%xmm6, %%xmm8, %%xmm7",
1219 "vdppd $0xF0, %%xmm6, %%xmm8, %%xmm7",
1223 "vdpps $0x00, %%xmm6, %%xmm8, %%xmm7",
1226 "vdpps $0x5A, %%xmm6, %%xmm8, %%xmm7",
1229 "vdpps $0x0F, %%xmm6, %%xmm8, %%xmm7",
1232 "vdpps $0xF0, %%xmm6, %%xmm8, %%xmm7",
1252 "vpalignr $0, %%xmm6, %%xmm8, %%xmm7",
1253 "vpalignr $3, (%%rax), %%xmm8, %%xmm7")
1255 "vpalignr $6, %%xmm6, %%xmm8, %%xmm7",
1256 "vpalignr $9, (%%rax), %%xmm8, %%xmm7")
1258 "vpalignr $12, %%xmm6, %%xmm8, %%xmm7",
1259 "vpalignr $15, (%%rax), %%xmm8, %%xmm7")
1261 GEN_test_Ronly(VMOVSD_REG_XMM, "vmovsd %%xmm9, %%xmm7, %%xmm8")
1263 GEN_test_Ronly(VMOVSS_REG_XMM, "vmovss %%xmm9, %%xmm7, %%xmm8")
1265 GEN_test_Monly(VMOVLPD_128_M64_XMM_XMM, "vmovlpd (%%rax), %%xmm8, %%xmm7")
1270 "vshufpd $0, %%xmm9, %%xmm8, %%xmm7",
1271 "vshufpd $1, (%%rax), %%xmm8, %%xmm7")
1273 "vshufpd $2, %%xmm9, %%xmm8, %%xmm7",
1274 "vshufpd $3, (%%rax), %%xmm8, %%xmm7")
1284 "vpermilps $0x00, %%xmm6, %%xmm8",
1285 "vpermilps $0x01, (%%rax), %%xmm8")
1287 "vpermilps $0xFE, %%xmm6, %%xmm8",
1288 "vpermilps $0xFF, (%%rax), %%xmm8")
1290 "vpermilps $0x30, %%xmm6, %%xmm8",
1291 "vpermilps $0x03, (%%rax), %%xmm8")
1293 "vpermilps $0x21, %%xmm6, %%xmm8",
1294 "vpermilps $0x12, (%%rax), %%xmm8")
1296 "vpermilps $0xD7, %%xmm6, %%xmm8",
1297 "vpermilps $0x6C, (%%rax), %%xmm8")
1299 "vpermilps $0xB5, %%xmm6, %%xmm8",
1300 "vpermilps $0x4A, (%%rax), %%xmm8")
1302 "vpermilps $0x85, %%xmm6, %%xmm8",
1303 "vpermilps $0xDC, (%%rax), %%xmm8")
1305 "vpermilps $0x29, %%xmm6, %%xmm8",
1306 "vpermilps $0x92, (%%rax), %%xmm8")
1309 "vblendps $0, %%xmm6, %%xmm8, %%xmm7",
1310 "vblendps $3, (%%rax), %%xmm8, %%xmm7")
1312 "vblendps $6, %%xmm6, %%xmm8, %%xmm7",
1313 "vblendps $9, (%%rax), %%xmm8, %%xmm7")
1315 "vblendps $12, %%xmm6, %%xmm8, %%xmm7",
1316 "vblendps $15, (%%rax), %%xmm8, %%xmm7")
1319 "vblendpd $0, %%xmm6, %%xmm8, %%xmm7",
1320 "vblendpd $1, (%%rax), %%xmm8, %%xmm7")
1322 "vblendpd $2, %%xmm6, %%xmm8, %%xmm7",
1323 "vblendpd $3, (%%rax), %%xmm8, %%xmm7")
1336 "vpblendw $0x00, %%xmm6, %%xmm8, %%xmm7",
1337 "vpblendw $0x01, (%%rax), %%xmm8, %%xmm7")
1339 "vpblendw $0xFE, %%xmm6, %%xmm8, %%xmm7",
1340 "vpblendw $0xFF, (%%rax), %%xmm8, %%xmm7")
1342 "vpblendw $0x30, %%xmm6, %%xmm8, %%xmm7",
1343 "vpblendw $0x03, (%%rax), %%xmm8, %%xmm7")
1345 "vpblendw $0x21, %%xmm6, %%xmm8, %%xmm7",
1346 "vpblendw $0x12, (%%rax), %%xmm8, %%xmm7")
1348 "vpblendw $0xD7, %%xmm6, %%xmm8, %%xmm7",
1349 "vpblendw $0x6C, (%%rax), %%xmm8, %%xmm7")
1351 "vpblendw $0xB5, %%xmm6, %%xmm8, %%xmm7",
1352 "vpblendw $0x4A, (%%rax), %%xmm8, %%xmm7")
1354 "vpblendw $0x85, %%xmm6, %%xmm8, %%xmm7",
1355 "vpblendw $0xDC, (%%rax), %%xmm8, %%xmm7")
1357 "vpblendw $0x29, %%xmm6, %%xmm8, %%xmm7",
1358 "vpblendw $0x92, (%%rax), %%xmm8, %%xmm7")
1365 "vsqrtss %%xmm6, %%xmm8, %%xmm7",
1366 "vsqrtss (%%rax), %%xmm8, %%xmm7")
1369 "vsqrtps %%xmm6, %%xmm8",
1370 "vsqrtps (%%rax), %%xmm8")
1377 "vsqrtpd %%xmm6, %%xmm8",
1378 "vsqrtpd (%%rax), %%xmm8")
1385 "vrsqrtss %%xmm6, %%xmm8, %%xmm7",
1386 "vrsqrtss (%%rax), %%xmm8, %%xmm7")
1389 "vrsqrtps %%xmm6, %%xmm8",
1390 "vrsqrtps (%%rax), %%xmm8")
1453 "vpermilps %%xmm6, %%xmm8, %%xmm7",
1454 "vpermilps (%%rax), %%xmm8, %%xmm7")
1457 "vpermilpd %%xmm6, %%xmm8, %%xmm7",
1458 "vpermilpd (%%rax), %%xmm8, %%xmm7")
1471 "vpsllw %%xmm6, %%xmm8, %%xmm9",
1473 "vpsllw 128(%%rax), %%xmm8, %%xmm9")
1478 "vpsrlw %%xmm6, %%xmm8, %%xmm9",
1480 "vpsrlw 128(%%rax), %%xmm8, %%xmm9")
1485 "vpsraw %%xmm6, %%xmm8, %%xmm9",
1487 "vpsraw 128(%%rax), %%xmm8, %%xmm9")
1492 "vpslld %%xmm6, %%xmm8, %%xmm9",
1494 "vpslld 128(%%rax), %%xmm8, %%xmm9")
1499 "vpsrld %%xmm6, %%xmm8, %%xmm9",
1501 "vpsrld 128(%%rax), %%xmm8, %%xmm9")
1506 "vpsrad %%xmm6, %%xmm8, %%xmm9",
1508 "vpsrad 128(%%rax), %%xmm8, %%xmm9")
1513 "vpsllq %%xmm6, %%xmm8, %%xmm9",
1515 "vpsllq 128(%%rax), %%xmm8, %%xmm9")
1520 "vpsrlq %%xmm6, %%xmm8, %%xmm9",
1522 "vpsrlq 128(%%rax), %%xmm8, %%xmm9")
1525 "vroundps $0x0, %%xmm8, %%xmm9",
1528 "vroundps $0x1, %%xmm8, %%xmm9",
1531 "vroundps $0x2, %%xmm8, %%xmm9",
1534 "vroundps $0x3, %%xmm8, %%xmm9",
1537 "vroundps $0x4, %%xmm8, %%xmm9",
1557 "vroundpd $0x0, %%xmm8, %%xmm9",
1560 "vroundpd $0x1, %%xmm8, %%xmm9",
1563 "vroundpd $0x2, %%xmm8, %%xmm9",
1566 "vroundpd $0x3, %%xmm8, %%xmm9",
1569 "vroundpd $0x4, %%xmm8, %%xmm9",
1589 "vpmaddwd %%xmm6, %%xmm8, %%xmm7",
1590 "vpmaddwd (%%rax), %%xmm8, %%xmm7")
1593 "vaddsubps %%xmm6, %%xmm8, %%xmm7",
1594 "vaddsubps (%%rax), %%xmm8, %%xmm7")
1601 "vaddsubpd %%xmm6, %%xmm8, %%xmm7",
1602 "vaddsubpd (%%rax), %%xmm8, %%xmm7")
1609 "vroundss $0x0, %%xmm8, %%xmm6, %%xmm9",
1612 "vroundss $0x1, %%xmm8, %%xmm6, %%xmm9",
1615 "vroundss $0x2, %%xmm8, %%xmm6, %%xmm9",
1618 "vroundss $0x3, %%xmm8, %%xmm6, %%xmm9",
1621 "vroundss $0x4, %%xmm8, %%xmm6, %%xmm9",
1624 "vroundss $0x5, %%xmm8, %%xmm6, %%xmm9",
1628 "vroundsd $0x0, %%xmm8, %%xmm6, %%xmm9",
1631 "vroundsd $0x1, %%xmm8, %%xmm6, %%xmm9",
1634 "vroundsd $0x2, %%xmm8, %%xmm6, %%xmm9",
1637 "vroundsd $0x3, %%xmm8, %%xmm6, %%xmm9",
1640 "vroundsd $0x4, %%xmm8, %%xmm6, %%xmm9",
1643 "vroundsd $0x5, %%xmm8, %%xmm6, %%xmm9",
1647 "vptest %%xmm6, %%xmm8; "
1649 "vptest (%%rax), %%xmm8; "
1656 "vmovups %%xmm6, %%xmm8;"
1657 "vptest %%xmm6, %%xmm8; "
1659 "vmovups (%%rax), %%xmm8;"
1660 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1661 "vxorpd %%xmm8,%%xmm7,%%xmm8;"
1662 "vptest (%%rax), %%xmm8; "
1679 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1694 "vtestps %%xmm6, %%xmm8; "
1696 "vtestps (%%rax), %%xmm8; "
1703 "vmovups %%xmm6, %%xmm8;"
1704 "vtestps %%xmm6, %%xmm8; "
1706 "vmovups (%%rax), %%xmm8;"
1707 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1708 "vxorpd %%xmm8,%%xmm7,%%xmm8;"
1709 "vtestps (%%rax), %%xmm8; "
1713 "vtestps %%xmm8, %%xmm9; "
1735 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1754 "vtestpd %%xmm6, %%xmm8; "
1756 "vtestpd (%%rax), %%xmm8; "
1763 "vmovups %%xmm6, %%xmm8;"
1764 "vtestpd %%xmm6, %%xmm8; "
1766 "vmovups (%%rax), %%xmm8;"
1767 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1768 "vxorpd %%xmm8,%%xmm7,%%xmm8;"
1769 "vtestpd (%%rax), %%xmm8; "
1773 "vtestpd %%xmm8, %%xmm9; "
1795 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1812 "vblendvps %%xmm9, %%xmm6, %%xmm8, %%xmm7",
1813 "vblendvps %%xmm9, (%%rax), %%xmm8, %%xmm7")
1820 "vblendvpd %%xmm9, %%xmm6, %%xmm8, %%xmm7",
1821 "vblendvpd %%xmm9, (%%rax), %%xmm8, %%xmm7")
1829 "vhaddps %%xmm6, %%xmm8, %%xmm7",
1830 "vhaddps (%%rax), %%xmm8, %%xmm7")
1837 "vhaddpd %%xmm6, %%xmm8, %%xmm7",
1838 "vhaddpd (%%rax), %%xmm8, %%xmm7")
1845 "vhsubps %%xmm6, %%xmm8, %%xmm7",
1846 "vhsubps (%%rax), %%xmm8, %%xmm7")
1853 "vhsubpd %%xmm6, %%xmm8, %%xmm7",
1854 "vhsubpd (%%rax), %%xmm8, %%xmm7")
1861 "vextractps $0, %%xmm8, %%r14d",
1862 "vextractps $0, %%xmm8, (%%rax)")
1865 "vextractps $1, %%xmm8, %%r14d",
1866 "vextractps $1, %%xmm8, (%%rax)")
1869 "vextractps $2, %%xmm8, %%r14d",
1870 "vextractps $2, %%xmm8, (%%rax)")
1873 "vextractps $3, %%xmm8, %%r14d",
1874 "vextractps $3, %%xmm8, (%%rax)")
1877 "vlddqu 1(%%rax), %%xmm8")
1887 "vmaskmovdqu %%xmm8, %%xmm9;"
1915 "vpacksswb %%xmm6, %%xmm8, %%xmm7",
1916 "vpacksswb (%%rax), %%xmm8, %%xmm7")
1919 "vpavgb %%xmm6, %%xmm8, %%xmm7",
1920 "vpavgb (%%rax), %%xmm8, %%xmm7")
1923 "vpavgw %%xmm6, %%xmm8, %%xmm7",
1924 "vpavgw (%%rax), %%xmm8, %%xmm7")
1927 "vpaddsb %%xmm6, %%xmm8, %%xmm7",
1928 "vpaddsb (%%rax), %%xmm8, %%xmm7")
1931 "vpaddsw %%xmm6, %%xmm8, %%xmm7",
1932 "vpaddsw (%%rax), %%xmm8, %%xmm7")
1935 "vphaddw %%xmm6, %%xmm8, %%xmm7",
1936 "vphaddw (%%rax), %%xmm8, %%xmm7")
1939 "vphaddd %%xmm6, %%xmm8, %%xmm7",
1940 "vphaddd (%%rax), %%xmm8, %%xmm7")
1943 "vphaddsw %%xmm6, %%xmm8, %%xmm7",
1944 "vphaddsw (%%rax), %%xmm8, %%xmm7")
1947 "vpmaddubsw %%xmm6, %%xmm8, %%xmm7",
1948 "vpmaddubsw (%%rax), %%xmm8, %%xmm7")
1951 "vphsubw %%xmm6, %%xmm8, %%xmm7",
1952 "vphsubw (%%rax), %%xmm8, %%xmm7")
1955 "vphsubd %%xmm6, %%xmm8, %%xmm7",
1956 "vphsubd (%%rax), %%xmm8, %%xmm7")
1959 "vphsubsw %%xmm6, %%xmm8, %%xmm7",
1960 "vphsubsw (%%rax), %%xmm8, %%xmm7")
1971 "vpmovsxbq %%xmm6, %%xmm8",
1972 "vpmovsxbq (%%rax), %%xmm8")
1975 "vpmovsxwq %%xmm6, %%xmm8",
1976 "vpmovsxwq (%%rax), %%xmm8")
1979 "vpackusdw %%xmm6, %%xmm8, %%xmm7",
1980 "vpackusdw (%%rax), %%xmm8, %%xmm7")
1983 "vpmovzxbq %%xmm6, %%xmm8",
1984 "vpmovzxbq (%%rax), %%xmm8")
1987 "vpmovzxwq %%xmm6, %%xmm8",
1988 "vpmovzxwq (%%rax), %%xmm8")
1991 "vpmovzxdq %%xmm6, %%xmm8",
1992 "vpmovzxdq (%%rax), %%xmm8")
1995 "vmpsadbw $0, %%xmm6, %%xmm8, %%xmm7",
1996 "vmpsadbw $0, (%%rax), %%xmm8, %%xmm7")
1998 "vmpsadbw $1, %%xmm6, %%xmm8, %%xmm7",
1999 "vmpsadbw $1, (%%rax), %%xmm8, %%xmm7")
2001 "vmpsadbw $2, %%xmm6, %%xmm8, %%xmm7",
2002 "vmpsadbw $2, (%%rax), %%xmm8, %%xmm7")
2004 "vmpsadbw $3, %%xmm6, %%xmm8, %%xmm7",
2005 "vmpsadbw $3, (%%rax), %%xmm8, %%xmm7")
2007 "vmpsadbw $4, %%xmm6, %%xmm8, %%xmm7",
2008 "vmpsadbw $4, (%%rax), %%xmm8, %%xmm7")
2010 "vmpsadbw $5, %%xmm6, %%xmm8, %%xmm7",
2011 "vmpsadbw $5, (%%rax), %%xmm8, %%xmm7")
2013 "vmpsadbw $6, %%xmm6, %%xmm8, %%xmm7",
2014 "vmpsadbw $6, (%%rax), %%xmm8, %%xmm7")
2016 "vmpsadbw $7, %%xmm6, %%xmm8, %%xmm7",
2017 "vmpsadbw $7, (%%rax), %%xmm8, %%xmm7")
2023 GEN_test_Monly(VMOVLPS_128_M64_XMM_XMM, "vmovlps (%%rax), %%xmm8, %%xmm7")
2028 "vrcpss %%xmm6, %%xmm8, %%xmm7",
2029 "vrcpss (%%rax), %%xmm8, %%xmm7")
2032 "vrcpps %%xmm6, %%xmm8",
2033 "vrcpps (%%rax), %%xmm8")
2040 "vpsadbw %%xmm6, %%xmm8, %%xmm7",
2041 "vpsadbw (%%rax), %%xmm8, %%xmm7")
2044 "vpsignb %%xmm6, %%xmm8, %%xmm7",
2045 "vpsignb (%%rax), %%xmm8, %%xmm7")
2048 "vpsignw %%xmm6, %%xmm8, %%xmm7",
2049 "vpsignw (%%rax), %%xmm8, %%xmm7")
2052 "vpsignd %%xmm6, %%xmm8, %%xmm7",
2053 "vpsignd (%%rax), %%xmm8, %%xmm7")
2056 "vpmulhrsw %%xmm6, %%xmm8, %%xmm7",
2057 "vpmulhrsw (%%rax), %%xmm8, %%xmm7")
2088 "vaesenc %%xmm6, %%xmm8, %%xmm7",
2089 "vaesenc (%%rax), %%xmm8, %%xmm7")
2092 "vaesenclast %%xmm6, %%xmm8, %%xmm7",
2093 "vaesenclast (%%rax), %%xmm8, %%xmm7")
2096 "vaesdec %%xmm6, %%xmm8, %%xmm7",
2097 "vaesdec (%%rax), %%xmm8, %%xmm7")
2100 "vaesdeclast %%xmm6, %%xmm8, %%xmm7",
2101 "vaesdeclast (%%rax), %%xmm8, %%xmm7")
2121 "vpclmulqdq $0x00, %%xmm6, %%xmm8, %%xmm7",
2122 "vpclmulqdq $0x00, (%%rax), %%xmm8, %%xmm7")
2124 "vpclmulqdq $0x01, %%xmm6, %%xmm8, %%xmm7",
2125 "vpclmulqdq $0x01, (%%rax), %%xmm8, %%xmm7")
2127 "vpclmulqdq $0x10, %%xmm6, %%xmm8, %%xmm7",
2128 "vpclmulqdq $0x10, (%%rax), %%xmm8, %%xmm7")
2130 "vpclmulqdq $0x11, %%xmm6, %%xmm8, %%xmm7",
2131 "vpclmulqdq $0x11, (%%rax), %%xmm8, %%xmm7")
2133 "vpclmulqdq $0xFF, %%xmm6, %%xmm8, %%xmm7",
2134 "vpclmulqdq $0xFF, (%%rax), %%xmm8, %%xmm7")
2137 "vcmpss $0x9, %%xmm6, %%xmm8, %%xmm7",
2138 "vcmpss $0x9, (%%rax), %%xmm8, %%xmm7")
2141 "vmaskmovps (%%rax), %%xmm8, %%xmm7;"
2151 "vmaskmovpd (%%rax), %%xmm8, %%xmm7;"
2161 "vmaskmovps %%xmm8, %%xmm7, (%%rax);"
2171 "vmaskmovpd %%xmm8, %%xmm7, (%%rax);"