Home | History | Annotate | Download | only in amd64

Lines Matching full:xmm7

83           : /*TRASH*/"xmm0","xmm7","xmm8","xmm6","xmm9","r14","memory","cc" \
102 : /*TRASH*/"xmm0","xmm8","xmm7","xmm9","r14","rax","memory","cc" \
116 "vpor %%xmm6, %%xmm8, %%xmm7",
117 "vpor (%%rax), %%xmm8, %%xmm7")
120 "vpxor %%xmm6, %%xmm8, %%xmm7",
121 "vpxor (%%rax), %%xmm8, %%xmm7")
124 "vpsubb %%xmm6, %%xmm8, %%xmm7",
125 "vpsubb (%%rax), %%xmm8, %%xmm7")
128 "vpsubd %%xmm6, %%xmm8, %%xmm7",
129 "vpsubd (%%rax), %%xmm8, %%xmm7")
132 "vpaddd %%xmm6, %%xmm8, %%xmm7",
133 "vpaddd (%%rax), %%xmm8, %%xmm7")
144 "vpblendvb %%xmm9, %%xmm6, %%xmm8, %%xmm7",
145 "vpblendvb %%xmm9, (%%rax), %%xmm8, %%xmm7")
148 "vpminsd %%xmm6, %%xmm8, %%xmm7",
149 "vpminsd (%%rax), %%xmm8, %%xmm7")
152 "vpmaxsd %%xmm6, %%xmm8, %%xmm7",
153 "vpmaxsd (%%rax), %%xmm8, %%xmm7")
156 "vandpd %%xmm6, %%xmm8, %%xmm7",
157 "vandpd (%%rax), %%xmm8, %%xmm7")
160 "vcvtsi2sdl %%r14d, %%xmm8, %%xmm7",
161 "vcvtsi2sdl (%%rax), %%xmm8, %%xmm7")
164 "vcvtsi2sdq %%r14, %%xmm8, %%xmm7",
165 "vcvtsi2sdq (%%rax), %%xmm8, %%xmm7")
168 "vcvtsi2ssq %%r14, %%xmm8, %%xmm7",
169 "vcvtsi2ssq (%%rax), %%xmm8, %%xmm7")
188 "vpshufb %%xmm6, %%xmm8, %%xmm7",
189 "vpshufb (%%rax), %%xmm8, %%xmm7")
192 "vcmpsd $0, %%xmm6, %%xmm8, %%xmm7",
193 "vcmpsd $0, (%%rax), %%xmm8, %%xmm7")
195 "vcmpsd $1, %%xmm6, %%xmm8, %%xmm7",
196 "vcmpsd $1, (%%rax), %%xmm8, %%xmm7")
198 "vcmpsd $2, %%xmm6, %%xmm8, %%xmm7",
199 "vcmpsd $2, (%%rax), %%xmm8, %%xmm7")
201 "vcmpsd $3, %%xmm6, %%xmm8, %%xmm7",
202 "vcmpsd $3, (%%rax), %%xmm8, %%xmm7")
204 "vcmpsd $4, %%xmm6, %%xmm8, %%xmm7",
205 "vcmpsd $4, (%%rax), %%xmm8, %%xmm7")
207 "vcmpsd $5, %%xmm6, %%xmm8, %%xmm7",
208 "vcmpsd $5, (%%rax), %%xmm8, %%xmm7")
210 "vcmpsd $6, %%xmm6, %%xmm8, %%xmm7",
211 "vcmpsd $6, (%%rax), %%xmm8, %%xmm7")
213 "vcmpsd $7, %%xmm6, %%xmm8, %%xmm7",
214 "vcmpsd $7, (%%rax), %%xmm8, %%xmm7")
216 "vcmpsd $0xA, %%xmm6, %%xmm8, %%xmm7",
217 "vcmpsd $0xA, (%%rax), %%xmm8, %%xmm7")
219 "vcmpsd $0xC, %%xmm6, %%xmm8, %%xmm7",
220 "vcmpsd $0xC, (%%rax), %%xmm8, %%xmm7")
222 "vcmpsd $0xD, %%xmm6, %%xmm8, %%xmm7",
223 "vcmpsd $0xD, (%%rax), %%xmm8, %%xmm7")
225 "vcmpsd $0xE, %%xmm6, %%xmm8, %%xmm7",
226 "vcmpsd $0xE, (%%rax), %%xmm8, %%xmm7")
228 "vcmpsd $0x11, %%xmm6, %%xmm8, %%xmm7",
229 "vcmpsd $0x11, (%%rax), %%xmm8, %%xmm7")
231 "vcmpsd $0x12, %%xmm6, %%xmm8, %%xmm7",
232 "vcmpsd $0x12, (%%rax), %%xmm8, %%xmm7")
234 "vcmpsd $0x16, %%xmm6, %%xmm8, %%xmm7",
235 "vcmpsd $0x16, (%%rax), %%xmm8, %%xmm7")
237 "vcmpsd $0x1E, %%xmm6, %%xmm8, %%xmm7",
238 "vcmpsd $0x1E, (%%rax), %%xmm8, %%xmm7")
241 "vsqrtsd %%xmm6, %%xmm8, %%xmm7",
242 "vsqrtsd (%%rax), %%xmm8, %%xmm7")
245 "vorps %%xmm6, %%xmm8, %%xmm7",
246 "vorps (%%rax), %%xmm8, %%xmm7")
249 "vandnps %%xmm6, %%xmm8, %%xmm7",
250 "vandnps (%%rax), %%xmm8, %%xmm7")
253 "vmaxss %%xmm6, %%xmm8, %%xmm7",
254 "vmaxss (%%rax), %%xmm8, %%xmm7")
257 "vminss %%xmm6, %%xmm8, %%xmm7",
258 "vminss (%%rax), %%xmm8, %%xmm7")
261 "vandps %%xmm6, %%xmm8, %%xmm7",
262 "vandps (%%rax), %%xmm8, %%xmm7")
265 "vcvtsi2ssl %%r14d, %%xmm8, %%xmm7",
266 "vcvtsi2ssl (%%rax), %%xmm8, %%xmm7")
269 "vunpcklps %%xmm6, %%xmm8, %%xmm7",
270 "vunpcklps (%%rax), %%xmm8, %%xmm7")
273 "vdivss %%xmm6, %%xmm8, %%xmm7",
274 "vdivss (%%rax), %%xmm8, %%xmm7")
277 "vaddss %%xmm6, %%xmm8, %%xmm7",
278 "vaddss (%%rax), %%xmm8, %%xmm7")
281 "vsubss %%xmm6, %%xmm8, %%xmm7",
282 "vsubss (%%rax), %%xmm8, %%xmm7")
285 "vmulss %%xmm6, %%xmm8, %%xmm7",
286 "vmulss (%%rax), %%xmm8, %%xmm7")
289 "vpunpcklbw %%xmm6, %%xmm8, %%xmm7",
290 "vpunpcklbw (%%rax), %%xmm8, %%xmm7")
293 "vpunpckhbw %%xmm6, %%xmm8, %%xmm7",
294 "vpunpckhbw (%%rax), %%xmm8, %%xmm7")
305 "vmovq %%xmm7, %%xmm8",
310 "vmovq %%xmm7, %%r14")
314 "vmovhlps %%xmm6, %%xmm8, %%xmm7")
322 "vmovlhps %%xmm6, %%xmm8, %%xmm7")
331 "vmovups %%xmm8, %%xmm7",
335 "vmovq %%r14, %%xmm7",
339 "vpcmpestrm $0x45, %%xmm7, %%xmm8; movapd %%xmm0, %%xmm9",
344 "vmovd %%xmm7, %%r14d")
347 "vcvtsd2ss %%xmm9, %%xmm8, %%xmm7",
348 "vcvtsd2ss (%%rax), %%xmm8, %%xmm7")
351 "vcvtss2sd %%xmm9, %%xmm8, %%xmm7",
352 "vcvtss2sd (%%rax), %%xmm8, %%xmm7")
355 "vpackuswb %%xmm9, %%xmm8, %%xmm7",
356 "vpackuswb (%%rax), %%xmm8, %%xmm7")
370 "vpand %%xmm9, %%xmm8, %%xmm7",
371 "vpand (%%rax), %%xmm8, %%xmm7")
380 "vpcmpeqb %%xmm9, %%xmm8, %%xmm7",
381 "vpcmpeqb (%%rax), %%xmm8, %%xmm7")
384 "vshufps $0x39, %%xmm9, %%xmm8, %%xmm7",
385 "vshufps $0xC6, (%%rax), %%xmm8, %%xmm7")
388 "vmulps %%xmm9, %%xmm8, %%xmm7",
389 "vmulps (%%rax), %%xmm8, %%xmm7")
392 "vsubps %%xmm9, %%xmm8, %%xmm7",
393 "vsubps (%%rax), %%xmm8, %%xmm7")
396 "vaddps %%xmm9, %%xmm8, %%xmm7",
397 "vaddps (%%rax), %%xmm8, %%xmm7")
400 "vmaxps %%xmm9, %%xmm8, %%xmm7",
401 "vmaxps (%%rax), %%xmm8, %%xmm7")
408 "vmaxpd %%xmm9, %%xmm8, %%xmm7",
409 "vmaxpd (%%rax), %%xmm8, %%xmm7")
416 "vminps %%xmm9, %%xmm8, %%xmm7",
417 "vminps (%%rax), %%xmm8, %%xmm7")
424 "vminpd %%xmm9, %%xmm8, %%xmm7",
425 "vminpd (%%rax), %%xmm8, %%xmm7")
432 "vcvtps2dq %%xmm8, %%xmm7",
436 "vpshuflw $0x39, %%xmm9, %%xmm7",
440 "vpshufhw $0x39, %%xmm9, %%xmm7",
444 "vpmullw %%xmm9, %%xmm8, %%xmm7",
445 "vpmullw (%%rax), %%xmm8, %%xmm7")
448 "vpaddusw %%xmm9, %%xmm8, %%xmm7",
449 "vpaddusw (%%rax), %%xmm8, %%xmm7")
452 "vpmulhuw %%xmm9, %%xmm8, %%xmm7",
453 "vpmulhuw (%%rax), %%xmm8, %%xmm7")
456 "vpaddusb %%xmm9, %%xmm8, %%xmm7",
457 "vpaddusb (%%rax), %%xmm8, %%xmm7")
460 "vpunpcklwd %%xmm6, %%xmm8, %%xmm7",
461 "vpunpcklwd (%%rax), %%xmm8, %%xmm7")
464 "vpunpckhwd %%xmm6, %%xmm8, %%xmm7",
465 "vpunpckhwd (%%rax), %%xmm8, %%xmm7")
468 "vpslld $0x5, %%xmm9, %%xmm7")
471 "vpsrld $0x5, %%xmm9, %%xmm7")
474 "vpsrad $0x5, %%xmm9, %%xmm7")
477 "vpsubusb %%xmm9, %%xmm8, %%xmm7",
478 "vpsubusb (%%rax), %%xmm8, %%xmm7")
481 "vpsubsb %%xmm9, %%xmm8, %%xmm7",
482 "vpsubsb (%%rax), %%xmm8, %%xmm7")
485 "vpsrldq $0x5, %%xmm9, %%xmm7")
488 "vpslldq $0x5, %%xmm9, %%xmm7")
491 "vpandn %%xmm9, %%xmm8, %%xmm7",
492 "vpandn (%%rax), %%xmm8, %%xmm7")
496 "vmovd %%xmm7, (%%rax)")
499 "vpinsrd $0, %%r14d, %%xmm8, %%xmm7",
500 "vpinsrd $3, (%%rax), %%xmm8, %%xmm7")
503 "vpunpcklqdq %%xmm6, %%xmm8, %%xmm7",
504 "vpunpcklqdq (%%rax), %%xmm8, %%xmm7")
507 "vpsrlw $0x5, %%xmm9, %%xmm7")
510 "vpsllw $0x5, %%xmm9, %%xmm7")
513 "vpaddw %%xmm6, %%xmm8, %%xmm7",
514 "vpaddw (%%rax), %%xmm8, %%xmm7")
517 "vpackssdw %%xmm9, %%xmm8, %%xmm7",
518 "vpackssdw (%%rax), %%xmm8, %%xmm7")
521 "vpunpckldq %%xmm6, %%xmm8, %%xmm7",
522 "vpunpckldq (%%rax), %%xmm8, %%xmm7")
525 "vinsertps $0x39, %%xmm6, %%xmm8, %%xmm7",
526 "vinsertps $0xC6, (%%rax), %%xmm8, %%xmm7")
538 "vmovupd %%xmm7, (%%rax)")
554 "vmovaps %%xmm7, (%%rax)")
562 "vmovapd %%xmm7, (%%rax)")
582 "vmovdqu %%xmm7, (%%rax)")
586 "vmovdqa %%xmm7, (%%rax)")
595 "vmovd %%r14d, %%xmm7",
599 "vmovddup %%xmm8, %%xmm7",
603 "vcmpss $0, %%xmm6, %%xmm8, %%xmm7",
604 "vcmpss $0, (%%rax), %%xmm8, %%xmm7")
606 "vcmpss $1, %%xmm6, %%xmm8, %%xmm7",
607 "vcmpss $1, (%%rax), %%xmm8, %%xmm7")
609 "vcmpss $2, %%xmm6, %%xmm8, %%xmm7",
610 "vcmpss $2, (%%rax), %%xmm8, %%xmm7")
612 "vcmpss $3, %%xmm6, %%xmm8, %%xmm7",
613 "vcmpss $3, (%%rax), %%xmm8, %%xmm7")
615 "vcmpss $4, %%xmm6, %%xmm8, %%xmm7",
616 "vcmpss $4, (%%rax), %%xmm8, %%xmm7")
618 "vcmpss $5, %%xmm6, %%xmm8, %%xmm7",
619 "vcmpss $5, (%%rax), %%xmm8, %%xmm7")
621 "vcmpss $6, %%xmm6, %%xmm8, %%xmm7",
622 "vcmpss $6, (%%rax), %%xmm8, %%xmm7")
624 "vcmpss $7, %%xmm6, %%xmm8, %%xmm7",
625 "vcmpss $7, (%%rax), %%xmm8, %%xmm7")
627 "vcmpss $0xA, %%xmm6, %%xmm8, %%xmm7",
628 "vcmpss $0xA, (%%rax), %%xmm8, %%xmm7")
630 "vcmpss $0xC, %%xmm6, %%xmm8, %%xmm7",
631 "vcmpss $0xC, (%%rax), %%xmm8, %%xmm7")
633 "vcmpss $0xD, %%xmm6, %%xmm8, %%xmm7",
634 "vcmpss $0xD, (%%rax), %%xmm8, %%xmm7")
636 "vcmpss $0xE, %%xmm6, %%xmm8, %%xmm7",
637 "vcmpss $0xE, (%%rax), %%xmm8, %%xmm7")
639 "vcmpss $0x11, %%xmm6, %%xmm8, %%xmm7",
640 "vcmpss $0x11, (%%rax), %%xmm8, %%xmm7")
642 "vcmpss $0x12, %%xmm6, %%xmm8, %%xmm7",
643 "vcmpss $0x12, (%%rax), %%xmm8, %%xmm7")
645 "vcmpss $0x16, %%xmm6, %%xmm8, %%xmm7",
646 "vcmpss $0x16, (%%rax), %%xmm8, %%xmm7")
648 "vcmpss $0x1E, %%xmm6, %%xmm8, %%xmm7",
649 "vcmpss $0x1E, (%%rax), %%xmm8, %%xmm7")
653 "vcvtpd2psx %%xmm8, %%xmm7",
673 "vpextrd $0x0, %%xmm7, %%r14d",
674 "vpextrd $0x0, %%xmm7, (%%rax)")
677 "vpextrd $0x3, %%xmm7, %%r14d",
678 "vpextrd $0x3, %%xmm7, (%%rax)")
681 "vpcmpeqd %%xmm6, %%xmm8, %%xmm7",
682 "vpcmpeqd (%%rax), %%xmm8, %%xmm7")
686 "vpshufd $0xC6, (%%rax), %%xmm7")
689 "vmaxsd %%xmm6, %%xmm8, %%xmm7",
690 "vmaxsd (%%rax), %%xmm8, %%xmm7")
693 "vdivsd %%xmm6, %%xmm8, %%xmm7",
694 "vdivsd (%%rax), %%xmm8, %%xmm7")
697 "vminsd %%xmm6, %%xmm8, %%xmm7",
698 "vminsd (%%rax), %%xmm8, %%xmm7")
701 "vsubsd %%xmm6, %%xmm8, %%xmm7",
702 "vsubsd (%%rax), %%xmm8, %%xmm7")
705 "vaddsd %%xmm6, %%xmm8, %%xmm7",
706 "vaddsd (%%rax), %%xmm8, %%xmm7")
709 "vmulsd %%xmm6, %%xmm8, %%xmm7",
710 "vmulsd (%%rax), %%xmm8, %%xmm7")
713 "vxorps %%xmm6, %%xmm8, %%xmm7",
714 "vxorps (%%rax), %%xmm8, %%xmm7")
717 "vxorpd %%xmm6, %%xmm8, %%xmm7",
718 "vxorpd (%%rax), %%xmm8, %%xmm7")
721 "vorpd %%xmm6, %%xmm8, %%xmm7",
722 "vorpd (%%rax), %%xmm8, %%xmm7")
725 "vandnpd %%xmm6, %%xmm8, %%xmm7",
726 "vandnpd (%%rax), %%xmm8, %%xmm7")
741 "vpinsrq $0, %%r14, %%xmm8, %%xmm7",
742 "vpinsrq $1, (%%rax), %%xmm8, %%xmm7")
745 "vpaddq %%xmm6, %%xmm8, %%xmm7",
746 "vpaddq (%%rax), %%xmm8, %%xmm7")
749 "vpsubq %%xmm6, %%xmm8, %%xmm7",
750 "vpsubq (%%rax), %%xmm8, %%xmm7")
753 "vpsubw %%xmm6, %%xmm8, %%xmm7",
754 "vpsubw (%%rax), %%xmm8, %%xmm7")
770 "vmovupd (%%rax), %%xmm7")
785 "vpcmpeqq %%xmm6, %%xmm8, %%xmm7",
786 "vpcmpeqq (%%rax), %%xmm8, %%xmm7")
789 "vsubpd %%xmm6, %%xmm8, %%xmm7",
790 "vsubpd (%%rax), %%xmm8, %%xmm7")
793 "vaddpd %%xmm6, %%xmm8, %%xmm7",
794 "vaddpd (%%rax), %%xmm8, %%xmm7")
797 "vunpcklpd %%xmm6, %%xmm8, %%xmm7",
798 "vunpcklpd (%%rax), %%xmm8, %%xmm7")
801 "vunpckhpd %%xmm6, %%xmm8, %%xmm7",
802 "vunpckhpd (%%rax), %%xmm8, %%xmm7")
805 "vunpckhps %%xmm6, %%xmm8, %%xmm7",
806 "vunpckhps (%%rax), %%xmm8, %%xmm7")
829 "vpcmpgtq %%xmm6, %%xmm8, %%xmm7",
830 "vpcmpgtq (%%rax), %%xmm8, %%xmm7")
833 "vpextrq $0x0, %%xmm7, %%r14",
834 "vpextrq $0x0, %%xmm7, (%%rax)")
837 "vpextrq $0x1, %%xmm7, %%r14",
838 "vpextrq $0x1, %%xmm7, (%%rax)")
841 "vpsrlq $0x5, %%xmm9, %%xmm7")
844 "vpmuludq %%xmm6, %%xmm8, %%xmm7",
845 "vpmuludq (%%rax), %%xmm8, %%xmm7")
848 "vpmuldq %%xmm6, %%xmm8, %%xmm7",
849 "vpmuldq (%%rax), %%xmm8, %%xmm7")
852 "vpsllq $0x5, %%xmm9, %%xmm7")
855 "vpmaxud %%xmm6, %%xmm8, %%xmm7",
856 "vpmaxud (%%rax), %%xmm8, %%xmm7")
859 "vpminud %%xmm6, %%xmm8, %%xmm7",
860 "vpminud (%%rax), %%xmm8, %%xmm7")
863 "vpmulld %%xmm6, %%xmm8, %%xmm7",
864 "vpmulld (%%rax), %%xmm8, %%xmm7")
867 "vpmaxuw %%xmm6, %%xmm8, %%xmm7",
868 "vpmaxuw (%%rax), %%xmm8, %%xmm7")
871 "vpextrw $0x0, %%xmm7, %%r14d")
874 "vpextrw $0x7, %%xmm7, %%r14d")
877 "vpminuw %%xmm6, %%xmm8, %%xmm7",
878 "vpminuw (%%rax), %%xmm8, %%xmm7")
882 "vphminposuw (%%rax), %%xmm7")
885 "vpmaxsw %%xmm6, %%xmm8, %%xmm7",
886 "vpmaxsw (%%rax), %%xmm8, %%xmm7")
889 "vpminsw %%xmm6, %%xmm8, %%xmm7",
890 "vpminsw (%%rax), %%xmm8, %%xmm7")
893 "vpmaxub %%xmm6, %%xmm8, %%xmm7",
894 "vpmaxub (%%rax), %%xmm8, %%xmm7")
929 "vpminub %%xmm6, %%xmm8, %%xmm7",
930 "vpminub (%%rax), %%xmm8, %%xmm7")
933 "vpmaxsb %%xmm6, %%xmm8, %%xmm7",
934 "vpmaxsb (%%rax), %%xmm8, %%xmm7")
937 "vpminsb %%xmm6, %%xmm8, %%xmm7",
938 "vpminsb (%%rax), %%xmm8, %%xmm7")
1013 "vcmppd $4, %%xmm6, %%xmm8, %%xmm7",
1014 "vcmppd $4, (%%rax), %%xmm8, %%xmm7")
1021 "vcmpps $4, %%xmm6, %%xmm8, %%xmm7",
1022 "vcmpps $4, (%%rax), %%xmm8, %%xmm7")
1033 "vdivpd %%xmm6, %%xmm8, %%xmm7",
1034 "vdivpd (%%rax), %%xmm8, %%xmm7")
1045 "vpsubusw %%xmm9, %%xmm8, %%xmm7",
1046 "vpsubusw (%%rax), %%xmm8, %%xmm7")
1049 "vpsubsw %%xmm9, %%xmm8, %%xmm7",
1050 "vpsubsw (%%rax), %%xmm8, %%xmm7")
1053 "vpcmpeqw %%xmm6, %%xmm8, %%xmm7",
1054 "vpcmpeqw (%%rax), %%xmm8, %%xmm7")
1057 "vpaddb %%xmm6, %%xmm8, %%xmm7",
1058 "vpaddb (%%rax), %%xmm8, %%xmm7")
1069 "vmovhpd (%%rax), %%xmm8, %%xmm7")
1072 "vmovhps (%%rax), %%xmm8, %%xmm7")
1076 "vcvtpd2psy %%ymm8, %%xmm7",
1080 "vpunpckhdq %%xmm6, %%xmm8, %%xmm7",
1081 "vpunpckhdq (%%rax), %%xmm8, %%xmm7")
1095 "vdivps %%xmm9, %%xmm8, %%xmm7",
1096 "vdivps (%%rax), %%xmm8, %%xmm7")
1136 "vpmulhw %%xmm9, %%xmm8, %%xmm7",
1137 "vpmulhw (%%rax), %%xmm8, %%xmm7")
1140 "vpunpckhqdq %%xmm6, %%xmm8, %%xmm7",
1141 "vpunpckhqdq (%%rax), %%xmm8, %%xmm7")
1144 "vpsraw $0x5, %%xmm9, %%xmm7")
1147 "vpcmpgtb %%xmm6, %%xmm8, %%xmm7",
1148 "vpcmpgtb (%%rax), %%xmm8, %%xmm7")
1151 "vpcmpgtw %%xmm6, %%xmm8, %%xmm7",
1152 "vpcmpgtw (%%rax), %%xmm8, %%xmm7")
1155 "vpcmpgtd %%xmm6, %%xmm8, %%xmm7",
1156 "vpcmpgtd (%%rax), %%xmm8, %%xmm7")
1167 "vpinsrb $0, %%r14d, %%xmm8, %%xmm7",
1168 "vpinsrb $3, (%%rax), %%xmm8, %%xmm7")
1170 "vpinsrb $6, %%r14d, %%xmm8, %%xmm7",
1171 "vpinsrb $9, (%%rax), %%xmm8, %%xmm7")
1173 "vpinsrb $12, %%r14d, %%xmm8, %%xmm7",
1174 "vpinsrb $15, (%%rax), %%xmm8, %%xmm7")
1177 "vpinsrw $0, %%r14d, %%xmm8, %%xmm7",
1178 "vpinsrw $3, (%%rax), %%xmm8, %%xmm7")
1180 "vpinsrw $2, %%r14d, %%xmm8, %%xmm7",
1181 "vpinsrw $3, (%%rax), %%xmm8, %%xmm7")
1183 "vpinsrw $4, %%r14d, %%xmm8, %%xmm7",
1184 "vpinsrw $5, (%%rax), %%xmm8, %%xmm7")
1186 "vpinsrw $6, %%r14d, %%xmm8, %%xmm7",
1187 "vpinsrw $7, (%%rax), %%xmm8, %%xmm7")
1202 "vdppd $0x00, %%xmm6, %%xmm8, %%xmm7",
1205 "vdppd $0x5A, %%xmm6, %%xmm8, %%xmm7",
1208 "vdppd $0x0F, %%xmm6, %%xmm8, %%xmm7",
1211 "vdppd $0xF0, %%xmm6, %%xmm8, %%xmm7",
1215 "vdpps $0x00, %%xmm6, %%xmm8, %%xmm7",
1218 "vdpps $0x5A, %%xmm6, %%xmm8, %%xmm7",
1221 "vdpps $0x0F, %%xmm6, %%xmm8, %%xmm7",
1224 "vdpps $0xF0, %%xmm6, %%xmm8, %%xmm7",
1244 "vpalignr $0, %%xmm6, %%xmm8, %%xmm7",
1245 "vpalignr $3, (%%rax), %%xmm8, %%xmm7")
1247 "vpalignr $6, %%xmm6, %%xmm8, %%xmm7",
1248 "vpalignr $9, (%%rax), %%xmm8, %%xmm7")
1250 "vpalignr $12, %%xmm6, %%xmm8, %%xmm7",
1251 "vpalignr $15, (%%rax), %%xmm8, %%xmm7")
1253 GEN_test_Ronly(VMOVSD_REG_XMM, "vmovsd %%xmm9, %%xmm7, %%xmm8")
1255 GEN_test_Ronly(VMOVSS_REG_XMM, "vmovss %%xmm9, %%xmm7, %%xmm8")
1257 GEN_test_Monly(VMOVLPD_128_M64_XMM_XMM, "vmovlpd (%%rax), %%xmm8, %%xmm7")
1259 GEN_test_Monly(VMOVLPD_128_XMM_M64, "vmovlpd %%xmm7, (%%rax)")
1262 "vshufpd $0, %%xmm9, %%xmm8, %%xmm7",
1263 "vshufpd $1, (%%rax), %%xmm8, %%xmm7")
1265 "vshufpd $2, %%xmm9, %%xmm8, %%xmm7",
1266 "vshufpd $3, (%%rax), %%xmm8, %%xmm7")
1301 "vblendps $0, %%xmm6, %%xmm8, %%xmm7",
1302 "vblendps $3, (%%rax), %%xmm8, %%xmm7")
1304 "vblendps $6, %%xmm6, %%xmm8, %%xmm7",
1305 "vblendps $9, (%%rax), %%xmm8, %%xmm7")
1307 "vblendps $12, %%xmm6, %%xmm8, %%xmm7",
1308 "vblendps $15, (%%rax), %%xmm8, %%xmm7")
1311 "vblendpd $0, %%xmm6, %%xmm8, %%xmm7",
1312 "vblendpd $1, (%%rax), %%xmm8, %%xmm7")
1314 "vblendpd $2, %%xmm6, %%xmm8, %%xmm7",
1315 "vblendpd $3, (%%rax), %%xmm8, %%xmm7")
1328 "vpblendw $0x00, %%xmm6, %%xmm8, %%xmm7",
1329 "vpblendw $0x01, (%%rax), %%xmm8, %%xmm7")
1331 "vpblendw $0xFE, %%xmm6, %%xmm8, %%xmm7",
1332 "vpblendw $0xFF, (%%rax), %%xmm8, %%xmm7")
1334 "vpblendw $0x30, %%xmm6, %%xmm8, %%xmm7",
1335 "vpblendw $0x03, (%%rax), %%xmm8, %%xmm7")
1337 "vpblendw $0x21, %%xmm6, %%xmm8, %%xmm7",
1338 "vpblendw $0x12, (%%rax), %%xmm8, %%xmm7")
1340 "vpblendw $0xD7, %%xmm6, %%xmm8, %%xmm7",
1341 "vpblendw $0x6C, (%%rax), %%xmm8, %%xmm7")
1343 "vpblendw $0xB5, %%xmm6, %%xmm8, %%xmm7",
1344 "vpblendw $0x4A, (%%rax), %%xmm8, %%xmm7")
1346 "vpblendw $0x85, %%xmm6, %%xmm8, %%xmm7",
1347 "vpblendw $0xDC, (%%rax), %%xmm8, %%xmm7")
1349 "vpblendw $0x29, %%xmm6, %%xmm8, %%xmm7",
1350 "vpblendw $0x92, (%%rax), %%xmm8, %%xmm7")
1357 "vsqrtss %%xmm6, %%xmm8, %%xmm7",
1358 "vsqrtss (%%rax), %%xmm8, %%xmm7")
1377 "vrsqrtss %%xmm6, %%xmm8, %%xmm7",
1378 "vrsqrtss (%%rax), %%xmm8, %%xmm7")
1398 "vcvttps2dq (%%rax), %%xmm7")
1406 "vcvtdq2ps (%%rax), %%xmm7")
1414 "vcvttpd2dqx (%%rax), %%xmm7")
1418 "vcvttpd2dqy (%%rax), %%xmm7")
1422 "vcvtpd2dqx (%%rax), %%xmm7")
1426 "vcvtpd2dqy (%%rax), %%xmm7")
1430 "vmovsldup (%%rax), %%xmm7")
1438 "vmovshdup (%%rax), %%xmm7")
1445 "vpermilps %%xmm6, %%xmm8, %%xmm7",
1446 "vpermilps (%%rax), %%xmm8, %%xmm7")
1449 "vpermilpd %%xmm6, %%xmm8, %%xmm7",
1450 "vpermilpd (%%rax), %%xmm8, %%xmm7")
1581 "vpmaddwd %%xmm6, %%xmm8, %%xmm7",
1582 "vpmaddwd (%%rax), %%xmm8, %%xmm7")
1585 "vaddsubps %%xmm6, %%xmm8, %%xmm7",
1586 "vaddsubps (%%rax), %%xmm8, %%xmm7")
1593 "vaddsubpd %%xmm6, %%xmm8, %%xmm7",
1594 "vaddsubpd (%%rax), %%xmm8, %%xmm7")
1652 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1653 "vxorpd %%xmm8,%%xmm7,%%xmm8;"
1671 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1673 "vmovups %%xmm7,512(%%rsp);"
1674 "vmovups %%xmm7,528(%%rsp);"
1699 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1700 "vxorpd %%xmm8,%%xmm7,%%xmm8;"
1727 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1729 "vmovups %%xmm7,512(%%rsp);"
1730 "vmovups %%xmm7,528(%%rsp);"
1759 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1760 "vxorpd %%xmm8,%%xmm7,%%xmm8;"
1787 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1789 "vmovups %%xmm7,512(%%rsp);"
1790 "vmovups %%xmm7,528(%%rsp);"
1804 "vblendvps %%xmm9, %%xmm6, %%xmm8, %%xmm7",
1805 "vblendvps %%xmm9, (%%rax), %%xmm8, %%xmm7")
1812 "vblendvpd %%xmm9, %%xmm6, %%xmm8, %%xmm7",
1813 "vblendvpd %%xmm9, (%%rax), %%xmm8, %%xmm7")
1821 "vhaddps %%xmm6, %%xmm8, %%xmm7",
1822 "vhaddps (%%rax), %%xmm8, %%xmm7")
1829 "vhaddpd %%xmm6, %%xmm8, %%xmm7",
1830 "vhaddpd (%%rax), %%xmm8, %%xmm7")
1837 "vhsubps %%xmm6, %%xmm8, %%xmm7",
1838 "vhsubps (%%rax), %%xmm8, %%xmm7")
1845 "vhsubpd %%xmm6, %%xmm8, %%xmm7",
1846 "vhsubpd (%%rax), %%xmm8, %%xmm7")
1907 "vpacksswb %%xmm6, %%xmm8, %%xmm7",
1908 "vpacksswb (%%rax), %%xmm8, %%xmm7")
1911 "vpavgb %%xmm6, %%xmm8, %%xmm7",
1912 "vpavgb (%%rax), %%xmm8, %%xmm7")
1915 "vpavgw %%xmm6, %%xmm8, %%xmm7",
1916 "vpavgw (%%rax), %%xmm8, %%xmm7")
1919 "vpaddsb %%xmm6, %%xmm8, %%xmm7",
1920 "vpaddsb (%%rax), %%xmm8, %%xmm7")
1923 "vpaddsw %%xmm6, %%xmm8, %%xmm7",
1924 "vpaddsw (%%rax), %%xmm8, %%xmm7")
1927 "vphaddw %%xmm6, %%xmm8, %%xmm7",
1928 "vphaddw (%%rax), %%xmm8, %%xmm7")
1931 "vphaddd %%xmm6, %%xmm8, %%xmm7",
1932 "vphaddd (%%rax), %%xmm8, %%xmm7")
1935 "vphaddsw %%xmm6, %%xmm8, %%xmm7",
1936 "vphaddsw (%%rax), %%xmm8, %%xmm7")
1939 "vpmaddubsw %%xmm6, %%xmm8, %%xmm7",
1940 "vpmaddubsw (%%rax), %%xmm8, %%xmm7")
1943 "vphsubw %%xmm6, %%xmm8, %%xmm7",
1944 "vphsubw (%%rax), %%xmm8, %%xmm7")
1947 "vphsubd %%xmm6, %%xmm8, %%xmm7",
1948 "vphsubd (%%rax), %%xmm8, %%xmm7")
1951 "vphsubsw %%xmm6, %%xmm8, %%xmm7",
1952 "vphsubsw (%%rax), %%xmm8, %%xmm7")
1955 "vpabsb %%xmm6, %%xmm7",
1956 "vpabsb (%%rax), %%xmm7")
1959 "vpabsw %%xmm6, %%xmm7",
1960 "vpabsw (%%rax), %%xmm7")
1971 "vpackusdw %%xmm6, %%xmm8, %%xmm7",
1972 "vpackusdw (%%rax), %%xmm8, %%xmm7")
1987 "vmpsadbw $0, %%xmm6, %%xmm8, %%xmm7",
1988 "vmpsadbw $0, (%%rax), %%xmm8, %%xmm7")
1990 "vmpsadbw $1, %%xmm6, %%xmm8, %%xmm7",
1991 "vmpsadbw $1, (%%rax), %%xmm8, %%xmm7")
1993 "vmpsadbw $2, %%xmm6, %%xmm8, %%xmm7",
1994 "vmpsadbw $2, (%%rax), %%xmm8, %%xmm7")
1996 "vmpsadbw $3, %%xmm6, %%xmm8, %%xmm7",
1997 "vmpsadbw $3, (%%rax), %%xmm8, %%xmm7")
1999 "vmpsadbw $4, %%xmm6, %%xmm8, %%xmm7",
2000 "vmpsadbw $4, (%%rax), %%xmm8, %%xmm7")
2002 "vmpsadbw $5, %%xmm6, %%xmm8, %%xmm7",
2003 "vmpsadbw $5, (%%rax), %%xmm8, %%xmm7")
2005 "vmpsadbw $6, %%xmm6, %%xmm8, %%xmm7",
2006 "vmpsadbw $6, (%%rax), %%xmm8, %%xmm7")
2008 "vmpsadbw $7, %%xmm6, %%xmm8, %%xmm7",
2009 "vmpsadbw $7, (%%rax), %%xmm8, %%xmm7")
2015 GEN_test_Monly(VMOVLPS_128_M64_XMM_XMM, "vmovlps (%%rax), %%xmm8, %%xmm7")
2017 GEN_test_Monly(VMOVLPS_128_XMM_M64, "vmovlps %%xmm7, (%%rax)")
2020 "vrcpss %%xmm6, %%xmm8, %%xmm7",
2021 "vrcpss (%%rax), %%xmm8, %%xmm7")
2032 "vpsadbw %%xmm6, %%xmm8, %%xmm7",
2033 "vpsadbw (%%rax), %%xmm8, %%xmm7")
2036 "vpsignb %%xmm6, %%xmm8, %%xmm7",
2037 "vpsignb (%%rax), %%xmm8, %%xmm7")
2040 "vpsignw %%xmm6, %%xmm8, %%xmm7",
2041 "vpsignw (%%rax), %%xmm8, %%xmm7")
2044 "vpsignd %%xmm6, %%xmm8, %%xmm7",
2045 "vpsignd (%%rax), %%xmm8, %%xmm7")
2048 "vpmulhrsw %%xmm6, %%xmm8, %%xmm7",
2049 "vpmulhrsw (%%rax), %%xmm8, %%xmm7")
2055 "vpextrw $0x0, %%xmm7, %%r14d",
2056 "vpextrw $0x0, %%xmm7, (%%rax)")
2058 "vpextrw $0x1, %%xmm7, %%r14d",
2059 "vpextrw $0x1, %%xmm7, (%%rax)")
2061 "vpextrw $0x2, %%xmm7, %%r14d",
2062 "vpextrw $0x2, %%xmm7, (%%rax)")
2064 "vpextrw $0x3, %%xmm7, %%r14d",
2065 "vpextrw $0x3, %%xmm7, (%%rax)")
2067 "vpextrw $0x4, %%xmm7, %%r14d",
2068 "vpextrw $0x4, %%xmm7, (%%rax)")
2070 "vpextrw $0x5, %%xmm7, %%r14d",
2071 "vpextrw $0x5, %%xmm7, (%%rax)")
2073 "vpextrw $0x6, %%xmm7, %%r14d",
2074 "vpextrw $0x6, %%xmm7, (%%rax)")
2076 "vpextrw $0x7, %%xmm7, %%r14d",
2077 "vpextrw $0x7, %%xmm7, (%%rax)")
2080 "vaesenc %%xmm6, %%xmm8, %%xmm7",
2081 "vaesenc (%%rax), %%xmm8, %%xmm7")
2084 "vaesenclast %%xmm6, %%xmm8, %%xmm7",
2085 "vaesenclast (%%rax), %%xmm8, %%xmm7")
2088 "vaesdec %%xmm6, %%xmm8, %%xmm7",
2089 "vaesdec (%%rax), %%xmm8, %%xmm7")
2092 "vaesdeclast %%xmm6, %%xmm8, %%xmm7",
2093 "vaesdeclast (%%rax), %%xmm8, %%xmm7")
2096 "vaesimc %%xmm6, %%xmm7",
2097 "vaesimc (%%rax), %%xmm7")
2100 "vaeskeygenassist $0x00, %%xmm6, %%xmm7",
2101 "vaeskeygenassist $0x00, (%%rax), %%xmm7")
2103 "vaeskeygenassist $0x31, %%xmm6, %%xmm7",
2104 "vaeskeygenassist $0x31, (%%rax), %%xmm7")
2106 "vaeskeygenassist $0xb2, %%xmm6, %%xmm7",
2107 "vaeskeygenassist $0xb2, (%%rax), %%xmm7")
2109 "vaeskeygenassist $0xFF, %%xmm6, %%xmm7",
2110 "vaeskeygenassist $0xFF, (%%rax), %%xmm7")
2113 "vpclmulqdq $0x00, %%xmm6, %%xmm8, %%xmm7",
2114 "vpclmulqdq $0x00, (%%rax), %%xmm8, %%xmm7")
2116 "vpclmulqdq $0x01, %%xmm6, %%xmm8, %%xmm7",
2117 "vpclmulqdq $0x01, (%%rax), %%xmm8, %%xmm7")
2119 "vpclmulqdq $0x10, %%xmm6, %%xmm8, %%xmm7",
2120 "vpclmulqdq $0x10, (%%rax), %%xmm8, %%xmm7")
2122 "vpclmulqdq $0x11, %%xmm6, %%xmm8, %%xmm7",
2123 "vpclmulqdq $0x11, (%%rax), %%xmm8, %%xmm7")
2125 "vpclmulqdq $0xFF, %%xmm6, %%xmm8, %%xmm7",
2126 "vpclmulqdq $0xFF, (%%rax), %%xmm8, %%xmm7")
2129 "vcmpss $0x9, %%xmm6, %%xmm8, %%xmm7",
2130 "vcmpss $0x9, (%%rax), %%xmm8, %%xmm7")