Lines Matching full:xmm8
84 : /*TRASH*/"xmm0","xmm7","xmm8","xmm6","xmm9","r14","memory","cc" \
104 "xmm0","xmm8","xmm7","xmm9","r14","rax","memory","cc" \
118 "vpor %%xmm6, %%xmm8, %%xmm7",
119 "vpor (%%rax), %%xmm8, %%xmm7")
122 "vpxor %%xmm6, %%xmm8, %%xmm7",
123 "vpxor (%%rax), %%xmm8, %%xmm7")
126 "vpsubb %%xmm6, %%xmm8, %%xmm7",
127 "vpsubb (%%rax), %%xmm8, %%xmm7")
130 "vpsubd %%xmm6, %%xmm8, %%xmm7",
131 "vpsubd (%%rax), %%xmm8, %%xmm7")
134 "vpaddd %%xmm6, %%xmm8, %%xmm7",
135 "vpaddd (%%rax), %%xmm8, %%xmm7")
138 "vpmovzxwd %%xmm6, %%xmm8",
139 "vpmovzxwd (%%rax), %%xmm8")
142 "vpmovzxbw %%xmm6, %%xmm8",
143 "vpmovzxbw (%%rax), %%xmm8")
146 "vpblendvb %%xmm9, %%xmm6, %%xmm8, %%xmm7",
147 "vpblendvb %%xmm9, (%%rax), %%xmm8, %%xmm7")
150 "vpminsd %%xmm6, %%xmm8, %%xmm7",
151 "vpminsd (%%rax), %%xmm8, %%xmm7")
154 "vpmaxsd %%xmm6, %%xmm8, %%xmm7",
155 "vpmaxsd (%%rax), %%xmm8, %%xmm7")
158 "vandpd %%xmm6, %%xmm8, %%xmm7",
159 "vandpd (%%rax), %%xmm8, %%xmm7")
162 "vcvtsi2sdl %%r14d, %%xmm8, %%xmm7",
163 "vcvtsi2sdl (%%rax), %%xmm8, %%xmm7")
166 "vcvtsi2sdq %%r14, %%xmm8, %%xmm7",
167 "vcvtsi2sdq (%%rax), %%xmm8, %%xmm7")
170 "vcvtsi2ssq %%r14, %%xmm8, %%xmm7",
171 "vcvtsi2ssq (%%rax), %%xmm8, %%xmm7")
174 "vcvttsd2si %%xmm8, %%r14d",
178 "vcvttsd2si %%xmm8, %%r14",
182 "vcvtsd2si %%xmm8, %%r14d",
186 "vcvtsd2si %%xmm8, %%r14",
190 "vpshufb %%xmm6, %%xmm8, %%xmm7",
191 "vpshufb (%%rax), %%xmm8, %%xmm7")
194 "vcmpsd $0, %%xmm6, %%xmm8, %%xmm7",
195 "vcmpsd $0, (%%rax), %%xmm8, %%xmm7")
197 "vcmpsd $1, %%xmm6, %%xmm8, %%xmm7",
198 "vcmpsd $1, (%%rax), %%xmm8, %%xmm7")
200 "vcmpsd $2, %%xmm6, %%xmm8, %%xmm7",
201 "vcmpsd $2, (%%rax), %%xmm8, %%xmm7")
203 "vcmpsd $3, %%xmm6, %%xmm8, %%xmm7",
204 "vcmpsd $3, (%%rax), %%xmm8, %%xmm7")
206 "vcmpsd $4, %%xmm6, %%xmm8, %%xmm7",
207 "vcmpsd $4, (%%rax), %%xmm8, %%xmm7")
209 "vcmpsd $5, %%xmm6, %%xmm8, %%xmm7",
210 "vcmpsd $5, (%%rax), %%xmm8, %%xmm7")
212 "vcmpsd $6, %%xmm6, %%xmm8, %%xmm7",
213 "vcmpsd $6, (%%rax), %%xmm8, %%xmm7")
215 "vcmpsd $7, %%xmm6, %%xmm8, %%xmm7",
216 "vcmpsd $7, (%%rax), %%xmm8, %%xmm7")
218 "vcmpsd $8, %%xmm6, %%xmm8, %%xmm7",
219 "vcmpsd $8, (%%rax), %%xmm8, %%xmm7")
221 "vcmpsd $0xA, %%xmm6, %%xmm8, %%xmm7",
222 "vcmpsd $0xA, (%%rax), %%xmm8, %%xmm7")
224 "vcmpsd $0xC, %%xmm6, %%xmm8, %%xmm7",
225 "vcmpsd $0xC, (%%rax), %%xmm8, %%xmm7")
227 "vcmpsd $0xD, %%xmm6, %%xmm8, %%xmm7",
228 "vcmpsd $0xD, (%%rax), %%xmm8, %%xmm7")
230 "vcmpsd $0xE, %%xmm6, %%xmm8, %%xmm7",
231 "vcmpsd $0xE, (%%rax), %%xmm8, %%xmm7")
233 "vcmpsd $0x10, %%xmm6, %%xmm8, %%xmm7",
234 "vcmpsd $0x10, (%%rax), %%xmm8, %%xmm7")
236 "vcmpsd $0x11, %%xmm6, %%xmm8, %%xmm7",
237 "vcmpsd $0x11, (%%rax), %%xmm8, %%xmm7")
239 "vcmpsd $0x12, %%xmm6, %%xmm8, %%xmm7",
240 "vcmpsd $0x12, (%%rax), %%xmm8, %%xmm7")
242 "vcmpsd $0x13, %%xmm6, %%xmm8, %%xmm7",
243 "vcmpsd $0x13, (%%rax), %%xmm8, %%xmm7")
245 "vcmpsd $0x14, %%xmm6, %%xmm8, %%xmm7",
246 "vcmpsd $0x14, (%%rax), %%xmm8, %%xmm7")
248 "vcmpsd $0x15, %%xmm6, %%xmm8, %%xmm7",
249 "vcmpsd $0x15, (%%rax), %%xmm8, %%xmm7")
251 "vcmpsd $0x16, %%xmm6, %%xmm8, %%xmm7",
252 "vcmpsd $0x16, (%%rax), %%xmm8, %%xmm7")
254 "vcmpsd $0x17, %%xmm6, %%xmm8, %%xmm7",
255 "vcmpsd $0x17, (%%rax), %%xmm8, %%xmm7")
257 "vcmpsd $0x18, %%xmm6, %%xmm8, %%xmm7",
258 "vcmpsd $0x18, (%%rax), %%xmm8, %%xmm7")
260 "vcmpsd $0x19, %%xmm6, %%xmm8, %%xmm7",
261 "vcmpsd $0x19, (%%rax), %%xmm8, %%xmm7")
263 "vcmpsd $0x1A, %%xmm6, %%xmm8, %%xmm7",
264 "vcmpsd $0x1A, (%%rax), %%xmm8, %%xmm7")
266 "vcmpsd $0x1C, %%xmm6, %%xmm8, %%xmm7",
267 "vcmpsd $0x1C, (%%rax), %%xmm8, %%xmm7")
269 "vcmpsd $0x1D, %%xmm6, %%xmm8, %%xmm7",
270 "vcmpsd $0x1D, (%%rax), %%xmm8, %%xmm7")
272 "vcmpsd $0x1E, %%xmm6, %%xmm8, %%xmm7",
273 "vcmpsd $0x1E, (%%rax), %%xmm8, %%xmm7")
276 "vsqrtsd %%xmm6, %%xmm8, %%xmm7",
277 "vsqrtsd (%%rax), %%xmm8, %%xmm7")
280 "vorps %%xmm6, %%xmm8, %%xmm7",
281 "vorps (%%rax), %%xmm8, %%xmm7")
284 "vandnps %%xmm6, %%xmm8, %%xmm7",
285 "vandnps (%%rax), %%xmm8, %%xmm7")
288 "vmaxss %%xmm6, %%xmm8, %%xmm7",
289 "vmaxss (%%rax), %%xmm8, %%xmm7")
292 "vminss %%xmm6, %%xmm8, %%xmm7",
293 "vminss (%%rax), %%xmm8, %%xmm7")
296 "vandps %%xmm6, %%xmm8, %%xmm7",
297 "vandps (%%rax), %%xmm8, %%xmm7")
300 "vcvtsi2ssl %%r14d, %%xmm8, %%xmm7",
301 "vcvtsi2ssl (%%rax), %%xmm8, %%xmm7")
304 "vunpcklps %%xmm6, %%xmm8, %%xmm7",
305 "vunpcklps (%%rax), %%xmm8, %%xmm7")
308 "vdivss %%xmm6, %%xmm8, %%xmm7",
309 "vdivss (%%rax), %%xmm8, %%xmm7")
312 "vaddss %%xmm6, %%xmm8, %%xmm7",
313 "vaddss (%%rax), %%xmm8, %%xmm7")
316 "vsubss %%xmm6, %%xmm8, %%xmm7",
317 "vsubss (%%rax), %%xmm8, %%xmm7")
320 "vmulss %%xmm6, %%xmm8, %%xmm7",
321 "vmulss (%%rax), %%xmm8, %%xmm7")
324 "vpunpcklbw %%xmm6, %%xmm8, %%xmm7",
325 "vpunpcklbw (%%rax), %%xmm8, %%xmm7")
328 "vpunpckhbw %%xmm6, %%xmm8, %%xmm7",
329 "vpunpckhbw (%%rax), %%xmm8, %%xmm7")
332 "vcvttss2si %%xmm8, %%r14d",
336 "vcvtss2si %%xmm8, %%r14d",
340 "vmovq %%xmm7, %%xmm8",
341 "vmovq (%%rax), %%xmm8")
349 "vmovhlps %%xmm6, %%xmm8, %%xmm7")
352 "vpabsd %%xmm6, %%xmm8",
353 "vpabsd (%%rax), %%xmm8")
357 "vmovlhps %%xmm6, %%xmm8, %%xmm7")
360 "vmovntdq %%xmm8, (%%rax)")
366 "vmovups %%xmm8, %%xmm7",
374 "vpcmpestrm $0x45, %%xmm7, %%xmm8; movapd %%xmm0, %%xmm9",
375 "vpcmpestrm $0x45, (%%rax), %%xmm8; movapd %%xmm0, %%xmm9")
382 "vcvtsd2ss %%xmm9, %%xmm8, %%xmm7",
383 "vcvtsd2ss (%%rax), %%xmm8, %%xmm7")
386 "vcvtss2sd %%xmm9, %%xmm8, %%xmm7",
387 "vcvtss2sd (%%rax), %%xmm8, %%xmm7")
390 "vpackuswb %%xmm9, %%xmm8, %%xmm7",
391 "vpackuswb (%%rax), %%xmm8, %%xmm7")
394 "vcvttss2si %%xmm8, %%r14",
398 "vcvtss2si %%xmm8, %%r14",
402 "vpmovmskb %%xmm8, %%r14")
405 "vpand %%xmm9, %%xmm8, %%xmm7",
406 "vpand (%%rax), %%xmm8, %%xmm7")
409 "vmovhpd %%xmm8, (%%rax)")
412 "vmovhps %%xmm8, (%%rax)")
415 "vpcmpeqb %%xmm9, %%xmm8, %%xmm7",
416 "vpcmpeqb (%%rax), %%xmm8, %%xmm7")
419 "vshufps $0x39, %%xmm9, %%xmm8, %%xmm7",
420 "vshufps $0xC6, (%%rax), %%xmm8, %%xmm7")
423 "vmulps %%xmm9, %%xmm8, %%xmm7",
424 "vmulps (%%rax), %%xmm8, %%xmm7")
427 "vsubps %%xmm9, %%xmm8, %%xmm7",
428 "vsubps (%%rax), %%xmm8, %%xmm7")
431 "vaddps %%xmm9, %%xmm8, %%xmm7",
432 "vaddps (%%rax), %%xmm8, %%xmm7")
435 "vmaxps %%xmm9, %%xmm8, %%xmm7",
436 "vmaxps (%%rax), %%xmm8, %%xmm7")
443 "vmaxpd %%xmm9, %%xmm8, %%xmm7",
444 "vmaxpd (%%rax), %%xmm8, %%xmm7")
451 "vminps %%xmm9, %%xmm8, %%xmm7",
452 "vminps (%%rax), %%xmm8, %%xmm7")
459 "vminpd %%xmm9, %%xmm8, %%xmm7",
460 "vminpd (%%rax), %%xmm8, %%xmm7")
467 "vcvtps2dq %%xmm8, %%xmm7",
468 "vcvtps2dq (%%rax), %%xmm8")
472 "vpshuflw $0xC6, (%%rax), %%xmm8")
476 "vpshufhw $0xC6, (%%rax), %%xmm8")
479 "vpmullw %%xmm9, %%xmm8, %%xmm7",
480 "vpmullw (%%rax), %%xmm8, %%xmm7")
483 xmm8, %%xmm7",
484 "vpaddusw (%%rax), %%xmm8, %%xmm7")
487 "vpmulhuw %%xmm9, %%xmm8, %%xmm7",
488 "vpmulhuw (%%rax), %%xmm8, %%xmm7")
491 "vpaddusb %%xmm9, %%xmm8, %%xmm7",
492 "vpaddusb (%%rax), %%xmm8, %%xmm7")
495 "vpunpcklwd %%xmm6, %%xmm8, %%xmm7",
496 "vpunpcklwd (%%rax), %%xmm8, %%xmm7")
499 "vpunpckhwd %%xmm6, %%xmm8, %%xmm7",
500 "vpunpckhwd (%%rax), %%xmm8, %%xmm7")
512 "vpsubusb %%xmm9, %%xmm8, %%xmm7",
513 "vpsubusb (%%rax), %%xmm8, %%xmm7")
516 "vpsubsb %%xmm9, %%xmm8, %%xmm7",
517 "vpsubsb (%%rax), %%xmm8, %%xmm7")
526 "vpandn %%xmm9, %%xmm8, %%xmm7",
527 "vpandn (%%rax), %%xmm8, %%xmm7")
534 "vpinsrd $0, %%r14d, %%xmm8, %%xmm7",
535 "vpinsrd $3, (%%rax), %%xmm8, %%xmm7")
538 "vpunpcklqdq %%xmm6, %%xmm8, %%xmm7",
539 "vpunpcklqdq (%%rax), %%xmm8, %%xmm7")
548 "vpaddw %%xmm6, %%xmm8, %%xmm7",
549 "vpaddw (%%rax), %%xmm8, %%xmm7")
552 "vpackssdw %%xmm9, %%xmm8, %%xmm7",
553 "vpackssdw (%%rax), %%xmm8, %%xmm7")
556 "vpunpckldq %%xmm6, %%xmm8, %%xmm7",
557 "vpunpckldq (%%rax), %%xmm8, %%xmm7")
560 "vinsertps $0x39, %%xmm6, %%xmm8, %%xmm7",
561 "vinsertps $0xC6, (%%rax), %%xmm8, %%xmm7")
563 GEN_test_Monly(VMOVSD_M64_XMM, "vmovsd (%%rax), %%xmm8")
565 GEN_test_Monly(VMOVSS_M64_XMM, "vmovss (%%rax), %%xmm8")
567 GEN_test_Monly(VMOVSD_XMM_M64, "vmovsd %%xmm8, (%%rax)")
569 GEN_test_Monly(VMOVSS_XMM_M32, "vmovss %%xmm8, (%%rax)")
576 "vmovapd %%xmm6, %%xmm8",
584 "vmovaps %%xmm6, %%xmm8",
604 "vmovdqu %%xmm6, %%xmm8",
608 "vmovdqa %%xmm6, %%xmm8",
627 GEN_test_Monly(VMOVQ_XMM_MEM64, "vmovq %%xmm8, (%%rax)")
634 "vmovddup %%xmm8, %%xmm7",
638 "vcmpss $0, %%xmm6, %%xmm8, %%xmm7",
639 "vcmpss $0, (%%rax), %%xmm8, %%xmm7")
641 "vcmpss $1, %%xmm6, %%xmm8, %%xmm7",
642 "vcmpss $1, (%%rax), %%xmm8, %%xmm7")
644 "vcmpss $2, %%xmm6, %%xmm8, %%xmm7",
645 "vcmpss $2, (%%rax), %%xmm8, %%xmm7")
647 "vcmpss $3, %%xmm6, %%xmm8, %%xmm7",
648 "vcmpss $3, (%%rax), %%xmm8, %%xmm7")
650 "vcmpss $4, %%xmm6, %%xmm8, %%xmm7",
651 "vcmpss $4, (%%rax), %%xmm8, %%xmm7")
653 "vcmpss $5, %%xmm6, %%xmm8, %%xmm7",
654 "vcmpss $5, (%%rax), %%xmm8, %%xmm7")
656 "vcmpss $6, %%xmm6, %%xmm8, %%xmm7",
657 "vcmpss $6, (%%rax), %%xmm8, %%xmm7")
659 "vcmpss $7, %%xmm6, %%xmm8, %%xmm7",
660 "vcmpss $7, (%%rax), %%xmm8, %%xmm7")
662 "vcmpss $8, %%xmm6, %%xmm8, %%xmm7",
663 "vcmpss $8, (%%rax), %%xmm8, %%xmm7")
665 "vcmpss $0xA, %%xmm6, %%xmm8, %%xmm7",
666 "vcmpss $0xA, (%%rax), %%xmm8, %%xmm7")
668 "vcmpss $0xC, %%xmm6, %%xmm8, %%xmm7",
669 "vcmpss $0xC, (%%rax), %%xmm8, %%xmm7")
671 "vcmpss $0xD, %%xmm6, %%xmm8, %%xmm7",
672 "vcmpss $0xD, (%%rax), %%xmm8, %%xmm7")
674 "vcmpss $0xE, %%xmm6, %%xmm8, %%xmm7",
675 "vcmpss $0xE, (%%rax), %%xmm8, %%xmm7")
677 "vcmpss $0x10, %%xmm6, %%xmm8, %%xmm7",
678 "vcmpss $0x10, (%%rax), %%xmm8, %%xmm7")
680 "vcmpss $0x11, %%xmm6, %%xmm8, %%xmm7",
681 "vcmpss $0x11, (%%rax), %%xmm8, %%xmm7")
683 "vcmpss $0x12, %%xmm6, %%xmm8, %%xmm7",
684 "vcmpss $0x12, (%%rax), %%xmm8, %%xmm7")
686 "vcmpss $0x13, %%xmm6, %%xmm8, %%xmm7",
687 "vcmpss $0x13, (%%rax), %%xmm8, %%xmm7")
689 "vcmpss $0x14, %%xmm6, %%xmm8, %%xmm7",
690 "vcmpss $0x14, (%%rax), %%xmm8, %%xmm7")
692 "vcmpss $0x15, %%xmm6, %%xmm8, %%xmm7",
693 "vcmpss $0x15, (%%rax), %%xmm8, %%xmm7")
695 "vcmpss $0x16, %%xmm6, %%xmm8, %%xmm7",
696 "vcmpss $0x16, (%%rax), %%xmm8, %%xmm7")
698 "vcmpss $0x17, %%xmm6, %%xmm8, %%xmm7",
699 "vcmpss $0x17, (%%rax), %%xmm8, %%xmm7")
701 "vcmpss $0x18, %%xmm6, %%xmm8, %%xmm7",
702 "vcmpss $0x18, (%%rax), %%xmm8, %%xmm7")
704 "vcmpss $0x19, %%xmm6, %%xmm8, %%xmm7",
705 "vcmpss $0x19, (%%rax), %%xmm8, %%xmm7")
707 "vcmpss $0x1A, %%xmm6, %%xmm8, %%xmm7",
708 "vcmpss $0x1A, (%%rax), %%xmm8, %%xmm7")
710 "vcmpss $0x1C, %%xmm6, %%xmm8, %%xmm7",
711 "vcmpss $0x1C, (%%rax), %%xmm8, %%xmm7")
713 "vcmpss $0x1D, %%xmm6, %%xmm8, %%xmm7",
714 "vcmpss $0x1D, (%%rax), %%xmm8, %%xmm7")
716 "vcmpss $0x1E, %%xmm6, %%xmm8, %%xmm7",
717 "vcmpss $0x1E, (%%rax), %%xmm8, %%xmm7")
721 "vcvtpd2psx %%xmm8, %%xmm7",
749 "vpcmpeqd %%xmm6, %%xmm8, %%xmm7",
750 "vpcmpeqd (%%rax), %%xmm8, %%xmm7")
753 "vpshufd $0x39, %%xmm9, %%xmm8",
757 "vmaxsd %%xmm6, %%xmm8, %%xmm7",
758 "vmaxsd (%%rax), %%xmm8, %%xmm7")
761 "vdivsd %%xmm6, %%xmm8, %%xmm7",
762 "vdivsd (%%rax), %%xmm8, %%xmm7")
765 "vminsd %%xmm6, %%xmm8, %%xmm7",
766 "vminsd (%%rax), %%xmm8, %%xmm7")
769 "vsubsd %%xmm6, %%xmm8, %%xmm7",
770 "vsubsd (%%rax), %%xmm8, %%xmm7")
773 "vaddsd %%xmm6, %%xmm8, %%xmm7",
774 "vaddsd (%%rax), %%xmm8, %%xmm7")
777 "vmulsd %%xmm6, %%xmm8, %%xmm7",
778 "vmulsd (%%rax), %%xmm8, %%xmm7")
781 "vxorps %%xmm6, %%xmm8, %%xmm7",
782 "vxorps (%%rax), %%xmm8, %%xmm7")
785 "vxorpd %%xmm6, %%xmm8, %%xmm7",
786 "vxorpd (%%rax), %%xmm8, %%xmm7")
789 "vorpd %%xmm6, %%xmm8, %%xmm7",
790 "vorpd (%%rax), %%xmm8, %%xmm7")
793 "vandnpd %%xmm6, %%xmm8, %%xmm7",
794 "vandnpd (%%rax), %%xmm8, %%xmm7")
797 "vcvtps2pd %%xmm6, %%xmm8",
798 "vcvtps2pd (%%rax), %%xmm8")
801 "vucomisd %%xmm6, %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14",
802 "vucomisd (%%rax), %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14")
805 "vucomiss %%xmm6, %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14",
806 "vucomiss (%%rax), %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14")
809 "vpinsrq $0, %%r14, %%xmm8, %%xmm7",
810 "vpinsrq $1, (%%rax), %%xmm8, %%xmm7")
813 "vpaddq %%xmm6, %%xmm8, %%xmm7",
814 "vpaddq (%%rax), %%xmm8, %%xmm7")
817 "vpsubq %%xmm6, %%xmm8, %%xmm7",
818 "vpsubq (%%rax), %%xmm8, %%xmm7")
821 "vpsubw %%xmm6, %%xmm8, %%xmm7",
822 "vpsubw (%%rax), %%xmm8, %%xmm7")
853 "vpcmpeqq %%xmm6, %%xmm8, %%xmm7",
854 "vpcmpeqq (%%rax), %%xmm8, %%xmm7")
857 "vsubpd %%xmm6, %%xmm8, %%xmm7",
858 "vsubpd (%%rax), %%xmm8, %%xmm7")
861 "vaddpd %%xmm6, %%xmm8, %%xmm7",
862 "vaddpd (%%rax), %%xmm8, %%xmm7")
865 "vunpcklpd %%xmm6, %%xmm8, %%xmm7",
866 "vunpcklpd (%%rax), %%xmm8, %%xmm7")
869 "vunpckhpd %%xmm6, %%xmm8, %%xmm7",
870 "vunpckhpd (%%rax), %%xmm8, %%xmm7")
873 "vunpckhps %%xmm6, %%xmm8, %%xmm7",
874 "vunpckhps (%%rax), %%xmm8, %%xmm7")
877 "vmovups %%xmm6, %%xmm8",
897 "vpcmpgtq %%xmm6, %%xmm8, %%xmm7",
898 "vpcmpgtq (%%rax), %%xmm8, %%xmm7")
912 "vpmuludq %%xmm6, %%xmm8, %%xmm7",
913 "vpmuludq (%%rax), %%xmm8, %%xmm7")
916 "vpmuldq %%xmm6, %%xmm8, %%xmm7",
917 "vpmuldq (%%rax), %%xmm8, %%xmm7")
923 "vpmaxud %%xmm6, %%xmm8, %%xmm7",
924 "vpmaxud (%%rax), %%xmm8, %%xmm7")
927 "vpminud %%xmm6, %%xmm8, %%xmm7",
928 "vpminud (%%rax), %%xmm8, %%xmm7")
931 "vpmulld %%xmm6, %%xmm8, %%xmm7",
932 "vpmulld (%%rax), %%xmm8, %%xmm7")
935 "vpmaxuw %%xmm6, %%xmm8, %%xmm7",
936 "vpmaxuw (%%rax), %%xmm8, %%xmm7")
945 "vpminuw %%xmm6, %%xmm8, %%xmm7",
946 "vpminuw (%%rax), %%xmm8, %%xmm7")
949 "vphminposuw %%xmm6, %%xmm8",
953 "vpmaxsw %%xmm6, %%xmm8, %%xmm7",
954 "vpmaxsw (%%rax), %%xmm8, %%xmm7")
957 "vpminsw %%xmm6, %%xmm8, %%xmm7",
958 "vpminsw (%%rax), %%xmm8, %%xmm7")
961 xmm8, %%xmm7",
962 "vpmaxub (%%rax), %%xmm8, %%xmm7")
965 "vpextrb $0x0, %%xmm8, %%r14",
966 "vpextrb $0x0, %%xmm8, (%%rax)")
969 "vpextrb $0x1, %%xmm8, %%r14",
970 "vpextrb $0x1, %%xmm8, (%%rax)")
973 "vpextrb $0x2, %%xmm8, %%r14",
974 "vpextrb $0x2, %%xmm8, (%%rax)")
977 "vpextrb $0x3, %%xmm8, %%r14",
978 "vpextrb $0x3, %%xmm8, (%%rax)")
981 "vpextrb $0x4, %%xmm8, %%r14",
982 "vpextrb $0x4, %%xmm8, (%%rax)")
985 "vpextrb $0x9, %%xmm8, %%r14",
986 "vpextrb $0x9, %%xmm8, (%%rax)")
989 "vpextrb $0xE, %%xmm8, %%r14",
990 "vpextrb $0xE, %%xmm8, (%%rax)")
993 "vpextrb $0xF, %%xmm8, %%r14",
994 "vpextrb $0xF, %%xmm8, (%%rax)")
997 "vpminub %%xmm6, %%xmm8, %%xmm7",
998 "vpminub (%%rax), %%xmm8, %%xmm7")
1001 "vpmaxsb %%xmm6, %%xmm8, %%xmm7",
1002 "vpmaxsb (%%rax), %%xmm8, %%xmm7")
1005 "vpminsb %%xmm6, %%xmm8, %%xmm7",
1006 "vpminsb (%%rax), %%xmm8, %%xmm7")
1047 "vpermilpd $0x0, %%xmm6, %%xmm8",
1048 "vpermilpd $0x1, (%%rax), %%xmm8")
1050 "vpermilpd $0x3, %%xmm6, %%xmm8",
1051 "vpermilpd $0x2, (%%rax), %%xmm8")
1081 "vcmppd $4, %%xmm6, %%xmm8, %%xmm7",
1082 "vcmppd $4, (%%rax), %%xmm8, %%xmm7")
1089 "vcmpps $4, %%xmm6, %%xmm8, %%xmm7",
1090 "vcmpps $4, (%%rax), %%xmm8, %%xmm7")
1097 "vcvtdq2pd %%xmm6, %%xmm8",
1098 "vcvtdq2pd (%%rax), %%xmm8")
1101 "vdivpd %%xmm6, %%xmm8, %%xmm7",
1102 "vdivpd (%%rax), %%xmm8, %%xmm7")
1109 "vpmovsxbw %%xmm6, %%xmm8",
1110 "vpmovsxbw (%%rax), %%xmm8")
1113 "vpsubusw %%xmm9, %%xmm8, %%xmm7",
1114 "vpsubusw (%%rax), %%xmm8, %%xmm7")
1117 "vpsubsw %%xmm9, %%xmm8, %%xmm7",
1118 "vpsubsw (%%rax), %%xmm8, %%xmm7")
1121 "vpcmpeqw %%xmm6, %%xmm8, %%xmm7",
1122 "vpcmpeqw (%%rax), %%xmm8, %%xmm7")
1125 "vpaddb %%xmm6, %%xmm8, %%xmm7",
1126 "vpaddb (%%rax), %%xmm8, %%xmm7")
1137 "vmovhpd (%%rax), %%xmm8, %%xmm7")
1140 "vmovhps (%%rax), %%xmm8, %%xmm7")
1148 "vpunpckhdq %%xmm6, %%xmm8, %%xmm7",
1149 "vpunpckhdq (%%rax), %%xmm8, %%xmm7")
1152 "vbroadcastss (%%rax), %%xmm8")
1155 "vpmovsxdq %%xmm6, %%xmm8",
1156 "vpmovsxdq (%%rax), %%xmm8")
1159 "vpmovsxwd %%xmm6, %%xmm8",
1160 "vpmovsxwd (%%rax), %%xmm8")
1163 "vdivps %%xmm9, %%xmm8, %%xmm7",
1164 "vdivps (%%rax), %%xmm8, %%xmm7")
1204 "vpmulhw %%xmm9, %%xmm8, %%xmm7",
1205 "vpmulhw (%%rax), %%xmm8, %%xmm7")
1208 "vpunpckhqdq %%xmm6, %%xmm8, %%xmm7",
1209 "vpunpckhqdq (%%rax), %%xmm8, %%xmm7")
1215 "vpcmpgtb %%xmm6, %%xmm8, %%xmm7",
1216 "vpcmpgtb (%%rax), %%xmm8, %%xmm7")
1219 "vpcmpgtw %%xmm6, %%xmm8, %%xmm7",
1220 "vpcmpgtw (%%rax), %%xmm8, %%xmm7")
1223 "vpcmpgtd %%xmm6, %%xmm8, %%xmm7",
1224 "vpcmpgtd (%%rax), %%xmm8, %%xmm7")
1227 "vpmovzxbd %%xmm6, %%xmm8",
1228 "vpmovzxbd (%%rax), %%xmm8")
1231 "vpmovsxbd %%xmm6, %%xmm8",
1232 "vpmovsxbd (%%rax), %%xmm8")
1235 "vpinsrb $0, %%r14d, %%xmm8, %%xmm7",
1236 "vpinsrb $3, (%%rax), %%xmm8, %%xmm7")
1238 "vpinsrb $6, %%r14d, %%xmm8, %%xmm7",
1239 "vpinsrb $9, (%%rax), %%xmm8, %%xmm7")
1241 "vpinsrb $12, %%r14d, %%xmm8, %%xmm7",
1242 "vpinsrb $15, (%%rax), %%xmm8, %%xmm7")
1245 "vpinsrw $0, %%r14d, %%xmm8, %%xmm7",
1246 "vpinsrw $3, (%%rax), %%xmm8, %%xmm7")
1248 "vpinsrw $2, %%r14d, %%xmm8, %%xmm7",
1249 "vpinsrw $3, (%%rax), %%xmm8, %%xmm7")
1251 "vpinsrw $4, %%r14d, %%xmm8, %%xmm7",
1252 "vpinsrw $5, (%%rax), %%xmm8, %%xmm7")
1254 "vpinsrw $6, %%r14d, %%xmm8, %%xmm7",
1255 "vpinsrw $7, (%%rax), %%xmm8, %%xmm7")
1258 "vcomisd %%xmm6, %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14",
1259 "vcomisd (%%rax), %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14")
1262 "vcomiss %%xmm6, %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14",
1263 "vcomiss (%%rax), %%xmm8; pushfq; popq %%r14; andq $0x8D5, %%r14")
1270 "vdppd $0x00, %%xmm6, %%xmm8, %%xmm7",
1273 "vdppd $0x5A, %%xmm6, %%xmm8, %%xmm7",
1276 "vdppd $0x0F, %%xmm6, %%xmm8, %%xmm7",
1279 "vdppd $0xF0, %%xmm6, %%xmm8, %%xmm7",
1283 "vdpps $0x00, %%xmm6, %%xmm8, %%xmm7",
1286 "vdpps $0x5A, %%xmm6, %%xmm8, %%xmm7",
1289 "vdpps $0x0F, %%xmm6, %%xmm8, %%xmm7",
1292 "vdpps $0xF0, %%xmm6, %%xmm8, %%xmm7",
1312 "vpalignr $0, %%xmm6, %%xmm8, %%xmm7",
1313 "vpalignr $3, (%%rax), %%xmm8, %%xmm7")
1315 "vpalignr $6, %%xmm6, %%xmm8, %%xmm7",
1316 "vpalignr $9, (%%rax), %%xmm8, %%xmm7")
1318 "vpalignr $12, %%xmm6, %%xmm8, %%xmm7",
1319 "vpalignr $15, (%%rax), %%xmm8, %%xmm7")
1321 GEN_test_Ronly(VMOVSD_REG_XMM, "vmovsd %%xmm9, %%xmm7, %%xmm8")
1323 GEN_test_Ronly(VMOVSS_REG_XMM, "vmovss %%xmm9, %%xmm7, %%xmm8")
1325 GEN_test_Monly(VMOVLPD_128_M64_XMM_XMM, "vmovlpd (%%rax), %%xmm8, %%xmm7")
1330 "vshufpd $0, %%xmm9, %%xmm8, %%xmm7",
1331 "vshufpd $1, (%%rax), %%xmm8, %%xmm7")
1333 "vshufpd $2, %%xmm9, %%xmm8, %%xmm7",
1334 "vshufpd $3, (%%rax), %%xmm8, %%xmm7")
1344 "vpermilps $0x00, %%xmm6, %%xmm8",
1345 "vpermilps $0x01, (%%rax), %%xmm8")
1347 "vpermilps $0xFE, %%xmm6, %%xmm8",
1348 "vpermilps $0xFF, (%%rax), %%xmm8")
1350 "vpermilps $0x30, %%xmm6, %%xmm8",
1351 "vpermilps $0x03, (%%rax), %%xmm8")
1353 "vpermilps $0x21, %%xmm6, %%xmm8",
1354 "vpermilps $0x12, (%%rax), %%xmm8")
1356 "vpermilps $0xD7, %%xmm6, %%xmm8",
1357 "vpermilps $0x6C, (%%rax), %%xmm8")
1359 "vpermilps $0xB5, %%xmm6, %%xmm8",
1360 "vpermilps $0x4A, (%%rax), %%xmm8")
1362 "vpermilps $0x85, %%xmm6, %%xmm8",
1363 "vpermilps $0xDC, (%%rax), %%xmm8")
1365 "vpermilps $0x29, %%xmm6, %%xmm8",
1366 "vpermilps $0x92, (%%rax), %%xmm8")
1369 "vblendps $0, %%xmm6, %%xmm8, %%xmm7",
1370 "vblendps $3, (%%rax), %%xmm8, %%xmm7")
1372 "vblendps $6, %%xmm6, %%xmm8, %%xmm7",
1373 "vblendps $9, (%%rax), %%xmm8, %%xmm7")
1375 "vblendps $12, %%xmm6, %%xmm8, %%xmm7",
1376 "vblendps $15, (%%rax), %%xmm8, %%xmm7")
1379 "vblendpd $0, %%xmm6, %%xmm8, %%xmm7",
1380 "vblendpd $1, (%%rax), %%xmm8, %%xmm7")
1382 "vblendpd $2, %%xmm6, %%xmm8, %%xmm7",
1383 "vblendpd $3, (%%rax), %%xmm8, %%xmm7")
1396 "vpblendw $0x00, %%xmm6, %%xmm8, %%xmm7",
1397 "vpblendw $0x01, (%%rax), %%xmm8, %%xmm7")
1399 "vpblendw $0xFE, %%xmm6, %%xmm8, %%xmm7",
1400 "vpblendw $0xFF, (%%rax), %%xmm8, %%xmm7")
1402 "vpblendw $0x30, %%xmm6, %%xmm8, %%xmm7",
1403 "vpblendw $0x03, (%%rax), %%xmm8, %%xmm7")
1405 "vpblendw $0x21, %%xmm6, %%xmm8, %%xmm7",
1406 "vpblendw $0x12, (%%rax), %%xmm8, %%xmm7")
1408 "vpblendw $0xD7, %%xmm6, %%xmm8, %%xmm7",
1409 "vpblendw $0x6C, (%%rax), %%xmm8, %%xmm7")
1411 "vpblendw $0xB5, %%xmm6, %%xmm8, %%xmm7",
1412 "vpblendw $0x4A, (%%rax), %%xmm8, %%xmm7")
1414 "vpblendw $0x85, %%xmm6, %%xmm8, %%xmm7",
1415 "vpblendw $0xDC, (%%rax), %%xmm8, %%xmm7")
1417 "vpblendw $0x29, %%xmm6, %%xmm8, %%xmm7",
1418 "vpblendw $0x92, (%%rax), %%xmm8, %%xmm7")
1425 "vsqrtss %%xmm6, %%xmm8, %%xmm7",
1426 "vsqrtss (%%rax), %%xmm8, %%xmm7")
1429 "vsqrtps %%xmm6, %%xmm8",
1430 "vsqrtps (%%rax), %%xmm8")
1437 "vsqrtpd %%xmm6, %%xmm8",
1438 "vsqrtpd (%%rax), %%xmm8")
1445 "vrsqrtss %%xmm6, %%xmm8, %%xmm7",
1446 "vrsqrtss (%%rax), %%xmm8, %%xmm7")
1449 "vrsqrtps %%xmm6, %%xmm8",
1450 "vrsqrtps (%%rax), %%xmm8")
1513 "vpermilps %%xmm6, %%xmm8, %%xmm7",
1514 "vpermilps (%%rax), %%xmm8, %%xmm7")
1517 "vpermilpd %%xmm6, %%xmm8, %%xmm7",
1518 "vpermilpd (%%rax), %%xmm8, %%xmm7")
1531 "vpsllw %%xmm6, %%xmm8, %%xmm9",
1533 "vpsllw 128(%%rax), %%xmm8, %%xmm9")
1538 "vpsrlw %%xmm6, %%xmm8, %%xmm9",
1540 "vpsrlw 128(%%rax), %%xmm8, %%xmm9")
1545 "vpsraw %%xmm6, %%xmm8, %%xmm9",
1547 "vpsraw 128(%%rax), %%xmm8, %%xmm9")
1552 "vpslld %%xmm6, %%xmm8, %%xmm9",
1554 "vpslld 128(%%rax), %%xmm8, %%xmm9")
1559 "vpsrld %%xmm6, %%xmm8, %%xmm9",
1561 "vpsrld 128(%%rax), %%xmm8, %%xmm9")
1566 "vpsrad %%xmm6, %%xmm8, %%xmm9",
1568 "vpsrad 128(%%rax), %%xmm8, %%xmm9")
1573 "vpsllq %%xmm6, %%xmm8, %%xmm9",
1575 "vpsllq 128(%%rax), %%xmm8, %%xmm9")
1580 "vpsrlq %%xmm6, %%xmm8, %%xmm9",
1582 "vpsrlq 128(%%rax), %%xmm8, %%xmm9")
1585 "vroundps $0x0, %%xmm8, %%xmm9",
1588 "vroundps $0x1, %%xmm8, %%xmm9",
1591 "vroundps $0x2, %%xmm8, %%xmm9",
1594 "vroundps $0x3, %%xmm8, %%xmm9",
1597 "vroundps $0x4, %%xmm8, %%xmm9",
1617 "vroundpd $0x0, %%xmm8, %%xmm9",
1620 "vroundpd $0x1, %%xmm8, %%xmm9",
1623 "vroundpd $0x2, %%xmm8, %%xmm9",
1626 "vroundpd $0x3, %%xmm8, %%xmm9",
1629 "vroundpd $0x4, %%xmm8, %%xmm9",
1649 "vpmaddwd %%xmm6, %%xmm8, %%xmm7",
1650 "vpmaddwd (%%rax), %%xmm8, %%xmm7")
1653 "vaddsubps %%xmm6, %%xmm8, %%xmm7",
1654 "vaddsubps (%%rax), %%xmm8, %%xmm7")
1661 "vaddsubpd %%xmm6, %%xmm8, %%xmm7",
1662 "vaddsubpd (%%rax), %%xmm8, %%xmm7")
1669 "vroundss $0x0, %%xmm8, %%xmm6, %%xmm9",
1672 "vroundss $0x1, %%xmm8, %%xmm6, %%xmm9",
1675 "vroundss $0x2, %%xmm8, %%xmm6, %%xmm9",
1678 "vroundss $0x3, %%xmm8, %%xmm6, %%xmm9",
1681 "vroundss $0x4, %%xmm8, %%xmm6, %%xmm9",
1684 "vroundss $0x5, %%xmm8, %%xmm6, %%xmm9",
1688 "vroundsd $0x0, %%xmm8, %%xmm6, %%xmm9",
1691 "vroundsd $0x1, %%xmm8, %%xmm6, %%xmm9",
1694 "vroundsd $0x2, %%xmm8, %%xmm6, %%xmm9",
1697 "vroundsd $0x3, %%xmm8, %%xmm6, %%xmm9",
1700 "vroundsd $0x4, %%xmm8, %%xmm6, %%xmm9",
1703 "vroundsd $0x5, %%xmm8, %%xmm6, %%xmm9",
1707 "vptest %%xmm6, %%xmm8; "
1709 "vptest (%%rax), %%xmm8; "
1716 "vmovups %%xmm6, %%xmm8;"
1717 "vptest %%xmm6, %%xmm8; "
1719 "vmovups (%%rax), %%xmm8;"
1720 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1721 "vxorpd %%xmm8,%%xmm7,%%xmm8;"
1722 "vptest (%%rax), %%xmm8; "
1739 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1754 "vtestps %%xmm6, %%xmm8; "
1756 "vtestps (%%rax), %%xmm8; "
1763 "vmovups %%xmm6, %%xmm8;"
1764 "vtestps %%xmm6, %%xmm8; "
1766 "vmovups (%%rax), %%xmm8;"
1767 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1768 "vxorpd %%xmm8,%%xmm7,%%xmm8;"
1769 "vtestps (%%rax), %%xmm8; "
1773 "vtestps %%xmm8, %%xmm9; "
1795 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1814 "vtestpd %%xmm6, %%xmm8; "
1816 "vtestpd (%%rax), %%xmm8; "
1823 "vmovups %%xmm6, %%xmm8;"
1824 "vtestpd %%xmm6, %%xmm8; "
1826 "vmovups (%%rax), %%xmm8;"
1827 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1828 "vxorpd %%xmm8,%%xmm7,%%xmm8;"
1829 "vtestpd (%%rax), %%xmm8; "
1833 "vtestpd %%xmm8, %%xmm9; "
1855 "vcmpeqpd %%xmm8,%%xmm8,%%xmm7;"
1872 "vblendvps %%xmm9, %%xmm6, %%xmm8, %%xmm7",
1873 "vblendvps %%xmm9, (%%rax), %%xmm8, %%xmm7")
1880 "vblendvpd %%xmm9, %%xmm6, %%xmm8, %%xmm7",
1881 "vblendvpd %%xmm9, (%%rax), %%xmm8, %%xmm7")
1889 "vhaddps %%xmm6, %%xmm8, %%xmm7",
1890 "vhaddps (%%rax), %%xmm8, %%xmm7")
1897 "vhaddpd %%xmm6, %%xmm8, %%xmm7",
1898 "vhaddpd (%%rax), %%xmm8, %%xmm7")
1905 "vhsubps %%xmm6, %%xmm8, %%xmm7",
1906 "vhsubps (%%rax), %%xmm8, %%xmm7")
1913 "vhsubpd %%xmm6, %%xmm8, %%xmm7",
1914 "vhsubpd (%%rax), %%xmm8, %%xmm7")
1921 "vextractps $0, %%xmm8, %%r14d",
1922 "vextractps $0, %%xmm8, (%%rax)")
1925 "vextractps $1, %%xmm8, %%r14d",
1926 "vextractps $1, %%xmm8, (%%rax)")
1929 "vextractps $2, %%xmm8, %%r14d",
1930 "vextractps $2, %%xmm8, (%%rax)")
1933 "vextractps $3, %%xmm8, %%r14d",
1934 "vextractps $3, %%xmm8, (%%rax)")
1937 "vlddqu 1(%%rax), %%xmm8")
1947 "vmaskmovdqu %%xmm8, %%xmm9;"
1975 "vpacksswb %%xmm6, %%xmm8, %%xmm7",
1976 "vpacksswb (%%rax), %%xmm8, %%xmm7")
1979 "vpavgb %%xmm6, %%xmm8, %%xmm7",
1980 "vpavgb (%%rax), %%xmm8, %%xmm7")
1983 "vpavgw %%xmm6, %%xmm8, %%xmm7",
1984 "vpavgw (%%rax), %%xmm8, %%xmm7")
1987 "vpaddsb %%xmm6, %%xmm8, %%xmm7",
1988 "vpaddsb (%%rax), %%xmm8, %%xmm7")
1991 "vpaddsw %%xmm6, %%xmm8, %%xmm7",
1992 "vpaddsw (%%rax), %%xmm8, %%xmm7")
1995 "vphaddw %%xmm6, %%xmm8, %%xmm7",
1996 "vphaddw (%%rax), %%xmm8, %%xmm7")
1999 "vphaddd %%xmm6, %%xmm8, %%xmm7",
2000 "vphaddd (%%rax), %%xmm8, %%xmm7")
2003 "vphaddsw %%xmm6, %%xmm8, %%xmm7",
2004 "vphaddsw (%%rax), %%xmm8, %%xmm7")
2007 "vpmaddubsw %%xmm6, %%xmm8, %%xmm7",
2008 "vpmaddubsw (%%rax), %%xmm8, %%xmm7")
2011 "vphsubw %%xmm6, %%xmm8, %%xmm7",
2012 "vphsubw (%%rax), %%xmm8, %%xmm7")
2015 "vphsubd %%xmm6, %%xmm8, %%xmm7",
2016 "vphsubd (%%rax), %%xmm8, %%xmm7")
2019 "vphsubsw %%xmm6, %%xmm8, %%xmm7",
2020 "vphsubsw (%%rax), %%xmm8, %%xmm7")
2031 "vpmovsxbq %%xmm6, %%xmm8",
2032 "vpmovsxbq (%%rax), %%xmm8")
2035 "vpmovsxwq %%xmm6, %%xmm8",
2036 "vpmovsxwq (%%rax), %%xmm8")
2039 "vpackusdw %%xmm6, %%xmm8, %%xmm7",
2040 "vpackusdw (%%rax), %%xmm8, %%xmm7")
2043 "vpmovzxbq %%xmm6, %%xmm8",
2044 "vpmovzxbq (%%rax), %%xmm8")
2047 "vpmovzxwq %%xmm6, %%xmm8",
2048 "vpmovzxwq (%%rax), %%xmm8")
2051 "vpmovzxdq %%xmm6, %%xmm8",
2052 "vpmovzxdq (%%rax), %%xmm8")
2055 "vmpsadbw $0, %%xmm6, %%xmm8, %%xmm7",
2056 "vmpsadbw $0, (%%rax), %%xmm8, %%xmm7")
2058 "vmpsadbw $1, %%xmm6, %%xmm8, %%xmm7",
2059 "vmpsadbw $1, (%%rax), %%xmm8, %%xmm7")
2061 "vmpsadbw $2, %%xmm6, %%xmm8, %%xmm7",
2062 "vmpsadbw $2, (%%rax), %%xmm8, %%xmm7")
2064 "vmpsadbw $3, %%xmm6, %%xmm8, %%xmm7",
2065 "vmpsadbw $3, (%%rax), %%xmm8, %%xmm7")
2067 "vmpsadbw $4, %%xmm6, %%xmm8, %%xmm7",
2068 "vmpsadbw $4, (%%rax), %%xmm8, %%xmm7")
2070 "vmpsadbw $5, %%xmm6, %%xmm8, %%xmm7",
2071 "vmpsadbw $5, (%%rax), %%xmm8, %%xmm7")
2073 "vmpsadbw $6, %%xmm6, %%xmm8, %%xmm7",
2074 "vmpsadbw $6, (%%rax), %%xmm8, %%xmm7")
2076 "vmpsadbw $7, %%xmm6, %%xmm8, %%xmm7",
2077 "vmpsadbw $7, (%%rax), %%xmm8, %%xmm7")
2083 GEN_test_Monly(VMOVLPS_128_M64_XMM_XMM, "vmovlps (%%rax), %%xmm8, %%xmm7")
2088 "vrcpss %%xmm6, %%xmm8, %%xmm7",
2089 "vrcpss (%%rax), %%xmm8, %%xmm7")
2092 "vrcpps %%xmm6, %%xmm8",
2093 "vrcpps (%%rax), %%xmm8")
2100 "vpsadbw %%xmm6, %%xmm8, %%xmm7",
2101 "vpsadbw (%%rax), %%xmm8, %%xmm7")
2104 "vpsignb %%xmm6, %%xmm8, %%xmm7",
2105 "vpsignb (%%rax), %%xmm8, %%xmm7")
2108 "vpsignw %%xmm6, %%xmm8, %%xmm7",
2109 "vpsignw (%%rax), %%xmm8, %%xmm7")
2112 "vpsignd %%xmm6, %%xmm8, %%xmm7",
2113 "vpsignd (%%rax), %%xmm8, %%xmm7")
2116 "vpmulhrsw %%xmm6, %%xmm8, %%xmm7",
2117 "vpmulhrsw (%%rax), %%xmm8, %%xmm7")
2148 "vaesenc %%xmm6, %%xmm8, %%xmm7",
2149 "vaesenc (%%rax), %%xmm8, %%xmm7")
2152 "vaesenclast %%xmm6, %%xmm8, %%xmm7",
2153 "vaesenclast (%%rax), %%xmm8, %%xmm7")
2156 "vaesdec %%xmm6, %%xmm8, %%xmm7",
2157 "vaesdec (%%rax), %%xmm8, %%xmm7")
2160 "vaesdeclast %%xmm6, %%xmm8, %%xmm7",
2161 "vaesdeclast (%%rax), %%xmm8, %%xmm7")
2181 "vpclmulqdq $0x00, %%xmm6, %%xmm8, %%xmm7",
2182 "vpclmulqdq $0x00, (%%rax), %%xmm8, %%xmm7")
2184 "vpclmulqdq $0x01, %%xmm6, %%xmm8, %%xmm7",
2185 "vpclmulqdq $0x01, (%%rax), %%xmm8, %%xmm7")
2187 "vpclmulqdq $0x10, %%xmm6, %%xmm8, %%xmm7",
2188 "vpclmulqdq $0x10, (%%rax), %%xmm8, %%xmm7")
2190 "vpclmulqdq $0x11, %%xmm6, %%xmm8, %%xmm7",
2191 "vpclmulqdq $0x11, (%%rax), %%xmm8, %%xmm7")
2193 "vpclmulqdq $0xFF, %%xmm6, %%xmm8, %%xmm7",
2194 "vpclmulqdq $0xFF, (%%rax), %%xmm8, %%xmm7")
2197 "vcmpss $0x9, %%xmm6, %%xmm8, %%xmm7",
2198 "vcmpss $0x9, (%%rax), %%xmm8, %%xmm7")
2201 "vmaskmovps (%%rax), %%xmm8, %%xmm7;"
2211 "vmaskmovpd (%%rax), %%xmm8, %%xmm7;"
2221 "vmaskmovps %%xmm8, %%xmm7, (%%rax);"
2231 "vmaskmovpd %%xmm8, %%xmm7, (%%rax);"