Home | History | Annotate | Download | only in amd64

Lines Matching full:movups

49    asm __volatile__("movups " VG_SYM(vecZ) ", %xmm0");
50 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm1");
51 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm2");
52 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm3");
53 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm4");
54 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm5");
55 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm6");
56 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm7");
57 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm8");
58 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm9");
59 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm10");
60 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm11");
61 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm12");
62 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm13");
63 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm14");
64 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm15");
66 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm0");
67 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm1");
68 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm2");
69 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm3");
70 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm4");
71 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm5");
72 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm6");
73 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm7");
74 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm8");
75 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm9");
76 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm10");
77 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm11");
78 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm12");
79 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm13");
80 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm14");
81 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm15");
100 asm __volatile__("movups (%0), %%xmm0" : : "r"(&vec0[0]) : "xmm0" );
101 asm __volatile__("movups (%0), %%xmm1" : : "r"(&vec1[0]) : "xmm1" );