Home | History | Annotate | Download | only in amd64

Lines Matching full:movups

57    asm __volatile__("movups " VG_SYM(vecZ) ", %xmm0");
58 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm1");
59 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm2");
60 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm3");
61 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm4");
62 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm5");
63 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm6");
64 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm7");
65 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm8");
66 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm9");
67 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm10");
68 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm11");
69 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm12");
70 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm13");
71 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm14");
72 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm15");
74 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm0");
75 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm1");
76 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm2");
77 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm3");
78 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm4");
79 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm5");
80 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm6");
81 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm7");
82 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm8");
83 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm9");
84 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm10");
85 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm11");
86 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm12");
87 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm13");
88 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm14");
89 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm15");
108 asm __volatile__("movups (%0), %%xmm0" : : "r"(&vec0[0]) : "xmm0" );
109 asm __volatile__("movups (%0), %%xmm1" : : "r"(&vec1[0]) : "xmm1" );