Home | History | Annotate | Download | only in amd64

Lines Matching refs:__volatile__

21       asm __volatile__("fxsave64 (%0)" : : "r" (p) : "memory" );
23 asm __volatile__("rex64/fxsave (%0)" : : "r" (p) : "memory" );
26 asm __volatile__("fxsave (%0)" : : "r" (p) : "memory" );
34 asm __volatile__("fxrstor64 (%0)" : : "r" (p) : "memory" );
36 asm __volatile__("rex64/fxrstor (%0)" : : "r" (p) : "memory" );
39 asm __volatile__("fxrstor (%0)" : : "r" (p) : "memory" );
45 asm __volatile__("finit");
46 asm __volatile__(
57 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm0");
58 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm1");
59 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm2");
60 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm3");
61 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm4");
62 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm5");
63 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm6");
64 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm7");
65 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm8");
66 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm9");
67 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm10");
68 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm11");
69 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm12");
70 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm13");
71 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm14");
72 asm __volatile__("movups " VG_SYM(vecZ) ", %xmm15");
74 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm0");
75 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm1");
76 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm2");
77 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm3");
78 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm4");
79 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm5");
80 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm6");
81 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm7");
82 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm8");
83 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm9");
84 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm10");
85 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm11");
86 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm12");
87 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm13");
88 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm14");
89 asm __volatile__("movups " VG_SYM(vecZ) "(%rip), %xmm15");
91 asm __volatile__(
100 asm __volatile__("finit");
101 asm __volatile__("fldpi");
102 asm __volatile__("fld1");
103 asm __volatile__("fldln2");
104 asm __volatile__("fldlg2");
105 asm __volatile__("fld %st(3)");
106 asm __volatile__("fld %st(3)");
107 asm __volatile__("fld1");
108 asm __volatile__("movups (%0), %%xmm0" : : "r"(&vec0[0]) : "xmm0" );
109 asm __volatile__("movups (%0), %%xmm1" : : "r"(&vec1[0]) : "xmm1" );
110 asm __volatile__("xorps %xmm2, %xmm2");
111 asm __volatile__("movaps %xmm0, %xmm3");
112 asm __volatile__("movaps %xmm1, %xmm4");
113 asm __volatile__("movaps %xmm2, %xmm5");
114 asm __volatile__("movaps %xmm0, %xmm6");
115 asm __volatile__("movaps %xmm1, %xmm7");
116 asm __volatile__("movaps %xmm1, %xmm8");
117 asm __volatile__("movaps %xmm2, %xmm9");
118 asm __volatile__("movaps %xmm0, %xmm10");
119 asm __volatile__("movaps %xmm1, %xmm11");
120 asm __volatile__("movaps %xmm1, %xmm12");
121 asm __volatile__("movaps %xmm2, %xmm13");
122 asm __volatile__("movaps %xmm0, %xmm14");
123 asm __volatile__("movaps %xmm1, %xmm15");