Home | History | Annotate | Download | only in priv

Lines Matching refs:gst

2049         ( VexGuestAMD64State* gst, HWord addr )
2069 do_get_x87( gst, &tmp );
2128 ( VexGuestAMD64State* gst, HWord addr )
2135 mxcsr = amd64g_create_mxcsr( gst->guest_SSEROUND );
2155 void LibVEX_GuestAMD64_fxsave ( /*IN*/VexGuestAMD64State* gst,
2159 amd64g_dirtyhelper_XSAVE_COMPONENT_0(gst, fp_state);
2162 amd64g_dirtyhelper_XSAVE_COMPONENT_1_EXCLUDING_XMMREGS(gst, fp_state);
2175 COPY_U128( xmm[0], gst->guest_YMM0 );
2176 COPY_U128( xmm[1], gst->guest_YMM1 );
2177 COPY_U128( xmm[2], gst->guest_YMM2 );
2178 COPY_U128( xmm[3], gst->guest_YMM3 );
2179 COPY_U128( xmm[4], gst->guest_YMM4 );
2180 COPY_U128( xmm[5], gst->guest_YMM5 );
2181 COPY_U128( xmm[6], gst->guest_YMM6 );
2182 COPY_U128( xmm[7], gst->guest_YMM7 );
2183 COPY_U128( xmm[8], gst->guest_YMM8 );
2184 COPY_U128( xmm[9], gst->guest_YMM9 );
2185 COPY_U128( xmm[10], gst->guest_YMM10 );
2186 COPY_U128( xmm[11], gst->guest_YMM11 );
2187 COPY_U128( xmm[12], gst->guest_YMM12 );
2188 COPY_U128( xmm[13], gst->guest_YMM13 );
2189 COPY_U128( xmm[14], gst->guest_YMM14 );
2190 COPY_U128( xmm[15], gst->guest_YMM15 );
2202 ( VexGuestAMD64State* gst, HWord addr )
2238 VexEmNote warnX87 = do_put_x87( True/*moveRegs*/, &tmp, gst );
2247 ( VexGuestAMD64State* gst, HWord addr )
2256 gst->guest_SSEROUND = w64 & 0xFFFFFFFFULL;
2271 /*MOD*/VexGuestAMD64State* gst )
2284 COPY_U128( gst->guest_YMM0, xmm[0] );
2285 COPY_U128( gst->guest_YMM1, xmm[1] );
2286 COPY_U128( gst->guest_YMM2, xmm[2] );
2287 COPY_U128( gst->guest_YMM3, xmm[3] );
2288 COPY_U128( gst->guest_YMM4, xmm[4] );
2289 COPY_U128( gst->guest_YMM5, xmm[5] );
2290 COPY_U128( gst->guest_YMM6, xmm[6] );
2291 COPY_U128( gst->guest_YMM7, xmm[7] );
2292 COPY_U128( gst->guest_YMM8, xmm[8] );
2293 COPY_U128( gst->guest_YMM9, xmm[9] );
2294 COPY_U128( gst->guest_YMM10, xmm[10] );
2295 COPY_U128( gst->guest_YMM11, xmm[11] );
2296 COPY_U128( gst->guest_YMM12, xmm[12] );
2297 COPY_U128( gst->guest_YMM13, xmm[13] );
2298 COPY_U128( gst->guest_YMM14, xmm[14] );
2299 COPY_U128( gst->guest_YMM15, xmm[15] );
2304 = amd64g_dirtyhelper_XRSTOR_COMPONENT_1_EXCLUDING_XMMREGS(gst, fp_state);
2306 = amd64g_dirtyhelper_XRSTOR_COMPONENT_0(gst, fp_state);
2322 void amd64g_dirtyhelper_FINIT ( VexGuestAMD64State* gst )
2325 gst->guest_FTOP = 0;
2327 gst->guest_FPTAG[i] = 0; /* empty */
2328 gst->guest_FPREG[i] = 0; /* IEEE754 64-bit zero */
2330 gst->guest_FPROUND = (ULong)Irrm_NEAREST;
2331 gst->guest_FC3210 = 0;
3896 VexGuestAMD64State* gst,
3910 V128* argL = (V128*)( ((UChar*)gst) + gstOffL );
3911 V128* argR = (V128*)( ((UChar*)gst) + gstOffR );
3984 gst->guest_YMM0[0] = resV.w32[0];
3985 gst->guest_YMM0[1] = resV.w32[1];
3986 gst->guest_YMM0[2] = resV.w32[2];
3987 gst->guest_YMM0[3] = resV.w32[3];
4238 VexGuestAMD64State* gst,
4244 V128* argD = (V128*)( ((UChar*)gst) + gstOffD );
4245 V128* argL = (V128*)( ((UChar*)gst) + gstOffL );
4246 V128* argR = (V128*)( ((UChar*)gst) + gstOffR );
4301 VexGuestAMD64State* gst,
4307 V128* argL = (V128*)( ((UChar*)gst) + gstOffL );
4308 V128* argR = (V128*)( ((UChar*)gst) + gstOffR );