1 /* 2 * This file was generated automatically by gen-mterp.py for 'x86'. 3 * 4 * --> DO NOT EDIT <-- 5 */ 6 7 /* File: c/header.cpp */ 8 /* 9 * Copyright (C) 2008 The Android Open Source Project 10 * 11 * Licensed under the Apache License, Version 2.0 (the "License"); 12 * you may not use this file except in compliance with the License. 13 * You may obtain a copy of the License at 14 * 15 * http://www.apache.org/licenses/LICENSE-2.0 16 * 17 * Unless required by applicable law or agreed to in writing, software 18 * distributed under the License is distributed on an "AS IS" BASIS, 19 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 20 * See the License for the specific language governing permissions and 21 * limitations under the License. 22 */ 23 24 /* common includes */ 25 #include "Dalvik.h" 26 #include "interp/InterpDefs.h" 27 #include "mterp/Mterp.h" 28 #include <math.h> // needed for fmod, fmodf 29 #include "mterp/common/FindInterface.h" 30 31 /* 32 * Configuration defines. These affect the C implementations, i.e. the 33 * portable interpreter(s) and C stubs. 34 * 35 * Some defines are controlled by the Makefile, e.g.: 36 * WITH_INSTR_CHECKS 37 * WITH_TRACKREF_CHECKS 38 * EASY_GDB 39 * NDEBUG 40 */ 41 42 #ifdef WITH_INSTR_CHECKS /* instruction-level paranoia (slow!) */ 43 # define CHECK_BRANCH_OFFSETS 44 # define CHECK_REGISTER_INDICES 45 #endif 46 47 /* 48 * Some architectures require 64-bit alignment for access to 64-bit data 49 * types. We can't just use pointers to copy 64-bit values out of our 50 * interpreted register set, because gcc may assume the pointer target is 51 * aligned and generate invalid code. 52 * 53 * There are two common approaches: 54 * (1) Use a union that defines a 32-bit pair and a 64-bit value. 55 * (2) Call memcpy(). 56 * 57 * Depending upon what compiler you're using and what options are specified, 58 * one may be faster than the other. For example, the compiler might 59 * convert a memcpy() of 8 bytes into a series of instructions and omit 60 * the call. The union version could cause some strange side-effects, 61 * e.g. for a while ARM gcc thought it needed separate storage for each 62 * inlined instance, and generated instructions to zero out ~700 bytes of 63 * stack space at the top of the interpreter. 64 * 65 * The default is to use memcpy(). The current gcc for ARM seems to do 66 * better with the union. 67 */ 68 #if defined(__ARM_EABI__) 69 # define NO_UNALIGN_64__UNION 70 #endif 71 72 73 //#define LOG_INSTR /* verbose debugging */ 74 /* set and adjust ANDROID_LOG_TAGS='*:i jdwp:i dalvikvm:i dalvikvmi:i' */ 75 76 /* 77 * Export another copy of the PC on every instruction; this is largely 78 * redundant with EXPORT_PC and the debugger code. This value can be 79 * compared against what we have stored on the stack with EXPORT_PC to 80 * help ensure that we aren't missing any export calls. 81 */ 82 #if WITH_EXTRA_GC_CHECKS > 1 83 # define EXPORT_EXTRA_PC() (self->currentPc2 = pc) 84 #else 85 # define EXPORT_EXTRA_PC() 86 #endif 87 88 /* 89 * Adjust the program counter. "_offset" is a signed int, in 16-bit units. 90 * 91 * Assumes the existence of "const u2* pc" and "const u2* curMethod->insns". 92 * 93 * We don't advance the program counter until we finish an instruction or 94 * branch, because we do want to have to unroll the PC if there's an 95 * exception. 96 */ 97 #ifdef CHECK_BRANCH_OFFSETS 98 # define ADJUST_PC(_offset) do { \ 99 int myoff = _offset; /* deref only once */ \ 100 if (pc + myoff < curMethod->insns || \ 101 pc + myoff >= curMethod->insns + dvmGetMethodInsnsSize(curMethod)) \ 102 { \ 103 char* desc; \ 104 desc = dexProtoCopyMethodDescriptor(&curMethod->prototype); \ 105 ALOGE("Invalid branch %d at 0x%04x in %s.%s %s", \ 106 myoff, (int) (pc - curMethod->insns), \ 107 curMethod->clazz->descriptor, curMethod->name, desc); \ 108 free(desc); \ 109 dvmAbort(); \ 110 } \ 111 pc += myoff; \ 112 EXPORT_EXTRA_PC(); \ 113 } while (false) 114 #else 115 # define ADJUST_PC(_offset) do { \ 116 pc += _offset; \ 117 EXPORT_EXTRA_PC(); \ 118 } while (false) 119 #endif 120 121 /* 122 * If enabled, log instructions as we execute them. 123 */ 124 #ifdef LOG_INSTR 125 # define ILOGD(...) ILOG(LOG_DEBUG, __VA_ARGS__) 126 # define ILOGV(...) ILOG(LOG_VERBOSE, __VA_ARGS__) 127 # define ILOG(_level, ...) do { \ 128 char debugStrBuf[128]; \ 129 snprintf(debugStrBuf, sizeof(debugStrBuf), __VA_ARGS__); \ 130 if (curMethod != NULL) \ 131 ALOG(_level, LOG_TAG"i", "%-2d|%04x%s", \ 132 self->threadId, (int)(pc - curMethod->insns), debugStrBuf); \ 133 else \ 134 ALOG(_level, LOG_TAG"i", "%-2d|####%s", \ 135 self->threadId, debugStrBuf); \ 136 } while(false) 137 void dvmDumpRegs(const Method* method, const u4* framePtr, bool inOnly); 138 # define DUMP_REGS(_meth, _frame, _inOnly) dvmDumpRegs(_meth, _frame, _inOnly) 139 static const char kSpacing[] = " "; 140 #else 141 # define ILOGD(...) ((void)0) 142 # define ILOGV(...) ((void)0) 143 # define DUMP_REGS(_meth, _frame, _inOnly) ((void)0) 144 #endif 145 146 /* get a long from an array of u4 */ 147 static inline s8 getLongFromArray(const u4* ptr, int idx) 148 { 149 #if defined(NO_UNALIGN_64__UNION) 150 union { s8 ll; u4 parts[2]; } conv; 151 152 ptr += idx; 153 conv.parts[0] = ptr[0]; 154 conv.parts[1] = ptr[1]; 155 return conv.ll; 156 #else 157 s8 val; 158 memcpy(&val, &ptr[idx], 8); 159 return val; 160 #endif 161 } 162 163 /* store a long into an array of u4 */ 164 static inline void putLongToArray(u4* ptr, int idx, s8 val) 165 { 166 #if defined(NO_UNALIGN_64__UNION) 167 union { s8 ll; u4 parts[2]; } conv; 168 169 ptr += idx; 170 conv.ll = val; 171 ptr[0] = conv.parts[0]; 172 ptr[1] = conv.parts[1]; 173 #else 174 memcpy(&ptr[idx], &val, 8); 175 #endif 176 } 177 178 /* get a double from an array of u4 */ 179 static inline double getDoubleFromArray(const u4* ptr, int idx) 180 { 181 #if defined(NO_UNALIGN_64__UNION) 182 union { double d; u4 parts[2]; } conv; 183 184 ptr += idx; 185 conv.parts[0] = ptr[0]; 186 conv.parts[1] = ptr[1]; 187 return conv.d; 188 #else 189 double dval; 190 memcpy(&dval, &ptr[idx], 8); 191 return dval; 192 #endif 193 } 194 195 /* store a double into an array of u4 */ 196 static inline void putDoubleToArray(u4* ptr, int idx, double dval) 197 { 198 #if defined(NO_UNALIGN_64__UNION) 199 union { double d; u4 parts[2]; } conv; 200 201 ptr += idx; 202 conv.d = dval; 203 ptr[0] = conv.parts[0]; 204 ptr[1] = conv.parts[1]; 205 #else 206 memcpy(&ptr[idx], &dval, 8); 207 #endif 208 } 209 210 /* 211 * If enabled, validate the register number on every access. Otherwise, 212 * just do an array access. 213 * 214 * Assumes the existence of "u4* fp". 215 * 216 * "_idx" may be referenced more than once. 217 */ 218 #ifdef CHECK_REGISTER_INDICES 219 # define GET_REGISTER(_idx) \ 220 ( (_idx) < curMethod->registersSize ? \ 221 (fp[(_idx)]) : (assert(!"bad reg"),1969) ) 222 # define SET_REGISTER(_idx, _val) \ 223 ( (_idx) < curMethod->registersSize ? \ 224 (fp[(_idx)] = (u4)(_val)) : (assert(!"bad reg"),1969) ) 225 # define GET_REGISTER_AS_OBJECT(_idx) ((Object *)GET_REGISTER(_idx)) 226 # define SET_REGISTER_AS_OBJECT(_idx, _val) SET_REGISTER(_idx, (s4)_val) 227 # define GET_REGISTER_INT(_idx) ((s4) GET_REGISTER(_idx)) 228 # define SET_REGISTER_INT(_idx, _val) SET_REGISTER(_idx, (s4)_val) 229 # define GET_REGISTER_WIDE(_idx) \ 230 ( (_idx) < curMethod->registersSize-1 ? \ 231 getLongFromArray(fp, (_idx)) : (assert(!"bad reg"),1969) ) 232 # define SET_REGISTER_WIDE(_idx, _val) \ 233 ( (_idx) < curMethod->registersSize-1 ? \ 234 (void)putLongToArray(fp, (_idx), (_val)) : assert(!"bad reg") ) 235 # define GET_REGISTER_FLOAT(_idx) \ 236 ( (_idx) < curMethod->registersSize ? \ 237 (*((float*) &fp[(_idx)])) : (assert(!"bad reg"),1969.0f) ) 238 # define SET_REGISTER_FLOAT(_idx, _val) \ 239 ( (_idx) < curMethod->registersSize ? \ 240 (*((float*) &fp[(_idx)]) = (_val)) : (assert(!"bad reg"),1969.0f) ) 241 # define GET_REGISTER_DOUBLE(_idx) \ 242 ( (_idx) < curMethod->registersSize-1 ? \ 243 getDoubleFromArray(fp, (_idx)) : (assert(!"bad reg"),1969.0) ) 244 # define SET_REGISTER_DOUBLE(_idx, _val) \ 245 ( (_idx) < curMethod->registersSize-1 ? \ 246 (void)putDoubleToArray(fp, (_idx), (_val)) : assert(!"bad reg") ) 247 #else 248 # define GET_REGISTER(_idx) (fp[(_idx)]) 249 # define SET_REGISTER(_idx, _val) (fp[(_idx)] = (_val)) 250 # define GET_REGISTER_AS_OBJECT(_idx) ((Object*) fp[(_idx)]) 251 # define SET_REGISTER_AS_OBJECT(_idx, _val) (fp[(_idx)] = (u4)(_val)) 252 # define GET_REGISTER_INT(_idx) ((s4)GET_REGISTER(_idx)) 253 # define SET_REGISTER_INT(_idx, _val) SET_REGISTER(_idx, (s4)_val) 254 # define GET_REGISTER_WIDE(_idx) getLongFromArray(fp, (_idx)) 255 # define SET_REGISTER_WIDE(_idx, _val) putLongToArray(fp, (_idx), (_val)) 256 # define GET_REGISTER_FLOAT(_idx) (*((float*) &fp[(_idx)])) 257 # define SET_REGISTER_FLOAT(_idx, _val) (*((float*) &fp[(_idx)]) = (_val)) 258 # define GET_REGISTER_DOUBLE(_idx) getDoubleFromArray(fp, (_idx)) 259 # define SET_REGISTER_DOUBLE(_idx, _val) putDoubleToArray(fp, (_idx), (_val)) 260 #endif 261 262 /* 263 * Get 16 bits from the specified offset of the program counter. We always 264 * want to load 16 bits at a time from the instruction stream -- it's more 265 * efficient than 8 and won't have the alignment problems that 32 might. 266 * 267 * Assumes existence of "const u2* pc". 268 */ 269 #define FETCH(_offset) (pc[(_offset)]) 270 271 /* 272 * Extract instruction byte from 16-bit fetch (_inst is a u2). 273 */ 274 #define INST_INST(_inst) ((_inst) & 0xff) 275 276 /* 277 * Replace the opcode (used when handling breakpoints). _opcode is a u1. 278 */ 279 #define INST_REPLACE_OP(_inst, _opcode) (((_inst) & 0xff00) | _opcode) 280 281 /* 282 * Extract the "vA, vB" 4-bit registers from the instruction word (_inst is u2). 283 */ 284 #define INST_A(_inst) (((_inst) >> 8) & 0x0f) 285 #define INST_B(_inst) ((_inst) >> 12) 286 287 /* 288 * Get the 8-bit "vAA" 8-bit register index from the instruction word. 289 * (_inst is u2) 290 */ 291 #define INST_AA(_inst) ((_inst) >> 8) 292 293 /* 294 * The current PC must be available to Throwable constructors, e.g. 295 * those created by the various exception throw routines, so that the 296 * exception stack trace can be generated correctly. If we don't do this, 297 * the offset within the current method won't be shown correctly. See the 298 * notes in Exception.c. 299 * 300 * This is also used to determine the address for precise GC. 301 * 302 * Assumes existence of "u4* fp" and "const u2* pc". 303 */ 304 #define EXPORT_PC() (SAVEAREA_FROM_FP(fp)->xtra.currentPc = pc) 305 306 /* 307 * Check to see if "obj" is NULL. If so, throw an exception. Assumes the 308 * pc has already been exported to the stack. 309 * 310 * Perform additional checks on debug builds. 311 * 312 * Use this to check for NULL when the instruction handler calls into 313 * something that could throw an exception (so we have already called 314 * EXPORT_PC at the top). 315 */ 316 static inline bool checkForNull(Object* obj) 317 { 318 if (obj == NULL) { 319 dvmThrowNullPointerException(NULL); 320 return false; 321 } 322 #ifdef WITH_EXTRA_OBJECT_VALIDATION 323 if (!dvmIsHeapAddress(obj)) { 324 ALOGE("Invalid object %p", obj); 325 dvmAbort(); 326 } 327 #endif 328 #ifndef NDEBUG 329 if (obj->clazz == NULL || ((u4) obj->clazz) <= 65536) { 330 /* probable heap corruption */ 331 ALOGE("Invalid object class %p (in %p)", obj->clazz, obj); 332 dvmAbort(); 333 } 334 #endif 335 return true; 336 } 337 338 /* 339 * Check to see if "obj" is NULL. If so, export the PC into the stack 340 * frame and throw an exception. 341 * 342 * Perform additional checks on debug builds. 343 * 344 * Use this to check for NULL when the instruction handler doesn't do 345 * anything else that can throw an exception. 346 */ 347 static inline bool checkForNullExportPC(Object* obj, u4* fp, const u2* pc) 348 { 349 if (obj == NULL) { 350 EXPORT_PC(); 351 dvmThrowNullPointerException(NULL); 352 return false; 353 } 354 #ifdef WITH_EXTRA_OBJECT_VALIDATION 355 if (!dvmIsHeapAddress(obj)) { 356 ALOGE("Invalid object %p", obj); 357 dvmAbort(); 358 } 359 #endif 360 #ifndef NDEBUG 361 if (obj->clazz == NULL || ((u4) obj->clazz) <= 65536) { 362 /* probable heap corruption */ 363 ALOGE("Invalid object class %p (in %p)", obj->clazz, obj); 364 dvmAbort(); 365 } 366 #endif 367 return true; 368 } 369 370 /* File: cstubs/stubdefs.cpp */ 371 /* 372 * In the C mterp stubs, "goto" is a function call followed immediately 373 * by a return. 374 */ 375 376 #define GOTO_TARGET_DECL(_target, ...) \ 377 extern "C" void dvmMterp_##_target(Thread* self, ## __VA_ARGS__); 378 379 /* (void)xxx to quiet unused variable compiler warnings. */ 380 #define GOTO_TARGET(_target, ...) \ 381 void dvmMterp_##_target(Thread* self, ## __VA_ARGS__) { \ 382 u2 ref, vsrc1, vsrc2, vdst; \ 383 u2 inst = FETCH(0); \ 384 const Method* methodToCall; \ 385 StackSaveArea* debugSaveArea; \ 386 (void)ref; (void)vsrc1; (void)vsrc2; (void)vdst; (void)inst; \ 387 (void)methodToCall; (void)debugSaveArea; 388 389 #define GOTO_TARGET_END } 390 391 /* 392 * Redefine what used to be local variable accesses into Thread struct 393 * references. (These are undefined down in "footer.cpp".) 394 */ 395 #define retval self->interpSave.retval 396 #define pc self->interpSave.pc 397 #define fp self->interpSave.curFrame 398 #define curMethod self->interpSave.method 399 #define methodClassDex self->interpSave.methodClassDex 400 #define debugTrackedRefStart self->interpSave.debugTrackedRefStart 401 402 /* ugh */ 403 #define STUB_HACK(x) x 404 #if defined(WITH_JIT) 405 #define JIT_STUB_HACK(x) x 406 #else 407 #define JIT_STUB_HACK(x) 408 #endif 409 410 /* 411 * InterpSave's pc and fp must be valid when breaking out to a 412 * "Reportxxx" routine. Because the portable interpreter uses local 413 * variables for these, we must flush prior. Stubs, however, use 414 * the interpSave vars directly, so this is a nop for stubs. 415 */ 416 #define PC_FP_TO_SELF() 417 #define PC_TO_SELF() 418 419 /* 420 * Opcode handler framing macros. Here, each opcode is a separate function 421 * that takes a "self" argument and returns void. We can't declare 422 * these "static" because they may be called from an assembly stub. 423 * (void)xxx to quiet unused variable compiler warnings. 424 */ 425 #define HANDLE_OPCODE(_op) \ 426 extern "C" void dvmMterp_##_op(Thread* self); \ 427 void dvmMterp_##_op(Thread* self) { \ 428 u4 ref; \ 429 u2 vsrc1, vsrc2, vdst; \ 430 u2 inst = FETCH(0); \ 431 (void)ref; (void)vsrc1; (void)vsrc2; (void)vdst; (void)inst; 432 433 #define OP_END } 434 435 /* 436 * Like the "portable" FINISH, but don't reload "inst", and return to caller 437 * when done. Further, debugger/profiler checks are handled 438 * before handler execution in mterp, so we don't do them here either. 439 */ 440 #if defined(WITH_JIT) 441 #define FINISH(_offset) { \ 442 ADJUST_PC(_offset); \ 443 if (self->interpBreak.ctl.subMode & kSubModeJitTraceBuild) { \ 444 dvmCheckJit(pc, self); \ 445 } \ 446 return; \ 447 } 448 #else 449 #define FINISH(_offset) { \ 450 ADJUST_PC(_offset); \ 451 return; \ 452 } 453 #endif 454 455 456 /* 457 * The "goto label" statements turn into function calls followed by 458 * return statements. Some of the functions take arguments, which in the 459 * portable interpreter are handled by assigning values to globals. 460 */ 461 462 #define GOTO_exceptionThrown() \ 463 do { \ 464 dvmMterp_exceptionThrown(self); \ 465 return; \ 466 } while(false) 467 468 #define GOTO_returnFromMethod() \ 469 do { \ 470 dvmMterp_returnFromMethod(self); \ 471 return; \ 472 } while(false) 473 474 #define GOTO_invoke(_target, _methodCallRange) \ 475 do { \ 476 dvmMterp_##_target(self, _methodCallRange); \ 477 return; \ 478 } while(false) 479 480 #define GOTO_invokeMethod(_methodCallRange, _methodToCall, _vsrc1, _vdst) \ 481 do { \ 482 dvmMterp_invokeMethod(self, _methodCallRange, _methodToCall, \ 483 _vsrc1, _vdst); \ 484 return; \ 485 } while(false) 486 487 /* 488 * As a special case, "goto bail" turns into a longjmp. 489 */ 490 #define GOTO_bail() \ 491 dvmMterpStdBail(self, false); 492 493 /* 494 * Periodically check for thread suspension. 495 * 496 * While we're at it, see if a debugger has attached or the profiler has 497 * started. 498 */ 499 #define PERIODIC_CHECKS(_pcadj) { \ 500 if (dvmCheckSuspendQuick(self)) { \ 501 EXPORT_PC(); /* need for precise GC */ \ 502 dvmCheckSuspendPending(self); \ 503 } \ 504 } 505 506 /* File: c/opcommon.cpp */ 507 /* forward declarations of goto targets */ 508 GOTO_TARGET_DECL(filledNewArray, bool methodCallRange); 509 GOTO_TARGET_DECL(invokeVirtual, bool methodCallRange); 510 GOTO_TARGET_DECL(invokeSuper, bool methodCallRange); 511 GOTO_TARGET_DECL(invokeInterface, bool methodCallRange); 512 GOTO_TARGET_DECL(invokeDirect, bool methodCallRange); 513 GOTO_TARGET_DECL(invokeStatic, bool methodCallRange); 514 GOTO_TARGET_DECL(invokeVirtualQuick, bool methodCallRange); 515 GOTO_TARGET_DECL(invokeSuperQuick, bool methodCallRange); 516 GOTO_TARGET_DECL(invokeMethod, bool methodCallRange, const Method* methodToCall, 517 u2 count, u2 regs); 518 GOTO_TARGET_DECL(returnFromMethod); 519 GOTO_TARGET_DECL(exceptionThrown); 520 521 /* 522 * =========================================================================== 523 * 524 * What follows are opcode definitions shared between multiple opcodes with 525 * minor substitutions handled by the C pre-processor. These should probably 526 * use the mterp substitution mechanism instead, with the code here moved 527 * into common fragment files (like the asm "binop.S"), although it's hard 528 * to give up the C preprocessor in favor of the much simpler text subst. 529 * 530 * =========================================================================== 531 */ 532 533 #define HANDLE_NUMCONV(_opcode, _opname, _fromtype, _totype) \ 534 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 535 vdst = INST_A(inst); \ 536 vsrc1 = INST_B(inst); \ 537 ILOGV("|%s v%d,v%d", (_opname), vdst, vsrc1); \ 538 SET_REGISTER##_totype(vdst, \ 539 GET_REGISTER##_fromtype(vsrc1)); \ 540 FINISH(1); 541 542 #define HANDLE_FLOAT_TO_INT(_opcode, _opname, _fromvtype, _fromrtype, \ 543 _tovtype, _tortype) \ 544 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 545 { \ 546 /* spec defines specific handling for +/- inf and NaN values */ \ 547 _fromvtype val; \ 548 _tovtype intMin, intMax, result; \ 549 vdst = INST_A(inst); \ 550 vsrc1 = INST_B(inst); \ 551 ILOGV("|%s v%d,v%d", (_opname), vdst, vsrc1); \ 552 val = GET_REGISTER##_fromrtype(vsrc1); \ 553 intMin = (_tovtype) 1 << (sizeof(_tovtype) * 8 -1); \ 554 intMax = ~intMin; \ 555 result = (_tovtype) val; \ 556 if (val >= intMax) /* +inf */ \ 557 result = intMax; \ 558 else if (val <= intMin) /* -inf */ \ 559 result = intMin; \ 560 else if (val != val) /* NaN */ \ 561 result = 0; \ 562 else \ 563 result = (_tovtype) val; \ 564 SET_REGISTER##_tortype(vdst, result); \ 565 } \ 566 FINISH(1); 567 568 #define HANDLE_INT_TO_SMALL(_opcode, _opname, _type) \ 569 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 570 vdst = INST_A(inst); \ 571 vsrc1 = INST_B(inst); \ 572 ILOGV("|int-to-%s v%d,v%d", (_opname), vdst, vsrc1); \ 573 SET_REGISTER(vdst, (_type) GET_REGISTER(vsrc1)); \ 574 FINISH(1); 575 576 /* NOTE: the comparison result is always a signed 4-byte integer */ 577 #define HANDLE_OP_CMPX(_opcode, _opname, _varType, _type, _nanVal) \ 578 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 579 { \ 580 int result; \ 581 u2 regs; \ 582 _varType val1, val2; \ 583 vdst = INST_AA(inst); \ 584 regs = FETCH(1); \ 585 vsrc1 = regs & 0xff; \ 586 vsrc2 = regs >> 8; \ 587 ILOGV("|cmp%s v%d,v%d,v%d", (_opname), vdst, vsrc1, vsrc2); \ 588 val1 = GET_REGISTER##_type(vsrc1); \ 589 val2 = GET_REGISTER##_type(vsrc2); \ 590 if (val1 == val2) \ 591 result = 0; \ 592 else if (val1 < val2) \ 593 result = -1; \ 594 else if (val1 > val2) \ 595 result = 1; \ 596 else \ 597 result = (_nanVal); \ 598 ILOGV("+ result=%d", result); \ 599 SET_REGISTER(vdst, result); \ 600 } \ 601 FINISH(2); 602 603 #define HANDLE_OP_IF_XX(_opcode, _opname, _cmp) \ 604 HANDLE_OPCODE(_opcode /*vA, vB, +CCCC*/) \ 605 vsrc1 = INST_A(inst); \ 606 vsrc2 = INST_B(inst); \ 607 if ((s4) GET_REGISTER(vsrc1) _cmp (s4) GET_REGISTER(vsrc2)) { \ 608 int branchOffset = (s2)FETCH(1); /* sign-extended */ \ 609 ILOGV("|if-%s v%d,v%d,+0x%04x", (_opname), vsrc1, vsrc2, \ 610 branchOffset); \ 611 ILOGV("> branch taken"); \ 612 if (branchOffset < 0) \ 613 PERIODIC_CHECKS(branchOffset); \ 614 FINISH(branchOffset); \ 615 } else { \ 616 ILOGV("|if-%s v%d,v%d,-", (_opname), vsrc1, vsrc2); \ 617 FINISH(2); \ 618 } 619 620 #define HANDLE_OP_IF_XXZ(_opcode, _opname, _cmp) \ 621 HANDLE_OPCODE(_opcode /*vAA, +BBBB*/) \ 622 vsrc1 = INST_AA(inst); \ 623 if ((s4) GET_REGISTER(vsrc1) _cmp 0) { \ 624 int branchOffset = (s2)FETCH(1); /* sign-extended */ \ 625 ILOGV("|if-%s v%d,+0x%04x", (_opname), vsrc1, branchOffset); \ 626 ILOGV("> branch taken"); \ 627 if (branchOffset < 0) \ 628 PERIODIC_CHECKS(branchOffset); \ 629 FINISH(branchOffset); \ 630 } else { \ 631 ILOGV("|if-%s v%d,-", (_opname), vsrc1); \ 632 FINISH(2); \ 633 } 634 635 #define HANDLE_UNOP(_opcode, _opname, _pfx, _sfx, _type) \ 636 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 637 vdst = INST_A(inst); \ 638 vsrc1 = INST_B(inst); \ 639 ILOGV("|%s v%d,v%d", (_opname), vdst, vsrc1); \ 640 SET_REGISTER##_type(vdst, _pfx GET_REGISTER##_type(vsrc1) _sfx); \ 641 FINISH(1); 642 643 #define HANDLE_OP_X_INT(_opcode, _opname, _op, _chkdiv) \ 644 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 645 { \ 646 u2 srcRegs; \ 647 vdst = INST_AA(inst); \ 648 srcRegs = FETCH(1); \ 649 vsrc1 = srcRegs & 0xff; \ 650 vsrc2 = srcRegs >> 8; \ 651 ILOGV("|%s-int v%d,v%d", (_opname), vdst, vsrc1); \ 652 if (_chkdiv != 0) { \ 653 s4 firstVal, secondVal, result; \ 654 firstVal = GET_REGISTER(vsrc1); \ 655 secondVal = GET_REGISTER(vsrc2); \ 656 if (secondVal == 0) { \ 657 EXPORT_PC(); \ 658 dvmThrowArithmeticException("divide by zero"); \ 659 GOTO_exceptionThrown(); \ 660 } \ 661 if ((u4)firstVal == 0x80000000 && secondVal == -1) { \ 662 if (_chkdiv == 1) \ 663 result = firstVal; /* division */ \ 664 else \ 665 result = 0; /* remainder */ \ 666 } else { \ 667 result = firstVal _op secondVal; \ 668 } \ 669 SET_REGISTER(vdst, result); \ 670 } else { \ 671 /* non-div/rem case */ \ 672 SET_REGISTER(vdst, \ 673 (s4) GET_REGISTER(vsrc1) _op (s4) GET_REGISTER(vsrc2)); \ 674 } \ 675 } \ 676 FINISH(2); 677 678 #define HANDLE_OP_SHX_INT(_opcode, _opname, _cast, _op) \ 679 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 680 { \ 681 u2 srcRegs; \ 682 vdst = INST_AA(inst); \ 683 srcRegs = FETCH(1); \ 684 vsrc1 = srcRegs & 0xff; \ 685 vsrc2 = srcRegs >> 8; \ 686 ILOGV("|%s-int v%d,v%d", (_opname), vdst, vsrc1); \ 687 SET_REGISTER(vdst, \ 688 _cast GET_REGISTER(vsrc1) _op (GET_REGISTER(vsrc2) & 0x1f)); \ 689 } \ 690 FINISH(2); 691 692 #define HANDLE_OP_X_INT_LIT16(_opcode, _opname, _op, _chkdiv) \ 693 HANDLE_OPCODE(_opcode /*vA, vB, #+CCCC*/) \ 694 vdst = INST_A(inst); \ 695 vsrc1 = INST_B(inst); \ 696 vsrc2 = FETCH(1); \ 697 ILOGV("|%s-int/lit16 v%d,v%d,#+0x%04x", \ 698 (_opname), vdst, vsrc1, vsrc2); \ 699 if (_chkdiv != 0) { \ 700 s4 firstVal, result; \ 701 firstVal = GET_REGISTER(vsrc1); \ 702 if ((s2) vsrc2 == 0) { \ 703 EXPORT_PC(); \ 704 dvmThrowArithmeticException("divide by zero"); \ 705 GOTO_exceptionThrown(); \ 706 } \ 707 if ((u4)firstVal == 0x80000000 && ((s2) vsrc2) == -1) { \ 708 /* won't generate /lit16 instr for this; check anyway */ \ 709 if (_chkdiv == 1) \ 710 result = firstVal; /* division */ \ 711 else \ 712 result = 0; /* remainder */ \ 713 } else { \ 714 result = firstVal _op (s2) vsrc2; \ 715 } \ 716 SET_REGISTER(vdst, result); \ 717 } else { \ 718 /* non-div/rem case */ \ 719 SET_REGISTER(vdst, GET_REGISTER(vsrc1) _op (s2) vsrc2); \ 720 } \ 721 FINISH(2); 722 723 #define HANDLE_OP_X_INT_LIT8(_opcode, _opname, _op, _chkdiv) \ 724 HANDLE_OPCODE(_opcode /*vAA, vBB, #+CC*/) \ 725 { \ 726 u2 litInfo; \ 727 vdst = INST_AA(inst); \ 728 litInfo = FETCH(1); \ 729 vsrc1 = litInfo & 0xff; \ 730 vsrc2 = litInfo >> 8; /* constant */ \ 731 ILOGV("|%s-int/lit8 v%d,v%d,#+0x%02x", \ 732 (_opname), vdst, vsrc1, vsrc2); \ 733 if (_chkdiv != 0) { \ 734 s4 firstVal, result; \ 735 firstVal = GET_REGISTER(vsrc1); \ 736 if ((s1) vsrc2 == 0) { \ 737 EXPORT_PC(); \ 738 dvmThrowArithmeticException("divide by zero"); \ 739 GOTO_exceptionThrown(); \ 740 } \ 741 if ((u4)firstVal == 0x80000000 && ((s1) vsrc2) == -1) { \ 742 if (_chkdiv == 1) \ 743 result = firstVal; /* division */ \ 744 else \ 745 result = 0; /* remainder */ \ 746 } else { \ 747 result = firstVal _op ((s1) vsrc2); \ 748 } \ 749 SET_REGISTER(vdst, result); \ 750 } else { \ 751 SET_REGISTER(vdst, \ 752 (s4) GET_REGISTER(vsrc1) _op (s1) vsrc2); \ 753 } \ 754 } \ 755 FINISH(2); 756 757 #define HANDLE_OP_SHX_INT_LIT8(_opcode, _opname, _cast, _op) \ 758 HANDLE_OPCODE(_opcode /*vAA, vBB, #+CC*/) \ 759 { \ 760 u2 litInfo; \ 761 vdst = INST_AA(inst); \ 762 litInfo = FETCH(1); \ 763 vsrc1 = litInfo & 0xff; \ 764 vsrc2 = litInfo >> 8; /* constant */ \ 765 ILOGV("|%s-int/lit8 v%d,v%d,#+0x%02x", \ 766 (_opname), vdst, vsrc1, vsrc2); \ 767 SET_REGISTER(vdst, \ 768 _cast GET_REGISTER(vsrc1) _op (vsrc2 & 0x1f)); \ 769 } \ 770 FINISH(2); 771 772 #define HANDLE_OP_X_INT_2ADDR(_opcode, _opname, _op, _chkdiv) \ 773 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 774 vdst = INST_A(inst); \ 775 vsrc1 = INST_B(inst); \ 776 ILOGV("|%s-int-2addr v%d,v%d", (_opname), vdst, vsrc1); \ 777 if (_chkdiv != 0) { \ 778 s4 firstVal, secondVal, result; \ 779 firstVal = GET_REGISTER(vdst); \ 780 secondVal = GET_REGISTER(vsrc1); \ 781 if (secondVal == 0) { \ 782 EXPORT_PC(); \ 783 dvmThrowArithmeticException("divide by zero"); \ 784 GOTO_exceptionThrown(); \ 785 } \ 786 if ((u4)firstVal == 0x80000000 && secondVal == -1) { \ 787 if (_chkdiv == 1) \ 788 result = firstVal; /* division */ \ 789 else \ 790 result = 0; /* remainder */ \ 791 } else { \ 792 result = firstVal _op secondVal; \ 793 } \ 794 SET_REGISTER(vdst, result); \ 795 } else { \ 796 SET_REGISTER(vdst, \ 797 (s4) GET_REGISTER(vdst) _op (s4) GET_REGISTER(vsrc1)); \ 798 } \ 799 FINISH(1); 800 801 #define HANDLE_OP_SHX_INT_2ADDR(_opcode, _opname, _cast, _op) \ 802 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 803 vdst = INST_A(inst); \ 804 vsrc1 = INST_B(inst); \ 805 ILOGV("|%s-int-2addr v%d,v%d", (_opname), vdst, vsrc1); \ 806 SET_REGISTER(vdst, \ 807 _cast GET_REGISTER(vdst) _op (GET_REGISTER(vsrc1) & 0x1f)); \ 808 FINISH(1); 809 810 #define HANDLE_OP_X_LONG(_opcode, _opname, _op, _chkdiv) \ 811 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 812 { \ 813 u2 srcRegs; \ 814 vdst = INST_AA(inst); \ 815 srcRegs = FETCH(1); \ 816 vsrc1 = srcRegs & 0xff; \ 817 vsrc2 = srcRegs >> 8; \ 818 ILOGV("|%s-long v%d,v%d,v%d", (_opname), vdst, vsrc1, vsrc2); \ 819 if (_chkdiv != 0) { \ 820 s8 firstVal, secondVal, result; \ 821 firstVal = GET_REGISTER_WIDE(vsrc1); \ 822 secondVal = GET_REGISTER_WIDE(vsrc2); \ 823 if (secondVal == 0LL) { \ 824 EXPORT_PC(); \ 825 dvmThrowArithmeticException("divide by zero"); \ 826 GOTO_exceptionThrown(); \ 827 } \ 828 if ((u8)firstVal == 0x8000000000000000ULL && \ 829 secondVal == -1LL) \ 830 { \ 831 if (_chkdiv == 1) \ 832 result = firstVal; /* division */ \ 833 else \ 834 result = 0; /* remainder */ \ 835 } else { \ 836 result = firstVal _op secondVal; \ 837 } \ 838 SET_REGISTER_WIDE(vdst, result); \ 839 } else { \ 840 SET_REGISTER_WIDE(vdst, \ 841 (s8) GET_REGISTER_WIDE(vsrc1) _op (s8) GET_REGISTER_WIDE(vsrc2)); \ 842 } \ 843 } \ 844 FINISH(2); 845 846 #define HANDLE_OP_SHX_LONG(_opcode, _opname, _cast, _op) \ 847 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 848 { \ 849 u2 srcRegs; \ 850 vdst = INST_AA(inst); \ 851 srcRegs = FETCH(1); \ 852 vsrc1 = srcRegs & 0xff; \ 853 vsrc2 = srcRegs >> 8; \ 854 ILOGV("|%s-long v%d,v%d,v%d", (_opname), vdst, vsrc1, vsrc2); \ 855 SET_REGISTER_WIDE(vdst, \ 856 _cast GET_REGISTER_WIDE(vsrc1) _op (GET_REGISTER(vsrc2) & 0x3f)); \ 857 } \ 858 FINISH(2); 859 860 #define HANDLE_OP_X_LONG_2ADDR(_opcode, _opname, _op, _chkdiv) \ 861 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 862 vdst = INST_A(inst); \ 863 vsrc1 = INST_B(inst); \ 864 ILOGV("|%s-long-2addr v%d,v%d", (_opname), vdst, vsrc1); \ 865 if (_chkdiv != 0) { \ 866 s8 firstVal, secondVal, result; \ 867 firstVal = GET_REGISTER_WIDE(vdst); \ 868 secondVal = GET_REGISTER_WIDE(vsrc1); \ 869 if (secondVal == 0LL) { \ 870 EXPORT_PC(); \ 871 dvmThrowArithmeticException("divide by zero"); \ 872 GOTO_exceptionThrown(); \ 873 } \ 874 if ((u8)firstVal == 0x8000000000000000ULL && \ 875 secondVal == -1LL) \ 876 { \ 877 if (_chkdiv == 1) \ 878 result = firstVal; /* division */ \ 879 else \ 880 result = 0; /* remainder */ \ 881 } else { \ 882 result = firstVal _op secondVal; \ 883 } \ 884 SET_REGISTER_WIDE(vdst, result); \ 885 } else { \ 886 SET_REGISTER_WIDE(vdst, \ 887 (s8) GET_REGISTER_WIDE(vdst) _op (s8)GET_REGISTER_WIDE(vsrc1));\ 888 } \ 889 FINISH(1); 890 891 #define HANDLE_OP_SHX_LONG_2ADDR(_opcode, _opname, _cast, _op) \ 892 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 893 vdst = INST_A(inst); \ 894 vsrc1 = INST_B(inst); \ 895 ILOGV("|%s-long-2addr v%d,v%d", (_opname), vdst, vsrc1); \ 896 SET_REGISTER_WIDE(vdst, \ 897 _cast GET_REGISTER_WIDE(vdst) _op (GET_REGISTER(vsrc1) & 0x3f)); \ 898 FINISH(1); 899 900 #define HANDLE_OP_X_FLOAT(_opcode, _opname, _op) \ 901 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 902 { \ 903 u2 srcRegs; \ 904 vdst = INST_AA(inst); \ 905 srcRegs = FETCH(1); \ 906 vsrc1 = srcRegs & 0xff; \ 907 vsrc2 = srcRegs >> 8; \ 908 ILOGV("|%s-float v%d,v%d,v%d", (_opname), vdst, vsrc1, vsrc2); \ 909 SET_REGISTER_FLOAT(vdst, \ 910 GET_REGISTER_FLOAT(vsrc1) _op GET_REGISTER_FLOAT(vsrc2)); \ 911 } \ 912 FINISH(2); 913 914 #define HANDLE_OP_X_DOUBLE(_opcode, _opname, _op) \ 915 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 916 { \ 917 u2 srcRegs; \ 918 vdst = INST_AA(inst); \ 919 srcRegs = FETCH(1); \ 920 vsrc1 = srcRegs & 0xff; \ 921 vsrc2 = srcRegs >> 8; \ 922 ILOGV("|%s-double v%d,v%d,v%d", (_opname), vdst, vsrc1, vsrc2); \ 923 SET_REGISTER_DOUBLE(vdst, \ 924 GET_REGISTER_DOUBLE(vsrc1) _op GET_REGISTER_DOUBLE(vsrc2)); \ 925 } \ 926 FINISH(2); 927 928 #define HANDLE_OP_X_FLOAT_2ADDR(_opcode, _opname, _op) \ 929 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 930 vdst = INST_A(inst); \ 931 vsrc1 = INST_B(inst); \ 932 ILOGV("|%s-float-2addr v%d,v%d", (_opname), vdst, vsrc1); \ 933 SET_REGISTER_FLOAT(vdst, \ 934 GET_REGISTER_FLOAT(vdst) _op GET_REGISTER_FLOAT(vsrc1)); \ 935 FINISH(1); 936 937 #define HANDLE_OP_X_DOUBLE_2ADDR(_opcode, _opname, _op) \ 938 HANDLE_OPCODE(_opcode /*vA, vB*/) \ 939 vdst = INST_A(inst); \ 940 vsrc1 = INST_B(inst); \ 941 ILOGV("|%s-double-2addr v%d,v%d", (_opname), vdst, vsrc1); \ 942 SET_REGISTER_DOUBLE(vdst, \ 943 GET_REGISTER_DOUBLE(vdst) _op GET_REGISTER_DOUBLE(vsrc1)); \ 944 FINISH(1); 945 946 #define HANDLE_OP_AGET(_opcode, _opname, _type, _regsize) \ 947 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 948 { \ 949 ArrayObject* arrayObj; \ 950 u2 arrayInfo; \ 951 EXPORT_PC(); \ 952 vdst = INST_AA(inst); \ 953 arrayInfo = FETCH(1); \ 954 vsrc1 = arrayInfo & 0xff; /* array ptr */ \ 955 vsrc2 = arrayInfo >> 8; /* index */ \ 956 ILOGV("|aget%s v%d,v%d,v%d", (_opname), vdst, vsrc1, vsrc2); \ 957 arrayObj = (ArrayObject*) GET_REGISTER(vsrc1); \ 958 if (!checkForNull((Object*) arrayObj)) \ 959 GOTO_exceptionThrown(); \ 960 if (GET_REGISTER(vsrc2) >= arrayObj->length) { \ 961 dvmThrowArrayIndexOutOfBoundsException( \ 962 arrayObj->length, GET_REGISTER(vsrc2)); \ 963 GOTO_exceptionThrown(); \ 964 } \ 965 SET_REGISTER##_regsize(vdst, \ 966 ((_type*)(void*)arrayObj->contents)[GET_REGISTER(vsrc2)]); \ 967 ILOGV("+ AGET[%d]=%#x", GET_REGISTER(vsrc2), GET_REGISTER(vdst)); \ 968 } \ 969 FINISH(2); 970 971 #define HANDLE_OP_APUT(_opcode, _opname, _type, _regsize) \ 972 HANDLE_OPCODE(_opcode /*vAA, vBB, vCC*/) \ 973 { \ 974 ArrayObject* arrayObj; \ 975 u2 arrayInfo; \ 976 EXPORT_PC(); \ 977 vdst = INST_AA(inst); /* AA: source value */ \ 978 arrayInfo = FETCH(1); \ 979 vsrc1 = arrayInfo & 0xff; /* BB: array ptr */ \ 980 vsrc2 = arrayInfo >> 8; /* CC: index */ \ 981 ILOGV("|aput%s v%d,v%d,v%d", (_opname), vdst, vsrc1, vsrc2); \ 982 arrayObj = (ArrayObject*) GET_REGISTER(vsrc1); \ 983 if (!checkForNull((Object*) arrayObj)) \ 984 GOTO_exceptionThrown(); \ 985 if (GET_REGISTER(vsrc2) >= arrayObj->length) { \ 986 dvmThrowArrayIndexOutOfBoundsException( \ 987 arrayObj->length, GET_REGISTER(vsrc2)); \ 988 GOTO_exceptionThrown(); \ 989 } \ 990 ILOGV("+ APUT[%d]=0x%08x", GET_REGISTER(vsrc2), GET_REGISTER(vdst));\ 991 ((_type*)(void*)arrayObj->contents)[GET_REGISTER(vsrc2)] = \ 992 GET_REGISTER##_regsize(vdst); \ 993 } \ 994 FINISH(2); 995 996 /* 997 * It's possible to get a bad value out of a field with sub-32-bit stores 998 * because the -quick versions always operate on 32 bits. Consider: 999 * short foo = -1 (sets a 32-bit register to 0xffffffff) 1000 * iput-quick foo (writes all 32 bits to the field) 1001 * short bar = 1 (sets a 32-bit register to 0x00000001) 1002 * iput-short (writes the low 16 bits to the field) 1003 * iget-quick foo (reads all 32 bits from the field, yielding 0xffff0001) 1004 * This can only happen when optimized and non-optimized code has interleaved 1005 * access to the same field. This is unlikely but possible. 1006 * 1007 * The easiest way to fix this is to always read/write 32 bits at a time. On 1008 * a device with a 16-bit data bus this is sub-optimal. (The alternative 1009 * approach is to have sub-int versions of iget-quick, but now we're wasting 1010 * Dalvik instruction space and making it less likely that handler code will 1011 * already be in the CPU i-cache.) 1012 */ 1013 #define HANDLE_IGET_X(_opcode, _opname, _ftype, _regsize) \ 1014 HANDLE_OPCODE(_opcode /*vA, vB, field@CCCC*/) \ 1015 { \ 1016 InstField* ifield; \ 1017 Object* obj; \ 1018 EXPORT_PC(); \ 1019 vdst = INST_A(inst); \ 1020 vsrc1 = INST_B(inst); /* object ptr */ \ 1021 ref = FETCH(1); /* field ref */ \ 1022 ILOGV("|iget%s v%d,v%d,field@0x%04x", (_opname), vdst, vsrc1, ref); \ 1023 obj = (Object*) GET_REGISTER(vsrc1); \ 1024 if (!checkForNull(obj)) \ 1025 GOTO_exceptionThrown(); \ 1026 ifield = (InstField*) dvmDexGetResolvedField(methodClassDex, ref); \ 1027 if (ifield == NULL) { \ 1028 ifield = dvmResolveInstField(curMethod->clazz, ref); \ 1029 if (ifield == NULL) \ 1030 GOTO_exceptionThrown(); \ 1031 } \ 1032 SET_REGISTER##_regsize(vdst, \ 1033 dvmGetField##_ftype(obj, ifield->byteOffset)); \ 1034 ILOGV("+ IGET '%s'=0x%08llx", ifield->field.name, \ 1035 (u8) GET_REGISTER##_regsize(vdst)); \ 1036 } \ 1037 FINISH(2); 1038 1039 #define HANDLE_IGET_X_QUICK(_opcode, _opname, _ftype, _regsize) \ 1040 HANDLE_OPCODE(_opcode /*vA, vB, field@CCCC*/) \ 1041 { \ 1042 Object* obj; \ 1043 vdst = INST_A(inst); \ 1044 vsrc1 = INST_B(inst); /* object ptr */ \ 1045 ref = FETCH(1); /* field offset */ \ 1046 ILOGV("|iget%s-quick v%d,v%d,field@+%u", \ 1047 (_opname), vdst, vsrc1, ref); \ 1048 obj = (Object*) GET_REGISTER(vsrc1); \ 1049 if (!checkForNullExportPC(obj, fp, pc)) \ 1050 GOTO_exceptionThrown(); \ 1051 SET_REGISTER##_regsize(vdst, dvmGetField##_ftype(obj, ref)); \ 1052 ILOGV("+ IGETQ %d=0x%08llx", ref, \ 1053 (u8) GET_REGISTER##_regsize(vdst)); \ 1054 } \ 1055 FINISH(2); 1056 1057 #define HANDLE_IPUT_X(_opcode, _opname, _ftype, _regsize) \ 1058 HANDLE_OPCODE(_opcode /*vA, vB, field@CCCC*/) \ 1059 { \ 1060 InstField* ifield; \ 1061 Object* obj; \ 1062 EXPORT_PC(); \ 1063 vdst = INST_A(inst); \ 1064 vsrc1 = INST_B(inst); /* object ptr */ \ 1065 ref = FETCH(1); /* field ref */ \ 1066 ILOGV("|iput%s v%d,v%d,field@0x%04x", (_opname), vdst, vsrc1, ref); \ 1067 obj = (Object*) GET_REGISTER(vsrc1); \ 1068 if (!checkForNull(obj)) \ 1069 GOTO_exceptionThrown(); \ 1070 ifield = (InstField*) dvmDexGetResolvedField(methodClassDex, ref); \ 1071 if (ifield == NULL) { \ 1072 ifield = dvmResolveInstField(curMethod->clazz, ref); \ 1073 if (ifield == NULL) \ 1074 GOTO_exceptionThrown(); \ 1075 } \ 1076 dvmSetField##_ftype(obj, ifield->byteOffset, \ 1077 GET_REGISTER##_regsize(vdst)); \ 1078 ILOGV("+ IPUT '%s'=0x%08llx", ifield->field.name, \ 1079 (u8) GET_REGISTER##_regsize(vdst)); \ 1080 } \ 1081 FINISH(2); 1082 1083 #define HANDLE_IPUT_X_QUICK(_opcode, _opname, _ftype, _regsize) \ 1084 HANDLE_OPCODE(_opcode /*vA, vB, field@CCCC*/) \ 1085 { \ 1086 Object* obj; \ 1087 vdst = INST_A(inst); \ 1088 vsrc1 = INST_B(inst); /* object ptr */ \ 1089 ref = FETCH(1); /* field offset */ \ 1090 ILOGV("|iput%s-quick v%d,v%d,field@0x%04x", \ 1091 (_opname), vdst, vsrc1, ref); \ 1092 obj = (Object*) GET_REGISTER(vsrc1); \ 1093 if (!checkForNullExportPC(obj, fp, pc)) \ 1094 GOTO_exceptionThrown(); \ 1095 dvmSetField##_ftype(obj, ref, GET_REGISTER##_regsize(vdst)); \ 1096 ILOGV("+ IPUTQ %d=0x%08llx", ref, \ 1097 (u8) GET_REGISTER##_regsize(vdst)); \ 1098 } \ 1099 FINISH(2); 1100 1101 /* 1102 * The JIT needs dvmDexGetResolvedField() to return non-null. 1103 * Because the portable interpreter is not involved with the JIT 1104 * and trace building, we only need the extra check here when this 1105 * code is massaged into a stub called from an assembly interpreter. 1106 * This is controlled by the JIT_STUB_HACK maco. 1107 */ 1108 1109 #define HANDLE_SGET_X(_opcode, _opname, _ftype, _regsize) \ 1110 HANDLE_OPCODE(_opcode /*vAA, field@BBBB*/) \ 1111 { \ 1112 StaticField* sfield; \ 1113 vdst = INST_AA(inst); \ 1114 ref = FETCH(1); /* field ref */ \ 1115 ILOGV("|sget%s v%d,sfield@0x%04x", (_opname), vdst, ref); \ 1116 sfield = (StaticField*)dvmDexGetResolvedField(methodClassDex, ref); \ 1117 if (sfield == NULL) { \ 1118 EXPORT_PC(); \ 1119 sfield = dvmResolveStaticField(curMethod->clazz, ref); \ 1120 if (sfield == NULL) \ 1121 GOTO_exceptionThrown(); \ 1122 if (dvmDexGetResolvedField(methodClassDex, ref) == NULL) { \ 1123 JIT_STUB_HACK(dvmJitEndTraceSelect(self,pc)); \ 1124 } \ 1125 } \ 1126 SET_REGISTER##_regsize(vdst, dvmGetStaticField##_ftype(sfield)); \ 1127 ILOGV("+ SGET '%s'=0x%08llx", \ 1128 sfield->field.name, (u8)GET_REGISTER##_regsize(vdst)); \ 1129 } \ 1130 FINISH(2); 1131 1132 #define HANDLE_SPUT_X(_opcode, _opname, _ftype, _regsize) \ 1133 HANDLE_OPCODE(_opcode /*vAA, field@BBBB*/) \ 1134 { \ 1135 StaticField* sfield; \ 1136 vdst = INST_AA(inst); \ 1137 ref = FETCH(1); /* field ref */ \ 1138 ILOGV("|sput%s v%d,sfield@0x%04x", (_opname), vdst, ref); \ 1139 sfield = (StaticField*)dvmDexGetResolvedField(methodClassDex, ref); \ 1140 if (sfield == NULL) { \ 1141 EXPORT_PC(); \ 1142 sfield = dvmResolveStaticField(curMethod->clazz, ref); \ 1143 if (sfield == NULL) \ 1144 GOTO_exceptionThrown(); \ 1145 if (dvmDexGetResolvedField(methodClassDex, ref) == NULL) { \ 1146 JIT_STUB_HACK(dvmJitEndTraceSelect(self,pc)); \ 1147 } \ 1148 } \ 1149 dvmSetStaticField##_ftype(sfield, GET_REGISTER##_regsize(vdst)); \ 1150 ILOGV("+ SPUT '%s'=0x%08llx", \ 1151 sfield->field.name, (u8)GET_REGISTER##_regsize(vdst)); \ 1152 } \ 1153 FINISH(2); 1154 1155 /* File: c/OP_IGET_WIDE_VOLATILE.cpp */ 1156 HANDLE_IGET_X(OP_IGET_WIDE_VOLATILE, "-wide-volatile", LongVolatile, _WIDE) 1157 OP_END 1158 1159 /* File: c/OP_IPUT_WIDE_VOLATILE.cpp */ 1160 HANDLE_IPUT_X(OP_IPUT_WIDE_VOLATILE, "-wide-volatile", LongVolatile, _WIDE) 1161 OP_END 1162 1163 /* File: c/OP_SGET_WIDE_VOLATILE.cpp */ 1164 HANDLE_SGET_X(OP_SGET_WIDE_VOLATILE, "-wide-volatile", LongVolatile, _WIDE) 1165 OP_END 1166 1167 /* File: c/OP_SPUT_WIDE_VOLATILE.cpp */ 1168 HANDLE_SPUT_X(OP_SPUT_WIDE_VOLATILE, "-wide-volatile", LongVolatile, _WIDE) 1169 OP_END 1170 1171 /* File: c/OP_EXECUTE_INLINE_RANGE.cpp */ 1172 HANDLE_OPCODE(OP_EXECUTE_INLINE_RANGE /*{vCCCC..v(CCCC+AA-1)}, inline@BBBB*/) 1173 { 1174 u4 arg0, arg1, arg2, arg3; 1175 arg0 = arg1 = arg2 = arg3 = 0; /* placate gcc */ 1176 1177 EXPORT_PC(); 1178 1179 vsrc1 = INST_AA(inst); /* #of args */ 1180 ref = FETCH(1); /* inline call "ref" */ 1181 vdst = FETCH(2); /* range base */ 1182 ILOGV("|execute-inline-range args=%d @%d {regs=v%d-v%d}", 1183 vsrc1, ref, vdst, vdst+vsrc1-1); 1184 1185 assert((vdst >> 16) == 0); // 16-bit type -or- high 16 bits clear 1186 assert(vsrc1 <= 4); 1187 1188 switch (vsrc1) { 1189 case 4: 1190 arg3 = GET_REGISTER(vdst+3); 1191 /* fall through */ 1192 case 3: 1193 arg2 = GET_REGISTER(vdst+2); 1194 /* fall through */ 1195 case 2: 1196 arg1 = GET_REGISTER(vdst+1); 1197 /* fall through */ 1198 case 1: 1199 arg0 = GET_REGISTER(vdst+0); 1200 /* fall through */ 1201 default: // case 0 1202 ; 1203 } 1204 1205 if (self->interpBreak.ctl.subMode & kSubModeDebuggerActive) { 1206 if (!dvmPerformInlineOp4Dbg(arg0, arg1, arg2, arg3, &retval, ref)) 1207 GOTO_exceptionThrown(); 1208 } else { 1209 if (!dvmPerformInlineOp4Std(arg0, arg1, arg2, arg3, &retval, ref)) 1210 GOTO_exceptionThrown(); 1211 } 1212 } 1213 FINISH(3); 1214 OP_END 1215 1216 /* File: c/OP_INVOKE_OBJECT_INIT_RANGE.cpp */ 1217 HANDLE_OPCODE(OP_INVOKE_OBJECT_INIT_RANGE /*{vCCCC..v(CCCC+AA-1)}, meth@BBBB*/) 1218 { 1219 Object* obj; 1220 1221 vsrc1 = FETCH(2); /* reg number of "this" pointer */ 1222 obj = GET_REGISTER_AS_OBJECT(vsrc1); 1223 1224 if (!checkForNullExportPC(obj, fp, pc)) 1225 GOTO_exceptionThrown(); 1226 1227 /* 1228 * The object should be marked "finalizable" when Object.<init> 1229 * completes normally. We're going to assume it does complete 1230 * (by virtue of being nothing but a return-void) and set it now. 1231 */ 1232 if (IS_CLASS_FLAG_SET(obj->clazz, CLASS_ISFINALIZABLE)) { 1233 EXPORT_PC(); 1234 dvmSetFinalizable(obj); 1235 if (dvmGetException(self)) 1236 GOTO_exceptionThrown(); 1237 } 1238 1239 if (self->interpBreak.ctl.subMode & kSubModeDebuggerActive) { 1240 /* behave like OP_INVOKE_DIRECT_RANGE */ 1241 GOTO_invoke(invokeDirect, true); 1242 } 1243 FINISH(3); 1244 } 1245 OP_END 1246 1247 /* File: c/OP_RETURN_VOID_BARRIER.cpp */ 1248 HANDLE_OPCODE(OP_RETURN_VOID_BARRIER /**/) 1249 ILOGV("|return-void"); 1250 #ifndef NDEBUG 1251 retval.j = 0xababababULL; /* placate valgrind */ 1252 #endif 1253 ANDROID_MEMBAR_STORE(); 1254 GOTO_returnFromMethod(); 1255 OP_END 1256 1257 /* File: c/gotoTargets.cpp */ 1258 /* 1259 * C footer. This has some common code shared by the various targets. 1260 */ 1261 1262 /* 1263 * Everything from here on is a "goto target". In the basic interpreter 1264 * we jump into these targets and then jump directly to the handler for 1265 * next instruction. Here, these are subroutines that return to the caller. 1266 */ 1267 1268 GOTO_TARGET(filledNewArray, bool methodCallRange, bool) 1269 { 1270 ClassObject* arrayClass; 1271 ArrayObject* newArray; 1272 u4* contents; 1273 char typeCh; 1274 int i; 1275 u4 arg5; 1276 1277 EXPORT_PC(); 1278 1279 ref = FETCH(1); /* class ref */ 1280 vdst = FETCH(2); /* first 4 regs -or- range base */ 1281 1282 if (methodCallRange) { 1283 vsrc1 = INST_AA(inst); /* #of elements */ 1284 arg5 = -1; /* silence compiler warning */ 1285 ILOGV("|filled-new-array-range args=%d @0x%04x {regs=v%d-v%d}", 1286 vsrc1, ref, vdst, vdst+vsrc1-1); 1287 } else { 1288 arg5 = INST_A(inst); 1289 vsrc1 = INST_B(inst); /* #of elements */ 1290 ILOGV("|filled-new-array args=%d @0x%04x {regs=0x%04x %x}", 1291 vsrc1, ref, vdst, arg5); 1292 } 1293 1294 /* 1295 * Resolve the array class. 1296 */ 1297 arrayClass = dvmDexGetResolvedClass(methodClassDex, ref); 1298 if (arrayClass == NULL) { 1299 arrayClass = dvmResolveClass(curMethod->clazz, ref, false); 1300 if (arrayClass == NULL) 1301 GOTO_exceptionThrown(); 1302 } 1303 /* 1304 if (!dvmIsArrayClass(arrayClass)) { 1305 dvmThrowRuntimeException( 1306 "filled-new-array needs array class"); 1307 GOTO_exceptionThrown(); 1308 } 1309 */ 1310 /* verifier guarantees this is an array class */ 1311 assert(dvmIsArrayClass(arrayClass)); 1312 assert(dvmIsClassInitialized(arrayClass)); 1313 1314 /* 1315 * Create an array of the specified type. 1316 */ 1317 LOGVV("+++ filled-new-array type is '%s'", arrayClass->descriptor); 1318 typeCh = arrayClass->descriptor[1]; 1319 if (typeCh == 'D' || typeCh == 'J') { 1320 /* category 2 primitives not allowed */ 1321 dvmThrowRuntimeException("bad filled array req"); 1322 GOTO_exceptionThrown(); 1323 } else if (typeCh != 'L' && typeCh != '[' && typeCh != 'I') { 1324 /* TODO: requires multiple "fill in" loops with different widths */ 1325 ALOGE("non-int primitives not implemented"); 1326 dvmThrowInternalError( 1327 "filled-new-array not implemented for anything but 'int'"); 1328 GOTO_exceptionThrown(); 1329 } 1330 1331 newArray = dvmAllocArrayByClass(arrayClass, vsrc1, ALLOC_DONT_TRACK); 1332 if (newArray == NULL) 1333 GOTO_exceptionThrown(); 1334 1335 /* 1336 * Fill in the elements. It's legal for vsrc1 to be zero. 1337 */ 1338 contents = (u4*)(void*)newArray->contents; 1339 if (methodCallRange) { 1340 for (i = 0; i < vsrc1; i++) 1341 contents[i] = GET_REGISTER(vdst+i); 1342 } else { 1343 assert(vsrc1 <= 5); 1344 if (vsrc1 == 5) { 1345 contents[4] = GET_REGISTER(arg5); 1346 vsrc1--; 1347 } 1348 for (i = 0; i < vsrc1; i++) { 1349 contents[i] = GET_REGISTER(vdst & 0x0f); 1350 vdst >>= 4; 1351 } 1352 } 1353 if (typeCh == 'L' || typeCh == '[') { 1354 dvmWriteBarrierArray(newArray, 0, newArray->length); 1355 } 1356 1357 retval.l = (Object*)newArray; 1358 } 1359 FINISH(3); 1360 GOTO_TARGET_END 1361 1362 1363 GOTO_TARGET(invokeVirtual, bool methodCallRange, bool) 1364 { 1365 Method* baseMethod; 1366 Object* thisPtr; 1367 1368 EXPORT_PC(); 1369 1370 vsrc1 = INST_AA(inst); /* AA (count) or BA (count + arg 5) */ 1371 ref = FETCH(1); /* method ref */ 1372 vdst = FETCH(2); /* 4 regs -or- first reg */ 1373 1374 /* 1375 * The object against which we are executing a method is always 1376 * in the first argument. 1377 */ 1378 if (methodCallRange) { 1379 assert(vsrc1 > 0); 1380 ILOGV("|invoke-virtual-range args=%d @0x%04x {regs=v%d-v%d}", 1381 vsrc1, ref, vdst, vdst+vsrc1-1); 1382 thisPtr = (Object*) GET_REGISTER(vdst); 1383 } else { 1384 assert((vsrc1>>4) > 0); 1385 ILOGV("|invoke-virtual args=%d @0x%04x {regs=0x%04x %x}", 1386 vsrc1 >> 4, ref, vdst, vsrc1 & 0x0f); 1387 thisPtr = (Object*) GET_REGISTER(vdst & 0x0f); 1388 } 1389 1390 if (!checkForNull(thisPtr)) 1391 GOTO_exceptionThrown(); 1392 1393 /* 1394 * Resolve the method. This is the correct method for the static 1395 * type of the object. We also verify access permissions here. 1396 */ 1397 baseMethod = dvmDexGetResolvedMethod(methodClassDex, ref); 1398 if (baseMethod == NULL) { 1399 baseMethod = dvmResolveMethod(curMethod->clazz, ref,METHOD_VIRTUAL); 1400 if (baseMethod == NULL) { 1401 ILOGV("+ unknown method or access denied"); 1402 GOTO_exceptionThrown(); 1403 } 1404 } 1405 1406 /* 1407 * Combine the object we found with the vtable offset in the 1408 * method. 1409 */ 1410 assert(baseMethod->methodIndex < thisPtr->clazz->vtableCount); 1411 methodToCall = thisPtr->clazz->vtable[baseMethod->methodIndex]; 1412 1413 #if defined(WITH_JIT) && defined(MTERP_STUB) 1414 self->methodToCall = methodToCall; 1415 self->callsiteClass = thisPtr->clazz; 1416 #endif 1417 1418 #if 0 1419 if (dvmIsAbstractMethod(methodToCall)) { 1420 /* 1421 * This can happen if you create two classes, Base and Sub, where 1422 * Sub is a sub-class of Base. Declare a protected abstract 1423 * method foo() in Base, and invoke foo() from a method in Base. 1424 * Base is an "abstract base class" and is never instantiated 1425 * directly. Now, Override foo() in Sub, and use Sub. This 1426 * Works fine unless Sub stops providing an implementation of 1427 * the method. 1428 */ 1429 dvmThrowAbstractMethodError("abstract method not implemented"); 1430 GOTO_exceptionThrown(); 1431 } 1432 #else 1433 assert(!dvmIsAbstractMethod(methodToCall) || 1434 methodToCall->nativeFunc != NULL); 1435 #endif 1436 1437 LOGVV("+++ base=%s.%s virtual[%d]=%s.%s", 1438 baseMethod->clazz->descriptor, baseMethod->name, 1439 (u4) baseMethod->methodIndex, 1440 methodToCall->clazz->descriptor, methodToCall->name); 1441 assert(methodToCall != NULL); 1442 1443 #if 0 1444 if (vsrc1 != methodToCall->insSize) { 1445 ALOGW("WRONG METHOD: base=%s.%s virtual[%d]=%s.%s", 1446 baseMethod->clazz->descriptor, baseMethod->name, 1447 (u4) baseMethod->methodIndex, 1448 methodToCall->clazz->descriptor, methodToCall->name); 1449 //dvmDumpClass(baseMethod->clazz); 1450 //dvmDumpClass(methodToCall->clazz); 1451 dvmDumpAllClasses(0); 1452 } 1453 #endif 1454 1455 GOTO_invokeMethod(methodCallRange, methodToCall, vsrc1, vdst); 1456 } 1457 GOTO_TARGET_END 1458 1459 GOTO_TARGET(invokeSuper, bool methodCallRange) 1460 { 1461 Method* baseMethod; 1462 u2 thisReg; 1463 1464 EXPORT_PC(); 1465 1466 vsrc1 = INST_AA(inst); /* AA (count) or BA (count + arg 5) */ 1467 ref = FETCH(1); /* method ref */ 1468 vdst = FETCH(2); /* 4 regs -or- first reg */ 1469 1470 if (methodCallRange) { 1471 ILOGV("|invoke-super-range args=%d @0x%04x {regs=v%d-v%d}", 1472 vsrc1, ref, vdst, vdst+vsrc1-1); 1473 thisReg = vdst; 1474 } else { 1475 ILOGV("|invoke-super args=%d @0x%04x {regs=0x%04x %x}", 1476 vsrc1 >> 4, ref, vdst, vsrc1 & 0x0f); 1477 thisReg = vdst & 0x0f; 1478 } 1479 1480 /* impossible in well-formed code, but we must check nevertheless */ 1481 if (!checkForNull((Object*) GET_REGISTER(thisReg))) 1482 GOTO_exceptionThrown(); 1483 1484 /* 1485 * Resolve the method. This is the correct method for the static 1486 * type of the object. We also verify access permissions here. 1487 * The first arg to dvmResolveMethod() is just the referring class 1488 * (used for class loaders and such), so we don't want to pass 1489 * the superclass into the resolution call. 1490 */ 1491 baseMethod = dvmDexGetResolvedMethod(methodClassDex, ref); 1492 if (baseMethod == NULL) { 1493 baseMethod = dvmResolveMethod(curMethod->clazz, ref,METHOD_VIRTUAL); 1494 if (baseMethod == NULL) { 1495 ILOGV("+ unknown method or access denied"); 1496 GOTO_exceptionThrown(); 1497 } 1498 } 1499 1500 /* 1501 * Combine the object we found with the vtable offset in the 1502 * method's class. 1503 * 1504 * We're using the current method's class' superclass, not the 1505 * superclass of "this". This is because we might be executing 1506 * in a method inherited from a superclass, and we want to run 1507 * in that class' superclass. 1508 */ 1509 if (baseMethod->methodIndex >= curMethod->clazz->super->vtableCount) { 1510 /* 1511 * Method does not exist in the superclass. Could happen if 1512 * superclass gets updated. 1513 */ 1514 dvmThrowNoSuchMethodError(baseMethod->name); 1515 GOTO_exceptionThrown(); 1516 } 1517 methodToCall = curMethod->clazz->super->vtable[baseMethod->methodIndex]; 1518 1519 #if 0 1520 if (dvmIsAbstractMethod(methodToCall)) { 1521 dvmThrowAbstractMethodError("abstract method not implemented"); 1522 GOTO_exceptionThrown(); 1523 } 1524 #else 1525 assert(!dvmIsAbstractMethod(methodToCall) || 1526 methodToCall->nativeFunc != NULL); 1527 #endif 1528 LOGVV("+++ base=%s.%s super-virtual=%s.%s", 1529 baseMethod->clazz->descriptor, baseMethod->name, 1530 methodToCall->clazz->descriptor, methodToCall->name); 1531 assert(methodToCall != NULL); 1532 1533 GOTO_invokeMethod(methodCallRange, methodToCall, vsrc1, vdst); 1534 } 1535 GOTO_TARGET_END 1536 1537 GOTO_TARGET(invokeInterface, bool methodCallRange) 1538 { 1539 Object* thisPtr; 1540 ClassObject* thisClass; 1541 1542 EXPORT_PC(); 1543 1544 vsrc1 = INST_AA(inst); /* AA (count) or BA (count + arg 5) */ 1545 ref = FETCH(1); /* method ref */ 1546 vdst = FETCH(2); /* 4 regs -or- first reg */ 1547 1548 /* 1549 * The object against which we are executing a method is always 1550 * in the first argument. 1551 */ 1552 if (methodCallRange) { 1553 assert(vsrc1 > 0); 1554 ILOGV("|invoke-interface-range args=%d @0x%04x {regs=v%d-v%d}", 1555 vsrc1, ref, vdst, vdst+vsrc1-1); 1556 thisPtr = (Object*) GET_REGISTER(vdst); 1557 } else { 1558 assert((vsrc1>>4) > 0); 1559 ILOGV("|invoke-interface args=%d @0x%04x {regs=0x%04x %x}", 1560 vsrc1 >> 4, ref, vdst, vsrc1 & 0x0f); 1561 thisPtr = (Object*) GET_REGISTER(vdst & 0x0f); 1562 } 1563 1564 if (!checkForNull(thisPtr)) 1565 GOTO_exceptionThrown(); 1566 1567 thisClass = thisPtr->clazz; 1568 1569 /* 1570 * Given a class and a method index, find the Method* with the 1571 * actual code we want to execute. 1572 */ 1573 methodToCall = dvmFindInterfaceMethodInCache(thisClass, ref, curMethod, 1574 methodClassDex); 1575 #if defined(WITH_JIT) && defined(MTERP_STUB) 1576 self->callsiteClass = thisClass; 1577 self->methodToCall = methodToCall; 1578 #endif 1579 if (methodToCall == NULL) { 1580 assert(dvmCheckException(self)); 1581 GOTO_exceptionThrown(); 1582 } 1583 1584 GOTO_invokeMethod(methodCallRange, methodToCall, vsrc1, vdst); 1585 } 1586 GOTO_TARGET_END 1587 1588 GOTO_TARGET(invokeDirect, bool methodCallRange) 1589 { 1590 u2 thisReg; 1591 1592 EXPORT_PC(); 1593 1594 vsrc1 = INST_AA(inst); /* AA (count) or BA (count + arg 5) */ 1595 ref = FETCH(1); /* method ref */ 1596 vdst = FETCH(2); /* 4 regs -or- first reg */ 1597 1598 if (methodCallRange) { 1599 ILOGV("|invoke-direct-range args=%d @0x%04x {regs=v%d-v%d}", 1600 vsrc1, ref, vdst, vdst+vsrc1-1); 1601 thisReg = vdst; 1602 } else { 1603 ILOGV("|invoke-direct args=%d @0x%04x {regs=0x%04x %x}", 1604 vsrc1 >> 4, ref, vdst, vsrc1 & 0x0f); 1605 thisReg = vdst & 0x0f; 1606 } 1607 1608 if (!checkForNull((Object*) GET_REGISTER(thisReg))) 1609 GOTO_exceptionThrown(); 1610 1611 methodToCall = dvmDexGetResolvedMethod(methodClassDex, ref); 1612 if (methodToCall == NULL) { 1613 methodToCall = dvmResolveMethod(curMethod->clazz, ref, 1614 METHOD_DIRECT); 1615 if (methodToCall == NULL) { 1616 ILOGV("+ unknown direct method"); // should be impossible 1617 GOTO_exceptionThrown(); 1618 } 1619 } 1620 GOTO_invokeMethod(methodCallRange, methodToCall, vsrc1, vdst); 1621 } 1622 GOTO_TARGET_END 1623 1624 GOTO_TARGET(invokeStatic, bool methodCallRange) 1625 EXPORT_PC(); 1626 1627 vsrc1 = INST_AA(inst); /* AA (count) or BA (count + arg 5) */ 1628 ref = FETCH(1); /* method ref */ 1629 vdst = FETCH(2); /* 4 regs -or- first reg */ 1630 1631 if (methodCallRange) 1632 ILOGV("|invoke-static-range args=%d @0x%04x {regs=v%d-v%d}", 1633 vsrc1, ref, vdst, vdst+vsrc1-1); 1634 else 1635 ILOGV("|invoke-static args=%d @0x%04x {regs=0x%04x %x}", 1636 vsrc1 >> 4, ref, vdst, vsrc1 & 0x0f); 1637 1638 methodToCall = dvmDexGetResolvedMethod(methodClassDex, ref); 1639 if (methodToCall == NULL) { 1640 methodToCall = dvmResolveMethod(curMethod->clazz, ref, METHOD_STATIC); 1641 if (methodToCall == NULL) { 1642 ILOGV("+ unknown method"); 1643 GOTO_exceptionThrown(); 1644 } 1645 1646 #if defined(WITH_JIT) && defined(MTERP_STUB) 1647 /* 1648 * The JIT needs dvmDexGetResolvedMethod() to return non-null. 1649 * Include the check if this code is being used as a stub 1650 * called from the assembly interpreter. 1651 */ 1652 if ((self->interpBreak.ctl.subMode & kSubModeJitTraceBuild) && 1653 (dvmDexGetResolvedMethod(methodClassDex, ref) == NULL)) { 1654 /* Class initialization is still ongoing */ 1655 dvmJitEndTraceSelect(self,pc); 1656 } 1657 #endif 1658 } 1659 GOTO_invokeMethod(methodCallRange, methodToCall, vsrc1, vdst); 1660 GOTO_TARGET_END 1661 1662 GOTO_TARGET(invokeVirtualQuick, bool methodCallRange) 1663 { 1664 Object* thisPtr; 1665 1666 EXPORT_PC(); 1667 1668 vsrc1 = INST_AA(inst); /* AA (count) or BA (count + arg 5) */ 1669 ref = FETCH(1); /* vtable index */ 1670 vdst = FETCH(2); /* 4 regs -or- first reg */ 1671 1672 /* 1673 * The object against which we are executing a method is always 1674 * in the first argument. 1675 */ 1676 if (methodCallRange) { 1677 assert(vsrc1 > 0); 1678 ILOGV("|invoke-virtual-quick-range args=%d @0x%04x {regs=v%d-v%d}", 1679 vsrc1, ref, vdst, vdst+vsrc1-1); 1680 thisPtr = (Object*) GET_REGISTER(vdst); 1681 } else { 1682 assert((vsrc1>>4) > 0); 1683 ILOGV("|invoke-virtual-quick args=%d @0x%04x {regs=0x%04x %x}", 1684 vsrc1 >> 4, ref, vdst, vsrc1 & 0x0f); 1685 thisPtr = (Object*) GET_REGISTER(vdst & 0x0f); 1686 } 1687 1688 if (!checkForNull(thisPtr)) 1689 GOTO_exceptionThrown(); 1690 1691 1692 /* 1693 * Combine the object we found with the vtable offset in the 1694 * method. 1695 */ 1696 assert(ref < (unsigned int) thisPtr->clazz->vtableCount); 1697 methodToCall = thisPtr->clazz->vtable[ref]; 1698 #if defined(WITH_JIT) && defined(MTERP_STUB) 1699 self->callsiteClass = thisPtr->clazz; 1700 self->methodToCall = methodToCall; 1701 #endif 1702 1703 #if 0 1704 if (dvmIsAbstractMethod(methodToCall)) { 1705 dvmThrowAbstractMethodError("abstract method not implemented"); 1706 GOTO_exceptionThrown(); 1707 } 1708 #else 1709 assert(!dvmIsAbstractMethod(methodToCall) || 1710 methodToCall->nativeFunc != NULL); 1711 #endif 1712 1713 LOGVV("+++ virtual[%d]=%s.%s", 1714 ref, methodToCall->clazz->descriptor, methodToCall->name); 1715 assert(methodToCall != NULL); 1716 1717 GOTO_invokeMethod(methodCallRange, methodToCall, vsrc1, vdst); 1718 } 1719 GOTO_TARGET_END 1720 1721 GOTO_TARGET(invokeSuperQuick, bool methodCallRange) 1722 { 1723 u2 thisReg; 1724 1725 EXPORT_PC(); 1726 1727 vsrc1 = INST_AA(inst); /* AA (count) or BA (count + arg 5) */ 1728 ref = FETCH(1); /* vtable index */ 1729 vdst = FETCH(2); /* 4 regs -or- first reg */ 1730 1731 if (methodCallRange) { 1732 ILOGV("|invoke-super-quick-range args=%d @0x%04x {regs=v%d-v%d}", 1733 vsrc1, ref, vdst, vdst+vsrc1-1); 1734 thisReg = vdst; 1735 } else { 1736 ILOGV("|invoke-super-quick args=%d @0x%04x {regs=0x%04x %x}", 1737 vsrc1 >> 4, ref, vdst, vsrc1 & 0x0f); 1738 thisReg = vdst & 0x0f; 1739 } 1740 /* impossible in well-formed code, but we must check nevertheless */ 1741 if (!checkForNull((Object*) GET_REGISTER(thisReg))) 1742 GOTO_exceptionThrown(); 1743 1744 #if 0 /* impossible in optimized + verified code */ 1745 if (ref >= curMethod->clazz->super->vtableCount) { 1746 dvmThrowNoSuchMethodError(NULL); 1747 GOTO_exceptionThrown(); 1748 } 1749 #else 1750 assert(ref < (unsigned int) curMethod->clazz->super->vtableCount); 1751 #endif 1752 1753 /* 1754 * Combine the object we found with the vtable offset in the 1755 * method's class. 1756 * 1757 * We're using the current method's class' superclass, not the 1758 * superclass of "this". This is because we might be executing 1759 * in a method inherited from a superclass, and we want to run 1760 * in the method's class' superclass. 1761 */ 1762 methodToCall = curMethod->clazz->super->vtable[ref]; 1763 1764 #if 0 1765 if (dvmIsAbstractMethod(methodToCall)) { 1766 dvmThrowAbstractMethodError("abstract method not implemented"); 1767 GOTO_exceptionThrown(); 1768 } 1769 #else 1770 assert(!dvmIsAbstractMethod(methodToCall) || 1771 methodToCall->nativeFunc != NULL); 1772 #endif 1773 LOGVV("+++ super-virtual[%d]=%s.%s", 1774 ref, methodToCall->clazz->descriptor, methodToCall->name); 1775 assert(methodToCall != NULL); 1776 GOTO_invokeMethod(methodCallRange, methodToCall, vsrc1, vdst); 1777 } 1778 GOTO_TARGET_END 1779 1780 1781 /* 1782 * General handling for return-void, return, and return-wide. Put the 1783 * return value in "retval" before jumping here. 1784 */ 1785 GOTO_TARGET(returnFromMethod) 1786 { 1787 StackSaveArea* saveArea; 1788 1789 /* 1790 * We must do this BEFORE we pop the previous stack frame off, so 1791 * that the GC can see the return value (if any) in the local vars. 1792 * 1793 * Since this is now an interpreter switch point, we must do it before 1794 * we do anything at all. 1795 */ 1796 PERIODIC_CHECKS(0); 1797 1798 ILOGV("> retval=0x%llx (leaving %s.%s %s)", 1799 retval.j, curMethod->clazz->descriptor, curMethod->name, 1800 curMethod->shorty); 1801 //DUMP_REGS(curMethod, fp); 1802 1803 saveArea = SAVEAREA_FROM_FP(fp); 1804 1805 #ifdef EASY_GDB 1806 debugSaveArea = saveArea; 1807 #endif 1808 1809 /* back up to previous frame and see if we hit a break */ 1810 fp = (u4*)saveArea->prevFrame; 1811 assert(fp != NULL); 1812 1813 /* Handle any special subMode requirements */ 1814 if (self->interpBreak.ctl.subMode != 0) { 1815 PC_FP_TO_SELF(); 1816 dvmReportReturn(self); 1817 } 1818 1819 if (dvmIsBreakFrame(fp)) { 1820 /* bail without popping the method frame from stack */ 1821 LOGVV("+++ returned into break frame"); 1822 GOTO_bail(); 1823 } 1824 1825 /* update thread FP, and reset local variables */ 1826 self->interpSave.curFrame = fp; 1827 curMethod = SAVEAREA_FROM_FP(fp)->method; 1828 self->interpSave.method = curMethod; 1829 //methodClass = curMethod->clazz; 1830 methodClassDex = curMethod->clazz->pDvmDex; 1831 pc = saveArea->savedPc; 1832 ILOGD("> (return to %s.%s %s)", curMethod->clazz->descriptor, 1833 curMethod->name, curMethod->shorty); 1834 1835 /* use FINISH on the caller's invoke instruction */ 1836 //u2 invokeInstr = INST_INST(FETCH(0)); 1837 if (true /*invokeInstr >= OP_INVOKE_VIRTUAL && 1838 invokeInstr <= OP_INVOKE_INTERFACE*/) 1839 { 1840 FINISH(3); 1841 } else { 1842 //ALOGE("Unknown invoke instr %02x at %d", 1843 // invokeInstr, (int) (pc - curMethod->insns)); 1844 assert(false); 1845 } 1846 } 1847 GOTO_TARGET_END 1848 1849 1850 /* 1851 * Jump here when the code throws an exception. 1852 * 1853 * By the time we get here, the Throwable has been created and the stack 1854 * trace has been saved off. 1855 */ 1856 GOTO_TARGET(exceptionThrown) 1857 { 1858 Object* exception; 1859 int catchRelPc; 1860 1861 PERIODIC_CHECKS(0); 1862 1863 /* 1864 * We save off the exception and clear the exception status. While 1865 * processing the exception we might need to load some Throwable 1866 * classes, and we don't want class loader exceptions to get 1867 * confused with this one. 1868 */ 1869 assert(dvmCheckException(self)); 1870 exception = dvmGetException(self); 1871 dvmAddTrackedAlloc(exception, self); 1872 dvmClearException(self); 1873 1874 ALOGV("Handling exception %s at %s:%d", 1875 exception->clazz->descriptor, curMethod->name, 1876 dvmLineNumFromPC(curMethod, pc - curMethod->insns)); 1877 1878 /* 1879 * Report the exception throw to any "subMode" watchers. 1880 * 1881 * TODO: if the exception was thrown by interpreted code, control 1882 * fell through native, and then back to us, we will report the 1883 * exception at the point of the throw and again here. We can avoid 1884 * this by not reporting exceptions when we jump here directly from 1885 * the native call code above, but then we won't report exceptions 1886 * that were thrown *from* the JNI code (as opposed to *through* it). 1887 * 1888 * The correct solution is probably to ignore from-native exceptions 1889 * here, and have the JNI exception code do the reporting to the 1890 * debugger. 1891 */ 1892 if (self->interpBreak.ctl.subMode != 0) { 1893 PC_FP_TO_SELF(); 1894 dvmReportExceptionThrow(self, exception); 1895 } 1896 1897 /* 1898 * We need to unroll to the catch block or the nearest "break" 1899 * frame. 1900 * 1901 * A break frame could indicate that we have reached an intermediate 1902 * native call, or have gone off the top of the stack and the thread 1903 * needs to exit. Either way, we return from here, leaving the 1904 * exception raised. 1905 * 1906 * If we do find a catch block, we want to transfer execution to 1907 * that point. 1908 * 1909 * Note this can cause an exception while resolving classes in 1910 * the "catch" blocks. 1911 */ 1912 catchRelPc = dvmFindCatchBlock(self, pc - curMethod->insns, 1913 exception, false, (void**)(void*)&fp); 1914 1915 /* 1916 * Restore the stack bounds after an overflow. This isn't going to 1917 * be correct in all circumstances, e.g. if JNI code devours the 1918 * exception this won't happen until some other exception gets 1919 * thrown. If the code keeps pushing the stack bounds we'll end 1920 * up aborting the VM. 1921 * 1922 * Note we want to do this *after* the call to dvmFindCatchBlock, 1923 * because that may need extra stack space to resolve exception 1924 * classes (e.g. through a class loader). 1925 * 1926 * It's possible for the stack overflow handling to cause an 1927 * exception (specifically, class resolution in a "catch" block 1928 * during the call above), so we could see the thread's overflow 1929 * flag raised but actually be running in a "nested" interpreter 1930 * frame. We don't allow doubled-up StackOverflowErrors, so 1931 * we can check for this by just looking at the exception type 1932 * in the cleanup function. Also, we won't unroll past the SOE 1933 * point because the more-recent exception will hit a break frame 1934 * as it unrolls to here. 1935 */ 1936 if (self->stackOverflowed) 1937 dvmCleanupStackOverflow(self, exception); 1938 1939 if (catchRelPc < 0) { 1940 /* falling through to JNI code or off the bottom of the stack */ 1941 #if DVM_SHOW_EXCEPTION >= 2 1942 ALOGD("Exception %s from %s:%d not caught locally", 1943 exception->clazz->descriptor, dvmGetMethodSourceFile(curMethod), 1944 dvmLineNumFromPC(curMethod, pc - curMethod->insns)); 1945 #endif 1946 dvmSetException(self, exception); 1947 dvmReleaseTrackedAlloc(exception, self); 1948 GOTO_bail(); 1949 } 1950 1951 #if DVM_SHOW_EXCEPTION >= 3 1952 { 1953 const Method* catchMethod = SAVEAREA_FROM_FP(fp)->method; 1954 ALOGD("Exception %s thrown from %s:%d to %s:%d", 1955 exception->clazz->descriptor, dvmGetMethodSourceFile(curMethod), 1956 dvmLineNumFromPC(curMethod, pc - curMethod->insns), 1957 dvmGetMethodSourceFile(catchMethod), 1958 dvmLineNumFromPC(catchMethod, catchRelPc)); 1959 } 1960 #endif 1961 1962 /* 1963 * Adjust local variables to match self->interpSave.curFrame and the 1964 * updated PC. 1965 */ 1966 //fp = (u4*) self->interpSave.curFrame; 1967 curMethod = SAVEAREA_FROM_FP(fp)->method; 1968 self->interpSave.method = curMethod; 1969 //methodClass = curMethod->clazz; 1970 methodClassDex = curMethod->clazz->pDvmDex; 1971 pc = curMethod->insns + catchRelPc; 1972 ILOGV("> pc <-- %s.%s %s", curMethod->clazz->descriptor, 1973 curMethod->name, curMethod->shorty); 1974 DUMP_REGS(curMethod, fp, false); // show all regs 1975 1976 /* 1977 * Restore the exception if the handler wants it. 1978 * 1979 * The Dalvik spec mandates that, if an exception handler wants to 1980 * do something with the exception, the first instruction executed 1981 * must be "move-exception". We can pass the exception along 1982 * through the thread struct, and let the move-exception instruction 1983 * clear it for us. 1984 * 1985 * If the handler doesn't call move-exception, we don't want to 1986 * finish here with an exception still pending. 1987 */ 1988 if (INST_INST(FETCH(0)) == OP_MOVE_EXCEPTION) 1989 dvmSetException(self, exception); 1990 1991 dvmReleaseTrackedAlloc(exception, self); 1992 FINISH(0); 1993 } 1994 GOTO_TARGET_END 1995 1996 1997 1998 /* 1999 * General handling for invoke-{virtual,super,direct,static,interface}, 2000 * including "quick" variants. 2001 * 2002 * Set "methodToCall" to the Method we're calling, and "methodCallRange" 2003 * depending on whether this is a "/range" instruction. 2004 * 2005 * For a range call: 2006 * "vsrc1" holds the argument count (8 bits) 2007 * "vdst" holds the first argument in the range 2008 * For a non-range call: 2009 * "vsrc1" holds the argument count (4 bits) and the 5th argument index 2010 * "vdst" holds four 4-bit register indices 2011 * 2012 * The caller must EXPORT_PC before jumping here, because any method 2013 * call can throw a stack overflow exception. 2014 */ 2015 GOTO_TARGET(invokeMethod, bool methodCallRange, const Method* _methodToCall, 2016 u2 count, u2 regs) 2017 { 2018 STUB_HACK(vsrc1 = count; vdst = regs; methodToCall = _methodToCall;); 2019 2020 //printf("range=%d call=%p count=%d regs=0x%04x\n", 2021 // methodCallRange, methodToCall, count, regs); 2022 //printf(" --> %s.%s %s\n", methodToCall->clazz->descriptor, 2023 // methodToCall->name, methodToCall->shorty); 2024 2025 u4* outs; 2026 int i; 2027 2028 /* 2029 * Copy args. This may corrupt vsrc1/vdst. 2030 */ 2031 if (methodCallRange) { 2032 // could use memcpy or a "Duff's device"; most functions have 2033 // so few args it won't matter much 2034 assert(vsrc1 <= curMethod->outsSize); 2035 assert(vsrc1 == methodToCall->insSize); 2036 outs = OUTS_FROM_FP(fp, vsrc1); 2037 for (i = 0; i < vsrc1; i++) 2038 outs[i] = GET_REGISTER(vdst+i); 2039 } else { 2040 u4 count = vsrc1 >> 4; 2041 2042 assert(count <= curMethod->outsSize); 2043 assert(count == methodToCall->insSize); 2044 assert(count <= 5); 2045 2046 outs = OUTS_FROM_FP(fp, count); 2047 #if 0 2048 if (count == 5) { 2049 outs[4] = GET_REGISTER(vsrc1 & 0x0f); 2050 count--; 2051 } 2052 for (i = 0; i < (int) count; i++) { 2053 outs[i] = GET_REGISTER(vdst & 0x0f); 2054 vdst >>= 4; 2055 } 2056 #else 2057 // This version executes fewer instructions but is larger 2058 // overall. Seems to be a teensy bit faster. 2059 assert((vdst >> 16) == 0); // 16 bits -or- high 16 bits clear 2060 switch (count) { 2061 case 5: 2062 outs[4] = GET_REGISTER(vsrc1 & 0x0f); 2063 case 4: 2064 outs[3] = GET_REGISTER(vdst >> 12); 2065 case 3: 2066 outs[2] = GET_REGISTER((vdst & 0x0f00) >> 8); 2067 case 2: 2068 outs[1] = GET_REGISTER((vdst & 0x00f0) >> 4); 2069 case 1: 2070 outs[0] = GET_REGISTER(vdst & 0x0f); 2071 default: 2072 ; 2073 } 2074 #endif 2075 } 2076 } 2077 2078 /* 2079 * (This was originally a "goto" target; I've kept it separate from the 2080 * stuff above in case we want to refactor things again.) 2081 * 2082 * At this point, we have the arguments stored in the "outs" area of 2083 * the current method's stack frame, and the method to call in 2084 * "methodToCall". Push a new stack frame. 2085 */ 2086 { 2087 StackSaveArea* newSaveArea; 2088 u4* newFp; 2089 2090 ILOGV("> %s%s.%s %s", 2091 dvmIsNativeMethod(methodToCall) ? "(NATIVE) " : "", 2092 methodToCall->clazz->descriptor, methodToCall->name, 2093 methodToCall->shorty); 2094 2095 newFp = (u4*) SAVEAREA_FROM_FP(fp) - methodToCall->registersSize; 2096 newSaveArea = SAVEAREA_FROM_FP(newFp); 2097 2098 /* verify that we have enough space */ 2099 if (true) { 2100 u1* bottom; 2101 bottom = (u1*) newSaveArea - methodToCall->outsSize * sizeof(u4); 2102 if (bottom < self->interpStackEnd) { 2103 /* stack overflow */ 2104 ALOGV("Stack overflow on method call (start=%p end=%p newBot=%p(%d) size=%d '%s')", 2105 self->interpStackStart, self->interpStackEnd, bottom, 2106 (u1*) fp - bottom, self->interpStackSize, 2107 methodToCall->name); 2108 dvmHandleStackOverflow(self, methodToCall); 2109 assert(dvmCheckException(self)); 2110 GOTO_exceptionThrown(); 2111 } 2112 //ALOGD("+++ fp=%p newFp=%p newSave=%p bottom=%p", 2113 // fp, newFp, newSaveArea, bottom); 2114 } 2115 2116 #ifdef LOG_INSTR 2117 if (methodToCall->registersSize > methodToCall->insSize) { 2118 /* 2119 * This makes valgrind quiet when we print registers that 2120 * haven't been initialized. Turn it off when the debug 2121 * messages are disabled -- we want valgrind to report any 2122 * used-before-initialized issues. 2123 */ 2124 memset(newFp, 0xcc, 2125 (methodToCall->registersSize - methodToCall->insSize) * 4); 2126 } 2127 #endif 2128 2129 #ifdef EASY_GDB 2130 newSaveArea->prevSave = SAVEAREA_FROM_FP(fp); 2131 #endif 2132 newSaveArea->prevFrame = fp; 2133 newSaveArea->savedPc = pc; 2134 #if defined(WITH_JIT) && defined(MTERP_STUB) 2135 newSaveArea->returnAddr = 0; 2136 #endif 2137 newSaveArea->method = methodToCall; 2138 2139 if (self->interpBreak.ctl.subMode != 0) { 2140 /* 2141 * We mark ENTER here for both native and non-native 2142 * calls. For native calls, we'll mark EXIT on return. 2143 * For non-native calls, EXIT is marked in the RETURN op. 2144 */ 2145 PC_TO_SELF(); 2146 dvmReportInvoke(self, methodToCall); 2147 } 2148 2149 if (!dvmIsNativeMethod(methodToCall)) { 2150 /* 2151 * "Call" interpreted code. Reposition the PC, update the 2152 * frame pointer and other local state, and continue. 2153 */ 2154 curMethod = methodToCall; 2155 self->interpSave.method = curMethod; 2156 methodClassDex = curMethod->clazz->pDvmDex; 2157 pc = methodToCall->insns; 2158 fp = newFp; 2159 self->interpSave.curFrame = fp; 2160 #ifdef EASY_GDB 2161 debugSaveArea = SAVEAREA_FROM_FP(newFp); 2162 #endif 2163 self->debugIsMethodEntry = true; // profiling, debugging 2164 ILOGD("> pc <-- %s.%s %s", curMethod->clazz->descriptor, 2165 curMethod->name, curMethod->shorty); 2166 DUMP_REGS(curMethod, fp, true); // show input args 2167 FINISH(0); // jump to method start 2168 } else { 2169 /* set this up for JNI locals, even if not a JNI native */ 2170 newSaveArea->xtra.localRefCookie = self->jniLocalRefTable.segmentState.all; 2171 2172 self->interpSave.curFrame = newFp; 2173 2174 DUMP_REGS(methodToCall, newFp, true); // show input args 2175 2176 if (self->interpBreak.ctl.subMode != 0) { 2177 dvmReportPreNativeInvoke(methodToCall, self, fp); 2178 } 2179 2180 ILOGD("> native <-- %s.%s %s", methodToCall->clazz->descriptor, 2181 methodToCall->name, methodToCall->shorty); 2182 2183 /* 2184 * Jump through native call bridge. Because we leave no 2185 * space for locals on native calls, "newFp" points directly 2186 * to the method arguments. 2187 */ 2188 (*methodToCall->nativeFunc)(newFp, &retval, methodToCall, self); 2189 2190 if (self->interpBreak.ctl.subMode != 0) { 2191 dvmReportPostNativeInvoke(methodToCall, self, fp); 2192 } 2193 2194 /* pop frame off */ 2195 dvmPopJniLocals(self, newSaveArea); 2196 self->interpSave.curFrame = fp; 2197 2198 /* 2199 * If the native code threw an exception, or interpreted code 2200 * invoked by the native call threw one and nobody has cleared 2201 * it, jump to our local exception handling. 2202 */ 2203 if (dvmCheckException(self)) { 2204 ALOGV("Exception thrown by/below native code"); 2205 GOTO_exceptionThrown(); 2206 } 2207 2208 ILOGD("> retval=0x%llx (leaving native)", retval.j); 2209 ILOGD("> (return from native %s.%s to %s.%s %s)", 2210 methodToCall->clazz->descriptor, methodToCall->name, 2211 curMethod->clazz->descriptor, curMethod->name, 2212 curMethod->shorty); 2213 2214 //u2 invokeInstr = INST_INST(FETCH(0)); 2215 if (true /*invokeInstr >= OP_INVOKE_VIRTUAL && 2216 invokeInstr <= OP_INVOKE_INTERFACE*/) 2217 { 2218 FINISH(3); 2219 } else { 2220 //ALOGE("Unknown invoke instr %02x at %d", 2221 // invokeInstr, (int) (pc - curMethod->insns)); 2222 assert(false); 2223 } 2224 } 2225 } 2226 assert(false); // should not get here 2227 GOTO_TARGET_END 2228 2229 /* File: cstubs/enddefs.cpp */ 2230 2231 /* undefine "magic" name remapping */ 2232 #undef retval 2233 #undef pc 2234 #undef fp 2235 #undef curMethod 2236 #undef methodClassDex 2237 #undef self 2238 #undef debugTrackedRefStart 2239 2240