Home | History | Annotate | Download | only in runtime
      1 /*
      2  * Copyright (C) 2008 The Android Open Source Project
      3  *
      4  * Licensed under the Apache License, Version 2.0 (the "License");
      5  * you may not use this file except in compliance with the License.
      6  * You may obtain a copy of the License at
      7  *
      8  *      http://www.apache.org/licenses/LICENSE-2.0
      9  *
     10  * Unless required by applicable law or agreed to in writing, software
     11  * distributed under the License is distributed on an "AS IS" BASIS,
     12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
     13  * See the License for the specific language governing permissions and
     14  * limitations under the License.
     15  */
     16 
     17 #include "debugger.h"
     18 
     19 #include <sys/uio.h>
     20 
     21 #include <set>
     22 
     23 #include "arch/context.h"
     24 #include "class_linker.h"
     25 #include "class_linker-inl.h"
     26 #include "dex_file-inl.h"
     27 #include "dex_instruction.h"
     28 #include "gc/accounting/card_table-inl.h"
     29 #include "gc/space/large_object_space.h"
     30 #include "gc/space/space-inl.h"
     31 #include "invoke_arg_array_builder.h"
     32 #include "jdwp/object_registry.h"
     33 #include "mirror/art_field-inl.h"
     34 #include "mirror/art_method-inl.h"
     35 #include "mirror/class.h"
     36 #include "mirror/class-inl.h"
     37 #include "mirror/class_loader.h"
     38 #include "mirror/object-inl.h"
     39 #include "mirror/object_array-inl.h"
     40 #include "mirror/throwable.h"
     41 #include "object_utils.h"
     42 #include "safe_map.h"
     43 #include "scoped_thread_state_change.h"
     44 #include "ScopedLocalRef.h"
     45 #include "ScopedPrimitiveArray.h"
     46 #include "sirt_ref.h"
     47 #include "stack_indirect_reference_table.h"
     48 #include "thread_list.h"
     49 #include "throw_location.h"
     50 #include "utf.h"
     51 #include "well_known_classes.h"
     52 
     53 #ifdef HAVE_ANDROID_OS
     54 #include "cutils/properties.h"
     55 #endif
     56 
     57 namespace art {
     58 
     59 static const size_t kMaxAllocRecordStackDepth = 16;  // Max 255.
     60 static const size_t kDefaultNumAllocRecords = 64*1024;  // Must be a power of 2.
     61 
     62 struct AllocRecordStackTraceElement {
     63   mirror::ArtMethod* method;
     64   uint32_t dex_pc;
     65 
     66   int32_t LineNumber() const SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
     67     return MethodHelper(method).GetLineNumFromDexPC(dex_pc);
     68   }
     69 };
     70 
     71 struct AllocRecord {
     72   mirror::Class* type;
     73   size_t byte_count;
     74   uint16_t thin_lock_id;
     75   AllocRecordStackTraceElement stack[kMaxAllocRecordStackDepth];  // Unused entries have NULL method.
     76 
     77   size_t GetDepth() {
     78     size_t depth = 0;
     79     while (depth < kMaxAllocRecordStackDepth && stack[depth].method != NULL) {
     80       ++depth;
     81     }
     82     return depth;
     83   }
     84 };
     85 
     86 struct Breakpoint {
     87   mirror::ArtMethod* method;
     88   uint32_t dex_pc;
     89   Breakpoint(mirror::ArtMethod* method, uint32_t dex_pc) : method(method), dex_pc(dex_pc) {}
     90 };
     91 
     92 static std::ostream& operator<<(std::ostream& os, const Breakpoint& rhs)
     93     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
     94   os << StringPrintf("Breakpoint[%s @%#x]", PrettyMethod(rhs.method).c_str(), rhs.dex_pc);
     95   return os;
     96 }
     97 
     98 struct SingleStepControl {
     99   // Are we single-stepping right now?
    100   bool is_active;
    101   Thread* thread;
    102 
    103   JDWP::JdwpStepSize step_size;
    104   JDWP::JdwpStepDepth step_depth;
    105 
    106   const mirror::ArtMethod* method;
    107   int32_t line_number;  // Or -1 for native methods.
    108   std::set<uint32_t> dex_pcs;
    109   int stack_depth;
    110 };
    111 
    112 class DebugInstrumentationListener : public instrumentation::InstrumentationListener {
    113  public:
    114   DebugInstrumentationListener() {}
    115   virtual ~DebugInstrumentationListener() {}
    116 
    117   virtual void MethodEntered(Thread* thread, mirror::Object* this_object,
    118                              const mirror::ArtMethod* method, uint32_t dex_pc)
    119       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    120     if (method->IsNative()) {
    121       // TODO: post location events is a suspension point and native method entry stubs aren't.
    122       return;
    123     }
    124     Dbg::PostLocationEvent(method, 0, this_object, Dbg::kMethodEntry);
    125   }
    126 
    127   virtual void MethodExited(Thread* thread, mirror::Object* this_object,
    128                             const mirror::ArtMethod* method,
    129                             uint32_t dex_pc, const JValue& return_value)
    130       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    131     UNUSED(return_value);
    132     if (method->IsNative()) {
    133       // TODO: post location events is a suspension point and native method entry stubs aren't.
    134       return;
    135     }
    136     Dbg::PostLocationEvent(method, dex_pc, this_object, Dbg::kMethodExit);
    137   }
    138 
    139   virtual void MethodUnwind(Thread* thread, const mirror::ArtMethod* method,
    140                             uint32_t dex_pc) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    141     // We're not recorded to listen to this kind of event, so complain.
    142     LOG(ERROR) << "Unexpected method unwind event in debugger " << PrettyMethod(method)
    143         << " " << dex_pc;
    144   }
    145 
    146   virtual void DexPcMoved(Thread* thread, mirror::Object* this_object,
    147                           const mirror::ArtMethod* method, uint32_t new_dex_pc)
    148       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    149     Dbg::UpdateDebugger(thread, this_object, method, new_dex_pc);
    150   }
    151 
    152   virtual void ExceptionCaught(Thread* thread, const ThrowLocation& throw_location,
    153                                mirror::ArtMethod* catch_method, uint32_t catch_dex_pc,
    154                                mirror::Throwable* exception_object)
    155       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    156     Dbg::PostException(thread, throw_location, catch_method, catch_dex_pc, exception_object);
    157   }
    158 } gDebugInstrumentationListener;
    159 
    160 // JDWP is allowed unless the Zygote forbids it.
    161 static bool gJdwpAllowed = true;
    162 
    163 // Was there a -Xrunjdwp or -agentlib:jdwp= argument on the command line?
    164 static bool gJdwpConfigured = false;
    165 
    166 // Broken-down JDWP options. (Only valid if IsJdwpConfigured() is true.)
    167 static JDWP::JdwpOptions gJdwpOptions;
    168 
    169 // Runtime JDWP state.
    170 static JDWP::JdwpState* gJdwpState = NULL;
    171 static bool gDebuggerConnected;  // debugger or DDMS is connected.
    172 static bool gDebuggerActive;     // debugger is making requests.
    173 static bool gDisposed;           // debugger called VirtualMachine.Dispose, so we should drop the connection.
    174 
    175 static bool gDdmThreadNotification = false;
    176 
    177 // DDMS GC-related settings.
    178 static Dbg::HpifWhen gDdmHpifWhen = Dbg::HPIF_WHEN_NEVER;
    179 static Dbg::HpsgWhen gDdmHpsgWhen = Dbg::HPSG_WHEN_NEVER;
    180 static Dbg::HpsgWhat gDdmHpsgWhat;
    181 static Dbg::HpsgWhen gDdmNhsgWhen = Dbg::HPSG_WHEN_NEVER;
    182 static Dbg::HpsgWhat gDdmNhsgWhat;
    183 
    184 static ObjectRegistry* gRegistry = NULL;
    185 
    186 // Recent allocation tracking.
    187 static Mutex gAllocTrackerLock DEFAULT_MUTEX_ACQUIRED_AFTER("AllocTracker lock");
    188 AllocRecord* Dbg::recent_allocation_records_ PT_GUARDED_BY(gAllocTrackerLock) = NULL;  // TODO: CircularBuffer<AllocRecord>
    189 static size_t gAllocRecordMax GUARDED_BY(gAllocTrackerLock) = 0;
    190 static size_t gAllocRecordHead GUARDED_BY(gAllocTrackerLock) = 0;
    191 static size_t gAllocRecordCount GUARDED_BY(gAllocTrackerLock) = 0;
    192 
    193 // Breakpoints and single-stepping.
    194 static std::vector<Breakpoint> gBreakpoints GUARDED_BY(Locks::breakpoint_lock_);
    195 static SingleStepControl gSingleStepControl GUARDED_BY(Locks::breakpoint_lock_);
    196 
    197 static bool IsBreakpoint(const mirror::ArtMethod* m, uint32_t dex_pc)
    198     LOCKS_EXCLUDED(Locks::breakpoint_lock_)
    199     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    200   MutexLock mu(Thread::Current(), *Locks::breakpoint_lock_);
    201   for (size_t i = 0; i < gBreakpoints.size(); ++i) {
    202     if (gBreakpoints[i].method == m && gBreakpoints[i].dex_pc == dex_pc) {
    203       VLOG(jdwp) << "Hit breakpoint #" << i << ": " << gBreakpoints[i];
    204       return true;
    205     }
    206   }
    207   return false;
    208 }
    209 
    210 static bool IsSuspendedForDebugger(ScopedObjectAccessUnchecked& soa, Thread* thread) {
    211   MutexLock mu(soa.Self(), *Locks::thread_suspend_count_lock_);
    212   // A thread may be suspended for GC; in this code, we really want to know whether
    213   // there's a debugger suspension active.
    214   return thread->IsSuspended() && thread->GetDebugSuspendCount() > 0;
    215 }
    216 
    217 static mirror::Array* DecodeArray(JDWP::RefTypeId id, JDWP::JdwpError& status)
    218     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    219   mirror::Object* o = gRegistry->Get<mirror::Object*>(id);
    220   if (o == NULL || o == ObjectRegistry::kInvalidObject) {
    221     status = JDWP::ERR_INVALID_OBJECT;
    222     return NULL;
    223   }
    224   if (!o->IsArrayInstance()) {
    225     status = JDWP::ERR_INVALID_ARRAY;
    226     return NULL;
    227   }
    228   status = JDWP::ERR_NONE;
    229   return o->AsArray();
    230 }
    231 
    232 static mirror::Class* DecodeClass(JDWP::RefTypeId id, JDWP::JdwpError& status)
    233     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    234   mirror::Object* o = gRegistry->Get<mirror::Object*>(id);
    235   if (o == NULL || o == ObjectRegistry::kInvalidObject) {
    236     status = JDWP::ERR_INVALID_OBJECT;
    237     return NULL;
    238   }
    239   if (!o->IsClass()) {
    240     status = JDWP::ERR_INVALID_CLASS;
    241     return NULL;
    242   }
    243   status = JDWP::ERR_NONE;
    244   return o->AsClass();
    245 }
    246 
    247 static JDWP::JdwpError DecodeThread(ScopedObjectAccessUnchecked& soa, JDWP::ObjectId thread_id, Thread*& thread)
    248     EXCLUSIVE_LOCKS_REQUIRED(Locks::thread_list_lock_)
    249     LOCKS_EXCLUDED(Locks::thread_suspend_count_lock_)
    250     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    251   mirror::Object* thread_peer = gRegistry->Get<mirror::Object*>(thread_id);
    252   if (thread_peer == NULL || thread_peer == ObjectRegistry::kInvalidObject) {
    253     // This isn't even an object.
    254     return JDWP::ERR_INVALID_OBJECT;
    255   }
    256 
    257   mirror::Class* java_lang_Thread = soa.Decode<mirror::Class*>(WellKnownClasses::java_lang_Thread);
    258   if (!java_lang_Thread->IsAssignableFrom(thread_peer->GetClass())) {
    259     // This isn't a thread.
    260     return JDWP::ERR_INVALID_THREAD;
    261   }
    262 
    263   thread = Thread::FromManagedThread(soa, thread_peer);
    264   if (thread == NULL) {
    265     // This is a java.lang.Thread without a Thread*. Must be a zombie.
    266     return JDWP::ERR_THREAD_NOT_ALIVE;
    267   }
    268   return JDWP::ERR_NONE;
    269 }
    270 
    271 static JDWP::JdwpTag BasicTagFromDescriptor(const char* descriptor) {
    272   // JDWP deliberately uses the descriptor characters' ASCII values for its enum.
    273   // Note that by "basic" we mean that we don't get more specific than JT_OBJECT.
    274   return static_cast<JDWP::JdwpTag>(descriptor[0]);
    275 }
    276 
    277 static JDWP::JdwpTag TagFromClass(mirror::Class* c)
    278     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    279   CHECK(c != NULL);
    280   if (c->IsArrayClass()) {
    281     return JDWP::JT_ARRAY;
    282   }
    283 
    284   ClassLinker* class_linker = Runtime::Current()->GetClassLinker();
    285   if (c->IsStringClass()) {
    286     return JDWP::JT_STRING;
    287   } else if (c->IsClassClass()) {
    288     return JDWP::JT_CLASS_OBJECT;
    289   } else if (class_linker->FindSystemClass("Ljava/lang/Thread;")->IsAssignableFrom(c)) {
    290     return JDWP::JT_THREAD;
    291   } else if (class_linker->FindSystemClass("Ljava/lang/ThreadGroup;")->IsAssignableFrom(c)) {
    292     return JDWP::JT_THREAD_GROUP;
    293   } else if (class_linker->FindSystemClass("Ljava/lang/ClassLoader;")->IsAssignableFrom(c)) {
    294     return JDWP::JT_CLASS_LOADER;
    295   } else {
    296     return JDWP::JT_OBJECT;
    297   }
    298 }
    299 
    300 /*
    301  * Objects declared to hold Object might actually hold a more specific
    302  * type.  The debugger may take a special interest in these (e.g. it
    303  * wants to display the contents of Strings), so we want to return an
    304  * appropriate tag.
    305  *
    306  * Null objects are tagged JT_OBJECT.
    307  */
    308 static JDWP::JdwpTag TagFromObject(const mirror::Object* o)
    309     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    310   return (o == NULL) ? JDWP::JT_OBJECT : TagFromClass(o->GetClass());
    311 }
    312 
    313 static bool IsPrimitiveTag(JDWP::JdwpTag tag) {
    314   switch (tag) {
    315   case JDWP::JT_BOOLEAN:
    316   case JDWP::JT_BYTE:
    317   case JDWP::JT_CHAR:
    318   case JDWP::JT_FLOAT:
    319   case JDWP::JT_DOUBLE:
    320   case JDWP::JT_INT:
    321   case JDWP::JT_LONG:
    322   case JDWP::JT_SHORT:
    323   case JDWP::JT_VOID:
    324     return true;
    325   default:
    326     return false;
    327   }
    328 }
    329 
    330 /*
    331  * Handle one of the JDWP name/value pairs.
    332  *
    333  * JDWP options are:
    334  *  help: if specified, show help message and bail
    335  *  transport: may be dt_socket or dt_shmem
    336  *  address: for dt_socket, "host:port", or just "port" when listening
    337  *  server: if "y", wait for debugger to attach; if "n", attach to debugger
    338  *  timeout: how long to wait for debugger to connect / listen
    339  *
    340  * Useful with server=n (these aren't supported yet):
    341  *  onthrow=<exception-name>: connect to debugger when exception thrown
    342  *  onuncaught=y|n: connect to debugger when uncaught exception thrown
    343  *  launch=<command-line>: launch the debugger itself
    344  *
    345  * The "transport" option is required, as is "address" if server=n.
    346  */
    347 static bool ParseJdwpOption(const std::string& name, const std::string& value) {
    348   if (name == "transport") {
    349     if (value == "dt_socket") {
    350       gJdwpOptions.transport = JDWP::kJdwpTransportSocket;
    351     } else if (value == "dt_android_adb") {
    352       gJdwpOptions.transport = JDWP::kJdwpTransportAndroidAdb;
    353     } else {
    354       LOG(ERROR) << "JDWP transport not supported: " << value;
    355       return false;
    356     }
    357   } else if (name == "server") {
    358     if (value == "n") {
    359       gJdwpOptions.server = false;
    360     } else if (value == "y") {
    361       gJdwpOptions.server = true;
    362     } else {
    363       LOG(ERROR) << "JDWP option 'server' must be 'y' or 'n'";
    364       return false;
    365     }
    366   } else if (name == "suspend") {
    367     if (value == "n") {
    368       gJdwpOptions.suspend = false;
    369     } else if (value == "y") {
    370       gJdwpOptions.suspend = true;
    371     } else {
    372       LOG(ERROR) << "JDWP option 'suspend' must be 'y' or 'n'";
    373       return false;
    374     }
    375   } else if (name == "address") {
    376     /* this is either <port> or <host>:<port> */
    377     std::string port_string;
    378     gJdwpOptions.host.clear();
    379     std::string::size_type colon = value.find(':');
    380     if (colon != std::string::npos) {
    381       gJdwpOptions.host = value.substr(0, colon);
    382       port_string = value.substr(colon + 1);
    383     } else {
    384       port_string = value;
    385     }
    386     if (port_string.empty()) {
    387       LOG(ERROR) << "JDWP address missing port: " << value;
    388       return false;
    389     }
    390     char* end;
    391     uint64_t port = strtoul(port_string.c_str(), &end, 10);
    392     if (*end != '\0' || port > 0xffff) {
    393       LOG(ERROR) << "JDWP address has junk in port field: " << value;
    394       return false;
    395     }
    396     gJdwpOptions.port = port;
    397   } else if (name == "launch" || name == "onthrow" || name == "oncaught" || name == "timeout") {
    398     /* valid but unsupported */
    399     LOG(INFO) << "Ignoring JDWP option '" << name << "'='" << value << "'";
    400   } else {
    401     LOG(INFO) << "Ignoring unrecognized JDWP option '" << name << "'='" << value << "'";
    402   }
    403 
    404   return true;
    405 }
    406 
    407 /*
    408  * Parse the latter half of a -Xrunjdwp/-agentlib:jdwp= string, e.g.:
    409  * "transport=dt_socket,address=8000,server=y,suspend=n"
    410  */
    411 bool Dbg::ParseJdwpOptions(const std::string& options) {
    412   VLOG(jdwp) << "ParseJdwpOptions: " << options;
    413 
    414   std::vector<std::string> pairs;
    415   Split(options, ',', pairs);
    416 
    417   for (size_t i = 0; i < pairs.size(); ++i) {
    418     std::string::size_type equals = pairs[i].find('=');
    419     if (equals == std::string::npos) {
    420       LOG(ERROR) << "Can't parse JDWP option '" << pairs[i] << "' in '" << options << "'";
    421       return false;
    422     }
    423     ParseJdwpOption(pairs[i].substr(0, equals), pairs[i].substr(equals + 1));
    424   }
    425 
    426   if (gJdwpOptions.transport == JDWP::kJdwpTransportUnknown) {
    427     LOG(ERROR) << "Must specify JDWP transport: " << options;
    428   }
    429   if (!gJdwpOptions.server && (gJdwpOptions.host.empty() || gJdwpOptions.port == 0)) {
    430     LOG(ERROR) << "Must specify JDWP host and port when server=n: " << options;
    431     return false;
    432   }
    433 
    434   gJdwpConfigured = true;
    435   return true;
    436 }
    437 
    438 void Dbg::StartJdwp() {
    439   if (!gJdwpAllowed || !IsJdwpConfigured()) {
    440     // No JDWP for you!
    441     return;
    442   }
    443 
    444   CHECK(gRegistry == NULL);
    445   gRegistry = new ObjectRegistry;
    446 
    447   // Init JDWP if the debugger is enabled. This may connect out to a
    448   // debugger, passively listen for a debugger, or block waiting for a
    449   // debugger.
    450   gJdwpState = JDWP::JdwpState::Create(&gJdwpOptions);
    451   if (gJdwpState == NULL) {
    452     // We probably failed because some other process has the port already, which means that
    453     // if we don't abort the user is likely to think they're talking to us when they're actually
    454     // talking to that other process.
    455     LOG(FATAL) << "Debugger thread failed to initialize";
    456   }
    457 
    458   // If a debugger has already attached, send the "welcome" message.
    459   // This may cause us to suspend all threads.
    460   if (gJdwpState->IsActive()) {
    461     ScopedObjectAccess soa(Thread::Current());
    462     if (!gJdwpState->PostVMStart()) {
    463       LOG(WARNING) << "Failed to post 'start' message to debugger";
    464     }
    465   }
    466 }
    467 
    468 void Dbg::StopJdwp() {
    469   delete gJdwpState;
    470   delete gRegistry;
    471   gRegistry = NULL;
    472 }
    473 
    474 void Dbg::GcDidFinish() {
    475   if (gDdmHpifWhen != HPIF_WHEN_NEVER) {
    476     ScopedObjectAccess soa(Thread::Current());
    477     LOG(DEBUG) << "Sending heap info to DDM";
    478     DdmSendHeapInfo(gDdmHpifWhen);
    479   }
    480   if (gDdmHpsgWhen != HPSG_WHEN_NEVER) {
    481     ScopedObjectAccess soa(Thread::Current());
    482     LOG(DEBUG) << "Dumping heap to DDM";
    483     DdmSendHeapSegments(false);
    484   }
    485   if (gDdmNhsgWhen != HPSG_WHEN_NEVER) {
    486     ScopedObjectAccess soa(Thread::Current());
    487     LOG(DEBUG) << "Dumping native heap to DDM";
    488     DdmSendHeapSegments(true);
    489   }
    490 }
    491 
    492 void Dbg::SetJdwpAllowed(bool allowed) {
    493   gJdwpAllowed = allowed;
    494 }
    495 
    496 DebugInvokeReq* Dbg::GetInvokeReq() {
    497   return Thread::Current()->GetInvokeReq();
    498 }
    499 
    500 Thread* Dbg::GetDebugThread() {
    501   return (gJdwpState != NULL) ? gJdwpState->GetDebugThread() : NULL;
    502 }
    503 
    504 void Dbg::ClearWaitForEventThread() {
    505   gJdwpState->ClearWaitForEventThread();
    506 }
    507 
    508 void Dbg::Connected() {
    509   CHECK(!gDebuggerConnected);
    510   VLOG(jdwp) << "JDWP has attached";
    511   gDebuggerConnected = true;
    512   gDisposed = false;
    513 }
    514 
    515 void Dbg::Disposed() {
    516   gDisposed = true;
    517 }
    518 
    519 bool Dbg::IsDisposed() {
    520   return gDisposed;
    521 }
    522 
    523 void Dbg::GoActive() {
    524   // Enable all debugging features, including scans for breakpoints.
    525   // This is a no-op if we're already active.
    526   // Only called from the JDWP handler thread.
    527   if (gDebuggerActive) {
    528     return;
    529   }
    530 
    531   {
    532     // TODO: dalvik only warned if there were breakpoints left over. clear in Dbg::Disconnected?
    533     MutexLock mu(Thread::Current(), *Locks::breakpoint_lock_);
    534     CHECK_EQ(gBreakpoints.size(), 0U);
    535   }
    536 
    537   Runtime* runtime = Runtime::Current();
    538   runtime->GetThreadList()->SuspendAll();
    539   Thread* self = Thread::Current();
    540   ThreadState old_state = self->SetStateUnsafe(kRunnable);
    541   CHECK_NE(old_state, kRunnable);
    542   runtime->GetInstrumentation()->AddListener(&gDebugInstrumentationListener,
    543                                              instrumentation::Instrumentation::kMethodEntered |
    544                                              instrumentation::Instrumentation::kMethodExited |
    545                                              instrumentation::Instrumentation::kDexPcMoved |
    546                                              instrumentation::Instrumentation::kExceptionCaught);
    547   gDebuggerActive = true;
    548   CHECK_EQ(self->SetStateUnsafe(old_state), kRunnable);
    549   runtime->GetThreadList()->ResumeAll();
    550 
    551   LOG(INFO) << "Debugger is active";
    552 }
    553 
    554 void Dbg::Disconnected() {
    555   CHECK(gDebuggerConnected);
    556 
    557   LOG(INFO) << "Debugger is no longer active";
    558 
    559   // Suspend all threads and exclusively acquire the mutator lock. Set the state of the thread
    560   // to kRunnable to avoid scoped object access transitions. Remove the debugger as a listener
    561   // and clear the object registry.
    562   Runtime* runtime = Runtime::Current();
    563   runtime->GetThreadList()->SuspendAll();
    564   Thread* self = Thread::Current();
    565   ThreadState old_state = self->SetStateUnsafe(kRunnable);
    566   runtime->GetInstrumentation()->RemoveListener(&gDebugInstrumentationListener,
    567                                                 instrumentation::Instrumentation::kMethodEntered |
    568                                                 instrumentation::Instrumentation::kMethodExited |
    569                                                 instrumentation::Instrumentation::kDexPcMoved |
    570                                                 instrumentation::Instrumentation::kExceptionCaught);
    571   gDebuggerActive = false;
    572   gRegistry->Clear();
    573   gDebuggerConnected = false;
    574   CHECK_EQ(self->SetStateUnsafe(old_state), kRunnable);
    575   runtime->GetThreadList()->ResumeAll();
    576 }
    577 
    578 bool Dbg::IsDebuggerActive() {
    579   return gDebuggerActive;
    580 }
    581 
    582 bool Dbg::IsJdwpConfigured() {
    583   return gJdwpConfigured;
    584 }
    585 
    586 int64_t Dbg::LastDebuggerActivity() {
    587   return gJdwpState->LastDebuggerActivity();
    588 }
    589 
    590 void Dbg::UndoDebuggerSuspensions() {
    591   Runtime::Current()->GetThreadList()->UndoDebuggerSuspensions();
    592 }
    593 
    594 std::string Dbg::GetClassName(JDWP::RefTypeId class_id) {
    595   mirror::Object* o = gRegistry->Get<mirror::Object*>(class_id);
    596   if (o == NULL) {
    597     return "NULL";
    598   }
    599   if (o == ObjectRegistry::kInvalidObject) {
    600     return StringPrintf("invalid object %p", reinterpret_cast<void*>(class_id));
    601   }
    602   if (!o->IsClass()) {
    603     return StringPrintf("non-class %p", o);  // This is only used for debugging output anyway.
    604   }
    605   return DescriptorToName(ClassHelper(o->AsClass()).GetDescriptor());
    606 }
    607 
    608 JDWP::JdwpError Dbg::GetClassObject(JDWP::RefTypeId id, JDWP::ObjectId& class_object_id) {
    609   JDWP::JdwpError status;
    610   mirror::Class* c = DecodeClass(id, status);
    611   if (c == NULL) {
    612     return status;
    613   }
    614   class_object_id = gRegistry->Add(c);
    615   return JDWP::ERR_NONE;
    616 }
    617 
    618 JDWP::JdwpError Dbg::GetSuperclass(JDWP::RefTypeId id, JDWP::RefTypeId& superclass_id) {
    619   JDWP::JdwpError status;
    620   mirror::Class* c = DecodeClass(id, status);
    621   if (c == NULL) {
    622     return status;
    623   }
    624   if (c->IsInterface()) {
    625     // http://code.google.com/p/android/issues/detail?id=20856
    626     superclass_id = 0;
    627   } else {
    628     superclass_id = gRegistry->Add(c->GetSuperClass());
    629   }
    630   return JDWP::ERR_NONE;
    631 }
    632 
    633 JDWP::JdwpError Dbg::GetClassLoader(JDWP::RefTypeId id, JDWP::ExpandBuf* pReply) {
    634   mirror::Object* o = gRegistry->Get<mirror::Object*>(id);
    635   if (o == NULL || o == ObjectRegistry::kInvalidObject) {
    636     return JDWP::ERR_INVALID_OBJECT;
    637   }
    638   expandBufAddObjectId(pReply, gRegistry->Add(o->GetClass()->GetClassLoader()));
    639   return JDWP::ERR_NONE;
    640 }
    641 
    642 JDWP::JdwpError Dbg::GetModifiers(JDWP::RefTypeId id, JDWP::ExpandBuf* pReply) {
    643   JDWP::JdwpError status;
    644   mirror::Class* c = DecodeClass(id, status);
    645   if (c == NULL) {
    646     return status;
    647   }
    648 
    649   uint32_t access_flags = c->GetAccessFlags() & kAccJavaFlagsMask;
    650 
    651   // Set ACC_SUPER; dex files don't contain this flag, but all classes are supposed to have it set.
    652   // Class.getModifiers doesn't return it, but JDWP does, so we set it here.
    653   access_flags |= kAccSuper;
    654 
    655   expandBufAdd4BE(pReply, access_flags);
    656 
    657   return JDWP::ERR_NONE;
    658 }
    659 
    660 JDWP::JdwpError Dbg::GetMonitorInfo(JDWP::ObjectId object_id, JDWP::ExpandBuf* reply)
    661     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    662   mirror::Object* o = gRegistry->Get<mirror::Object*>(object_id);
    663   if (o == NULL || o == ObjectRegistry::kInvalidObject) {
    664     return JDWP::ERR_INVALID_OBJECT;
    665   }
    666 
    667   // Ensure all threads are suspended while we read objects' lock words.
    668   Thread* self = Thread::Current();
    669   Locks::mutator_lock_->SharedUnlock(self);
    670   Locks::mutator_lock_->ExclusiveLock(self);
    671 
    672   MonitorInfo monitor_info(o);
    673 
    674   Locks::mutator_lock_->ExclusiveUnlock(self);
    675   Locks::mutator_lock_->SharedLock(self);
    676 
    677   if (monitor_info.owner != NULL) {
    678     expandBufAddObjectId(reply, gRegistry->Add(monitor_info.owner->GetPeer()));
    679   } else {
    680     expandBufAddObjectId(reply, gRegistry->Add(NULL));
    681   }
    682   expandBufAdd4BE(reply, monitor_info.entry_count);
    683   expandBufAdd4BE(reply, monitor_info.waiters.size());
    684   for (size_t i = 0; i < monitor_info.waiters.size(); ++i) {
    685     expandBufAddObjectId(reply, gRegistry->Add(monitor_info.waiters[i]->GetPeer()));
    686   }
    687   return JDWP::ERR_NONE;
    688 }
    689 
    690 JDWP::JdwpError Dbg::GetOwnedMonitors(JDWP::ObjectId thread_id,
    691                                       std::vector<JDWP::ObjectId>& monitors,
    692                                       std::vector<uint32_t>& stack_depths)
    693     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    694   ScopedObjectAccessUnchecked soa(Thread::Current());
    695   MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
    696   Thread* thread;
    697   JDWP::JdwpError error = DecodeThread(soa, thread_id, thread);
    698   if (error != JDWP::ERR_NONE) {
    699     return error;
    700   }
    701   if (!IsSuspendedForDebugger(soa, thread)) {
    702     return JDWP::ERR_THREAD_NOT_SUSPENDED;
    703   }
    704 
    705   struct OwnedMonitorVisitor : public StackVisitor {
    706     OwnedMonitorVisitor(Thread* thread, Context* context)
    707         SHARED_LOCKS_REQUIRED(Locks::mutator_lock_)
    708         : StackVisitor(thread, context), current_stack_depth(0) {}
    709 
    710     // TODO: Enable annotalysis. We know lock is held in constructor, but abstraction confuses
    711     // annotalysis.
    712     bool VisitFrame() NO_THREAD_SAFETY_ANALYSIS {
    713       if (!GetMethod()->IsRuntimeMethod()) {
    714         Monitor::VisitLocks(this, AppendOwnedMonitors, this);
    715         ++current_stack_depth;
    716       }
    717       return true;
    718     }
    719 
    720     static void AppendOwnedMonitors(mirror::Object* owned_monitor, void* arg) {
    721       OwnedMonitorVisitor* visitor = reinterpret_cast<OwnedMonitorVisitor*>(arg);
    722       visitor->monitors.push_back(owned_monitor);
    723       visitor->stack_depths.push_back(visitor->current_stack_depth);
    724     }
    725 
    726     size_t current_stack_depth;
    727     std::vector<mirror::Object*> monitors;
    728     std::vector<uint32_t> stack_depths;
    729   };
    730   UniquePtr<Context> context(Context::Create());
    731   OwnedMonitorVisitor visitor(thread, context.get());
    732   visitor.WalkStack();
    733 
    734   for (size_t i = 0; i < visitor.monitors.size(); ++i) {
    735     monitors.push_back(gRegistry->Add(visitor.monitors[i]));
    736     stack_depths.push_back(visitor.stack_depths[i]);
    737   }
    738 
    739   return JDWP::ERR_NONE;
    740 }
    741 
    742 JDWP::JdwpError Dbg::GetContendedMonitor(JDWP::ObjectId thread_id, JDWP::ObjectId& contended_monitor)
    743     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    744   ScopedObjectAccessUnchecked soa(Thread::Current());
    745   MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
    746   Thread* thread;
    747   JDWP::JdwpError error = DecodeThread(soa, thread_id, thread);
    748   if (error != JDWP::ERR_NONE) {
    749     return error;
    750   }
    751   if (!IsSuspendedForDebugger(soa, thread)) {
    752     return JDWP::ERR_THREAD_NOT_SUSPENDED;
    753   }
    754 
    755   contended_monitor = gRegistry->Add(Monitor::GetContendedMonitor(thread));
    756 
    757   return JDWP::ERR_NONE;
    758 }
    759 
    760 JDWP::JdwpError Dbg::GetInstanceCounts(const std::vector<JDWP::RefTypeId>& class_ids,
    761                                        std::vector<uint64_t>& counts)
    762     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    763   std::vector<mirror::Class*> classes;
    764   counts.clear();
    765   for (size_t i = 0; i < class_ids.size(); ++i) {
    766     JDWP::JdwpError status;
    767     mirror::Class* c = DecodeClass(class_ids[i], status);
    768     if (c == NULL) {
    769       return status;
    770     }
    771     classes.push_back(c);
    772     counts.push_back(0);
    773   }
    774 
    775   Runtime::Current()->GetHeap()->CountInstances(classes, false, &counts[0]);
    776   return JDWP::ERR_NONE;
    777 }
    778 
    779 JDWP::JdwpError Dbg::GetInstances(JDWP::RefTypeId class_id, int32_t max_count, std::vector<JDWP::ObjectId>& instances)
    780     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    781   JDWP::JdwpError status;
    782   mirror::Class* c = DecodeClass(class_id, status);
    783   if (c == NULL) {
    784     return status;
    785   }
    786 
    787   std::vector<mirror::Object*> raw_instances;
    788   Runtime::Current()->GetHeap()->GetInstances(c, max_count, raw_instances);
    789   for (size_t i = 0; i < raw_instances.size(); ++i) {
    790     instances.push_back(gRegistry->Add(raw_instances[i]));
    791   }
    792   return JDWP::ERR_NONE;
    793 }
    794 
    795 JDWP::JdwpError Dbg::GetReferringObjects(JDWP::ObjectId object_id, int32_t max_count,
    796                                          std::vector<JDWP::ObjectId>& referring_objects)
    797     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    798   mirror::Object* o = gRegistry->Get<mirror::Object*>(object_id);
    799   if (o == NULL || o == ObjectRegistry::kInvalidObject) {
    800     return JDWP::ERR_INVALID_OBJECT;
    801   }
    802 
    803   std::vector<mirror::Object*> raw_instances;
    804   Runtime::Current()->GetHeap()->GetReferringObjects(o, max_count, raw_instances);
    805   for (size_t i = 0; i < raw_instances.size(); ++i) {
    806     referring_objects.push_back(gRegistry->Add(raw_instances[i]));
    807   }
    808   return JDWP::ERR_NONE;
    809 }
    810 
    811 JDWP::JdwpError Dbg::DisableCollection(JDWP::ObjectId object_id)
    812     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    813   gRegistry->DisableCollection(object_id);
    814   return JDWP::ERR_NONE;
    815 }
    816 
    817 JDWP::JdwpError Dbg::EnableCollection(JDWP::ObjectId object_id)
    818     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    819   gRegistry->EnableCollection(object_id);
    820   return JDWP::ERR_NONE;
    821 }
    822 
    823 JDWP::JdwpError Dbg::IsCollected(JDWP::ObjectId object_id, bool& is_collected)
    824     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    825   is_collected = gRegistry->IsCollected(object_id);
    826   return JDWP::ERR_NONE;
    827 }
    828 
    829 void Dbg::DisposeObject(JDWP::ObjectId object_id, uint32_t reference_count)
    830     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    831   gRegistry->DisposeObject(object_id, reference_count);
    832 }
    833 
    834 JDWP::JdwpError Dbg::GetReflectedType(JDWP::RefTypeId class_id, JDWP::ExpandBuf* pReply) {
    835   JDWP::JdwpError status;
    836   mirror::Class* c = DecodeClass(class_id, status);
    837   if (c == NULL) {
    838     return status;
    839   }
    840 
    841   expandBufAdd1(pReply, c->IsInterface() ? JDWP::TT_INTERFACE : JDWP::TT_CLASS);
    842   expandBufAddRefTypeId(pReply, class_id);
    843   return JDWP::ERR_NONE;
    844 }
    845 
    846 void Dbg::GetClassList(std::vector<JDWP::RefTypeId>& classes) {
    847   // Get the complete list of reference classes (i.e. all classes except
    848   // the primitive types).
    849   // Returns a newly-allocated buffer full of RefTypeId values.
    850   struct ClassListCreator {
    851     explicit ClassListCreator(std::vector<JDWP::RefTypeId>& classes) : classes(classes) {
    852     }
    853 
    854     static bool Visit(mirror::Class* c, void* arg) {
    855       return reinterpret_cast<ClassListCreator*>(arg)->Visit(c);
    856     }
    857 
    858     // TODO: Enable annotalysis. We know lock is held in constructor, but abstraction confuses
    859     // annotalysis.
    860     bool Visit(mirror::Class* c) NO_THREAD_SAFETY_ANALYSIS {
    861       if (!c->IsPrimitive()) {
    862         classes.push_back(gRegistry->AddRefType(c));
    863       }
    864       return true;
    865     }
    866 
    867     std::vector<JDWP::RefTypeId>& classes;
    868   };
    869 
    870   ClassListCreator clc(classes);
    871   Runtime::Current()->GetClassLinker()->VisitClasses(ClassListCreator::Visit, &clc);
    872 }
    873 
    874 JDWP::JdwpError Dbg::GetClassInfo(JDWP::RefTypeId class_id, JDWP::JdwpTypeTag* pTypeTag, uint32_t* pStatus, std::string* pDescriptor) {
    875   JDWP::JdwpError status;
    876   mirror::Class* c = DecodeClass(class_id, status);
    877   if (c == NULL) {
    878     return status;
    879   }
    880 
    881   if (c->IsArrayClass()) {
    882     *pStatus = JDWP::CS_VERIFIED | JDWP::CS_PREPARED;
    883     *pTypeTag = JDWP::TT_ARRAY;
    884   } else {
    885     if (c->IsErroneous()) {
    886       *pStatus = JDWP::CS_ERROR;
    887     } else {
    888       *pStatus = JDWP::CS_VERIFIED | JDWP::CS_PREPARED | JDWP::CS_INITIALIZED;
    889     }
    890     *pTypeTag = c->IsInterface() ? JDWP::TT_INTERFACE : JDWP::TT_CLASS;
    891   }
    892 
    893   if (pDescriptor != NULL) {
    894     *pDescriptor = ClassHelper(c).GetDescriptor();
    895   }
    896   return JDWP::ERR_NONE;
    897 }
    898 
    899 void Dbg::FindLoadedClassBySignature(const char* descriptor, std::vector<JDWP::RefTypeId>& ids) {
    900   std::vector<mirror::Class*> classes;
    901   Runtime::Current()->GetClassLinker()->LookupClasses(descriptor, classes);
    902   ids.clear();
    903   for (size_t i = 0; i < classes.size(); ++i) {
    904     ids.push_back(gRegistry->Add(classes[i]));
    905   }
    906 }
    907 
    908 JDWP::JdwpError Dbg::GetReferenceType(JDWP::ObjectId object_id, JDWP::ExpandBuf* pReply)
    909     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
    910   mirror::Object* o = gRegistry->Get<mirror::Object*>(object_id);
    911   if (o == NULL || o == ObjectRegistry::kInvalidObject) {
    912     return JDWP::ERR_INVALID_OBJECT;
    913   }
    914 
    915   JDWP::JdwpTypeTag type_tag;
    916   if (o->GetClass()->IsArrayClass()) {
    917     type_tag = JDWP::TT_ARRAY;
    918   } else if (o->GetClass()->IsInterface()) {
    919     type_tag = JDWP::TT_INTERFACE;
    920   } else {
    921     type_tag = JDWP::TT_CLASS;
    922   }
    923   JDWP::RefTypeId type_id = gRegistry->AddRefType(o->GetClass());
    924 
    925   expandBufAdd1(pReply, type_tag);
    926   expandBufAddRefTypeId(pReply, type_id);
    927 
    928   return JDWP::ERR_NONE;
    929 }
    930 
    931 JDWP::JdwpError Dbg::GetSignature(JDWP::RefTypeId class_id, std::string& signature) {
    932   JDWP::JdwpError status;
    933   mirror::Class* c = DecodeClass(class_id, status);
    934   if (c == NULL) {
    935     return status;
    936   }
    937   signature = ClassHelper(c).GetDescriptor();
    938   return JDWP::ERR_NONE;
    939 }
    940 
    941 JDWP::JdwpError Dbg::GetSourceFile(JDWP::RefTypeId class_id, std::string& result) {
    942   JDWP::JdwpError status;
    943   mirror::Class* c = DecodeClass(class_id, status);
    944   if (c == NULL) {
    945     return status;
    946   }
    947   result = ClassHelper(c).GetSourceFile();
    948   return JDWP::ERR_NONE;
    949 }
    950 
    951 JDWP::JdwpError Dbg::GetObjectTag(JDWP::ObjectId object_id, uint8_t& tag) {
    952   mirror::Object* o = gRegistry->Get<mirror::Object*>(object_id);
    953   if (o == ObjectRegistry::kInvalidObject) {
    954     return JDWP::ERR_INVALID_OBJECT;
    955   }
    956   tag = TagFromObject(o);
    957   return JDWP::ERR_NONE;
    958 }
    959 
    960 size_t Dbg::GetTagWidth(JDWP::JdwpTag tag) {
    961   switch (tag) {
    962   case JDWP::JT_VOID:
    963     return 0;
    964   case JDWP::JT_BYTE:
    965   case JDWP::JT_BOOLEAN:
    966     return 1;
    967   case JDWP::JT_CHAR:
    968   case JDWP::JT_SHORT:
    969     return 2;
    970   case JDWP::JT_FLOAT:
    971   case JDWP::JT_INT:
    972     return 4;
    973   case JDWP::JT_ARRAY:
    974   case JDWP::JT_OBJECT:
    975   case JDWP::JT_STRING:
    976   case JDWP::JT_THREAD:
    977   case JDWP::JT_THREAD_GROUP:
    978   case JDWP::JT_CLASS_LOADER:
    979   case JDWP::JT_CLASS_OBJECT:
    980     return sizeof(JDWP::ObjectId);
    981   case JDWP::JT_DOUBLE:
    982   case JDWP::JT_LONG:
    983     return 8;
    984   default:
    985     LOG(FATAL) << "Unknown tag " << tag;
    986     return -1;
    987   }
    988 }
    989 
    990 JDWP::JdwpError Dbg::GetArrayLength(JDWP::ObjectId array_id, int& length) {
    991   JDWP::JdwpError status;
    992   mirror::Array* a = DecodeArray(array_id, status);
    993   if (a == NULL) {
    994     return status;
    995   }
    996   length = a->GetLength();
    997   return JDWP::ERR_NONE;
    998 }
    999 
   1000 JDWP::JdwpError Dbg::OutputArray(JDWP::ObjectId array_id, int offset, int count, JDWP::ExpandBuf* pReply) {
   1001   JDWP::JdwpError status;
   1002   mirror::Array* a = DecodeArray(array_id, status);
   1003   if (a == NULL) {
   1004     return status;
   1005   }
   1006 
   1007   if (offset < 0 || count < 0 || offset > a->GetLength() || a->GetLength() - offset < count) {
   1008     LOG(WARNING) << __FUNCTION__ << " access out of bounds: offset=" << offset << "; count=" << count;
   1009     return JDWP::ERR_INVALID_LENGTH;
   1010   }
   1011   std::string descriptor(ClassHelper(a->GetClass()).GetDescriptor());
   1012   JDWP::JdwpTag tag = BasicTagFromDescriptor(descriptor.c_str() + 1);
   1013 
   1014   expandBufAdd1(pReply, tag);
   1015   expandBufAdd4BE(pReply, count);
   1016 
   1017   if (IsPrimitiveTag(tag)) {
   1018     size_t width = GetTagWidth(tag);
   1019     uint8_t* dst = expandBufAddSpace(pReply, count * width);
   1020     if (width == 8) {
   1021       const uint64_t* src8 = reinterpret_cast<uint64_t*>(a->GetRawData(sizeof(uint64_t)));
   1022       for (int i = 0; i < count; ++i) JDWP::Write8BE(&dst, src8[offset + i]);
   1023     } else if (width == 4) {
   1024       const uint32_t* src4 = reinterpret_cast<uint32_t*>(a->GetRawData(sizeof(uint32_t)));
   1025       for (int i = 0; i < count; ++i) JDWP::Write4BE(&dst, src4[offset + i]);
   1026     } else if (width == 2) {
   1027       const uint16_t* src2 = reinterpret_cast<uint16_t*>(a->GetRawData(sizeof(uint16_t)));
   1028       for (int i = 0; i < count; ++i) JDWP::Write2BE(&dst, src2[offset + i]);
   1029     } else {
   1030       const uint8_t* src = reinterpret_cast<uint8_t*>(a->GetRawData(sizeof(uint8_t)));
   1031       memcpy(dst, &src[offset * width], count * width);
   1032     }
   1033   } else {
   1034     mirror::ObjectArray<mirror::Object>* oa = a->AsObjectArray<mirror::Object>();
   1035     for (int i = 0; i < count; ++i) {
   1036       mirror::Object* element = oa->Get(offset + i);
   1037       JDWP::JdwpTag specific_tag = (element != NULL) ? TagFromObject(element) : tag;
   1038       expandBufAdd1(pReply, specific_tag);
   1039       expandBufAddObjectId(pReply, gRegistry->Add(element));
   1040     }
   1041   }
   1042 
   1043   return JDWP::ERR_NONE;
   1044 }
   1045 
   1046 template <typename T> void CopyArrayData(mirror::Array* a, JDWP::Request& src, int offset, int count) {
   1047   DCHECK(a->GetClass()->IsPrimitiveArray());
   1048 
   1049   T* dst = &(reinterpret_cast<T*>(a->GetRawData(sizeof(T)))[offset * sizeof(T)]);
   1050   for (int i = 0; i < count; ++i) {
   1051     *dst++ = src.ReadValue(sizeof(T));
   1052   }
   1053 }
   1054 
   1055 JDWP::JdwpError Dbg::SetArrayElements(JDWP::ObjectId array_id, int offset, int count,
   1056                                       JDWP::Request& request)
   1057     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1058   JDWP::JdwpError status;
   1059   mirror::Array* dst = DecodeArray(array_id, status);
   1060   if (dst == NULL) {
   1061     return status;
   1062   }
   1063 
   1064   if (offset < 0 || count < 0 || offset > dst->GetLength() || dst->GetLength() - offset < count) {
   1065     LOG(WARNING) << __FUNCTION__ << " access out of bounds: offset=" << offset << "; count=" << count;
   1066     return JDWP::ERR_INVALID_LENGTH;
   1067   }
   1068   std::string descriptor(ClassHelper(dst->GetClass()).GetDescriptor());
   1069   JDWP::JdwpTag tag = BasicTagFromDescriptor(descriptor.c_str() + 1);
   1070 
   1071   if (IsPrimitiveTag(tag)) {
   1072     size_t width = GetTagWidth(tag);
   1073     if (width == 8) {
   1074       CopyArrayData<uint64_t>(dst, request, offset, count);
   1075     } else if (width == 4) {
   1076       CopyArrayData<uint32_t>(dst, request, offset, count);
   1077     } else if (width == 2) {
   1078       CopyArrayData<uint16_t>(dst, request, offset, count);
   1079     } else {
   1080       CopyArrayData<uint8_t>(dst, request, offset, count);
   1081     }
   1082   } else {
   1083     mirror::ObjectArray<mirror::Object>* oa = dst->AsObjectArray<mirror::Object>();
   1084     for (int i = 0; i < count; ++i) {
   1085       JDWP::ObjectId id = request.ReadObjectId();
   1086       mirror::Object* o = gRegistry->Get<mirror::Object*>(id);
   1087       if (o == ObjectRegistry::kInvalidObject) {
   1088         return JDWP::ERR_INVALID_OBJECT;
   1089       }
   1090       oa->Set(offset + i, o);
   1091     }
   1092   }
   1093 
   1094   return JDWP::ERR_NONE;
   1095 }
   1096 
   1097 JDWP::ObjectId Dbg::CreateString(const std::string& str) {
   1098   return gRegistry->Add(mirror::String::AllocFromModifiedUtf8(Thread::Current(), str.c_str()));
   1099 }
   1100 
   1101 JDWP::JdwpError Dbg::CreateObject(JDWP::RefTypeId class_id, JDWP::ObjectId& new_object) {
   1102   JDWP::JdwpError status;
   1103   mirror::Class* c = DecodeClass(class_id, status);
   1104   if (c == NULL) {
   1105     return status;
   1106   }
   1107   new_object = gRegistry->Add(c->AllocObject(Thread::Current()));
   1108   return JDWP::ERR_NONE;
   1109 }
   1110 
   1111 /*
   1112  * Used by Eclipse's "Display" view to evaluate "new byte[5]" to get "(byte[]) [0, 0, 0, 0, 0]".
   1113  */
   1114 JDWP::JdwpError Dbg::CreateArrayObject(JDWP::RefTypeId array_class_id, uint32_t length,
   1115                                        JDWP::ObjectId& new_array) {
   1116   JDWP::JdwpError status;
   1117   mirror::Class* c = DecodeClass(array_class_id, status);
   1118   if (c == NULL) {
   1119     return status;
   1120   }
   1121   new_array = gRegistry->Add(mirror::Array::Alloc(Thread::Current(), c, length));
   1122   return JDWP::ERR_NONE;
   1123 }
   1124 
   1125 bool Dbg::MatchType(JDWP::RefTypeId instance_class_id, JDWP::RefTypeId class_id) {
   1126   JDWP::JdwpError status;
   1127   mirror::Class* c1 = DecodeClass(instance_class_id, status);
   1128   CHECK(c1 != NULL);
   1129   mirror::Class* c2 = DecodeClass(class_id, status);
   1130   CHECK(c2 != NULL);
   1131   return c1->IsAssignableFrom(c2);
   1132 }
   1133 
   1134 static JDWP::FieldId ToFieldId(const mirror::ArtField* f)
   1135     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1136 #ifdef MOVING_GARBAGE_COLLECTOR
   1137   UNIMPLEMENTED(FATAL);
   1138 #else
   1139   return static_cast<JDWP::FieldId>(reinterpret_cast<uintptr_t>(f));
   1140 #endif
   1141 }
   1142 
   1143 static JDWP::MethodId ToMethodId(const mirror::ArtMethod* m)
   1144     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1145 #ifdef MOVING_GARBAGE_COLLECTOR
   1146   UNIMPLEMENTED(FATAL);
   1147 #else
   1148   return static_cast<JDWP::MethodId>(reinterpret_cast<uintptr_t>(m));
   1149 #endif
   1150 }
   1151 
   1152 static mirror::ArtField* FromFieldId(JDWP::FieldId fid)
   1153     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1154 #ifdef MOVING_GARBAGE_COLLECTOR
   1155   UNIMPLEMENTED(FATAL);
   1156 #else
   1157   return reinterpret_cast<mirror::ArtField*>(static_cast<uintptr_t>(fid));
   1158 #endif
   1159 }
   1160 
   1161 static mirror::ArtMethod* FromMethodId(JDWP::MethodId mid)
   1162     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1163 #ifdef MOVING_GARBAGE_COLLECTOR
   1164   UNIMPLEMENTED(FATAL);
   1165 #else
   1166   return reinterpret_cast<mirror::ArtMethod*>(static_cast<uintptr_t>(mid));
   1167 #endif
   1168 }
   1169 
   1170 static void SetLocation(JDWP::JdwpLocation& location, mirror::ArtMethod* m, uint32_t dex_pc)
   1171     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1172   if (m == NULL) {
   1173     memset(&location, 0, sizeof(location));
   1174   } else {
   1175     mirror::Class* c = m->GetDeclaringClass();
   1176     location.type_tag = c->IsInterface() ? JDWP::TT_INTERFACE : JDWP::TT_CLASS;
   1177     location.class_id = gRegistry->Add(c);
   1178     location.method_id = ToMethodId(m);
   1179     location.dex_pc = dex_pc;
   1180   }
   1181 }
   1182 
   1183 std::string Dbg::GetMethodName(JDWP::MethodId method_id)
   1184     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1185   mirror::ArtMethod* m = FromMethodId(method_id);
   1186   return MethodHelper(m).GetName();
   1187 }
   1188 
   1189 std::string Dbg::GetFieldName(JDWP::FieldId field_id)
   1190     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1191   mirror::ArtField* f = FromFieldId(field_id);
   1192   return FieldHelper(f).GetName();
   1193 }
   1194 
   1195 /*
   1196  * Augment the access flags for synthetic methods and fields by setting
   1197  * the (as described by the spec) "0xf0000000 bit".  Also, strip out any
   1198  * flags not specified by the Java programming language.
   1199  */
   1200 static uint32_t MangleAccessFlags(uint32_t accessFlags) {
   1201   accessFlags &= kAccJavaFlagsMask;
   1202   if ((accessFlags & kAccSynthetic) != 0) {
   1203     accessFlags |= 0xf0000000;
   1204   }
   1205   return accessFlags;
   1206 }
   1207 
   1208 static const uint16_t kEclipseWorkaroundSlot = 1000;
   1209 
   1210 /*
   1211  * Eclipse appears to expect that the "this" reference is in slot zero.
   1212  * If it's not, the "variables" display will show two copies of "this",
   1213  * possibly because it gets "this" from SF.ThisObject and then displays
   1214  * all locals with nonzero slot numbers.
   1215  *
   1216  * So, we remap the item in slot 0 to 1000, and remap "this" to zero.  On
   1217  * SF.GetValues / SF.SetValues we map them back.
   1218  *
   1219  * TODO: jdb uses the value to determine whether a variable is a local or an argument,
   1220  * by checking whether it's less than the number of arguments. To make that work, we'd
   1221  * have to "mangle" all the arguments to come first, not just the implicit argument 'this'.
   1222  */
   1223 static uint16_t MangleSlot(uint16_t slot, const char* name) {
   1224   uint16_t newSlot = slot;
   1225   if (strcmp(name, "this") == 0) {
   1226     newSlot = 0;
   1227   } else if (slot == 0) {
   1228     newSlot = kEclipseWorkaroundSlot;
   1229   }
   1230   return newSlot;
   1231 }
   1232 
   1233 static uint16_t DemangleSlot(uint16_t slot, mirror::ArtMethod* m)
   1234     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1235   if (slot == kEclipseWorkaroundSlot) {
   1236     return 0;
   1237   } else if (slot == 0) {
   1238     const DexFile::CodeItem* code_item = MethodHelper(m).GetCodeItem();
   1239     CHECK(code_item != NULL) << PrettyMethod(m);
   1240     return code_item->registers_size_ - code_item->ins_size_;
   1241   }
   1242   return slot;
   1243 }
   1244 
   1245 JDWP::JdwpError Dbg::OutputDeclaredFields(JDWP::RefTypeId class_id, bool with_generic, JDWP::ExpandBuf* pReply) {
   1246   JDWP::JdwpError status;
   1247   mirror::Class* c = DecodeClass(class_id, status);
   1248   if (c == NULL) {
   1249     return status;
   1250   }
   1251 
   1252   size_t instance_field_count = c->NumInstanceFields();
   1253   size_t static_field_count = c->NumStaticFields();
   1254 
   1255   expandBufAdd4BE(pReply, instance_field_count + static_field_count);
   1256 
   1257   for (size_t i = 0; i < instance_field_count + static_field_count; ++i) {
   1258     mirror::ArtField* f = (i < instance_field_count) ? c->GetInstanceField(i) : c->GetStaticField(i - instance_field_count);
   1259     FieldHelper fh(f);
   1260     expandBufAddFieldId(pReply, ToFieldId(f));
   1261     expandBufAddUtf8String(pReply, fh.GetName());
   1262     expandBufAddUtf8String(pReply, fh.GetTypeDescriptor());
   1263     if (with_generic) {
   1264       static const char genericSignature[1] = "";
   1265       expandBufAddUtf8String(pReply, genericSignature);
   1266     }
   1267     expandBufAdd4BE(pReply, MangleAccessFlags(f->GetAccessFlags()));
   1268   }
   1269   return JDWP::ERR_NONE;
   1270 }
   1271 
   1272 JDWP::JdwpError Dbg::OutputDeclaredMethods(JDWP::RefTypeId class_id, bool with_generic,
   1273                                            JDWP::ExpandBuf* pReply) {
   1274   JDWP::JdwpError status;
   1275   mirror::Class* c = DecodeClass(class_id, status);
   1276   if (c == NULL) {
   1277     return status;
   1278   }
   1279 
   1280   size_t direct_method_count = c->NumDirectMethods();
   1281   size_t virtual_method_count = c->NumVirtualMethods();
   1282 
   1283   expandBufAdd4BE(pReply, direct_method_count + virtual_method_count);
   1284 
   1285   for (size_t i = 0; i < direct_method_count + virtual_method_count; ++i) {
   1286     mirror::ArtMethod* m = (i < direct_method_count) ? c->GetDirectMethod(i) : c->GetVirtualMethod(i - direct_method_count);
   1287     MethodHelper mh(m);
   1288     expandBufAddMethodId(pReply, ToMethodId(m));
   1289     expandBufAddUtf8String(pReply, mh.GetName());
   1290     expandBufAddUtf8String(pReply, mh.GetSignature());
   1291     if (with_generic) {
   1292       static const char genericSignature[1] = "";
   1293       expandBufAddUtf8String(pReply, genericSignature);
   1294     }
   1295     expandBufAdd4BE(pReply, MangleAccessFlags(m->GetAccessFlags()));
   1296   }
   1297   return JDWP::ERR_NONE;
   1298 }
   1299 
   1300 JDWP::JdwpError Dbg::OutputDeclaredInterfaces(JDWP::RefTypeId class_id, JDWP::ExpandBuf* pReply) {
   1301   JDWP::JdwpError status;
   1302   mirror::Class* c = DecodeClass(class_id, status);
   1303   if (c == NULL) {
   1304     return status;
   1305   }
   1306 
   1307   ClassHelper kh(c);
   1308   size_t interface_count = kh.NumDirectInterfaces();
   1309   expandBufAdd4BE(pReply, interface_count);
   1310   for (size_t i = 0; i < interface_count; ++i) {
   1311     expandBufAddRefTypeId(pReply, gRegistry->AddRefType(kh.GetDirectInterface(i)));
   1312   }
   1313   return JDWP::ERR_NONE;
   1314 }
   1315 
   1316 void Dbg::OutputLineTable(JDWP::RefTypeId, JDWP::MethodId method_id, JDWP::ExpandBuf* pReply)
   1317     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1318   struct DebugCallbackContext {
   1319     int numItems;
   1320     JDWP::ExpandBuf* pReply;
   1321 
   1322     static bool Callback(void* context, uint32_t address, uint32_t line_number) {
   1323       DebugCallbackContext* pContext = reinterpret_cast<DebugCallbackContext*>(context);
   1324       expandBufAdd8BE(pContext->pReply, address);
   1325       expandBufAdd4BE(pContext->pReply, line_number);
   1326       pContext->numItems++;
   1327       return false;
   1328     }
   1329   };
   1330   mirror::ArtMethod* m = FromMethodId(method_id);
   1331   MethodHelper mh(m);
   1332   uint64_t start, end;
   1333   if (m->IsNative()) {
   1334     start = -1;
   1335     end = -1;
   1336   } else {
   1337     start = 0;
   1338     // Return the index of the last instruction
   1339     end = mh.GetCodeItem()->insns_size_in_code_units_ - 1;
   1340   }
   1341 
   1342   expandBufAdd8BE(pReply, start);
   1343   expandBufAdd8BE(pReply, end);
   1344 
   1345   // Add numLines later
   1346   size_t numLinesOffset = expandBufGetLength(pReply);
   1347   expandBufAdd4BE(pReply, 0);
   1348 
   1349   DebugCallbackContext context;
   1350   context.numItems = 0;
   1351   context.pReply = pReply;
   1352 
   1353   mh.GetDexFile().DecodeDebugInfo(mh.GetCodeItem(), m->IsStatic(), m->GetDexMethodIndex(),
   1354                                   DebugCallbackContext::Callback, NULL, &context);
   1355 
   1356   JDWP::Set4BE(expandBufGetBuffer(pReply) + numLinesOffset, context.numItems);
   1357 }
   1358 
   1359 void Dbg::OutputVariableTable(JDWP::RefTypeId, JDWP::MethodId method_id, bool with_generic, JDWP::ExpandBuf* pReply) {
   1360   struct DebugCallbackContext {
   1361     JDWP::ExpandBuf* pReply;
   1362     size_t variable_count;
   1363     bool with_generic;
   1364 
   1365     static void Callback(void* context, uint16_t slot, uint32_t startAddress, uint32_t endAddress, const char* name, const char* descriptor, const char* signature) {
   1366       DebugCallbackContext* pContext = reinterpret_cast<DebugCallbackContext*>(context);
   1367 
   1368       VLOG(jdwp) << StringPrintf("    %2zd: %d(%d) '%s' '%s' '%s' actual slot=%d mangled slot=%d", pContext->variable_count, startAddress, endAddress - startAddress, name, descriptor, signature, slot, MangleSlot(slot, name));
   1369 
   1370       slot = MangleSlot(slot, name);
   1371 
   1372       expandBufAdd8BE(pContext->pReply, startAddress);
   1373       expandBufAddUtf8String(pContext->pReply, name);
   1374       expandBufAddUtf8String(pContext->pReply, descriptor);
   1375       if (pContext->with_generic) {
   1376         expandBufAddUtf8String(pContext->pReply, signature);
   1377       }
   1378       expandBufAdd4BE(pContext->pReply, endAddress - startAddress);
   1379       expandBufAdd4BE(pContext->pReply, slot);
   1380 
   1381       ++pContext->variable_count;
   1382     }
   1383   };
   1384   mirror::ArtMethod* m = FromMethodId(method_id);
   1385   MethodHelper mh(m);
   1386   const DexFile::CodeItem* code_item = mh.GetCodeItem();
   1387 
   1388   // arg_count considers doubles and longs to take 2 units.
   1389   // variable_count considers everything to take 1 unit.
   1390   std::string shorty(mh.GetShorty());
   1391   expandBufAdd4BE(pReply, mirror::ArtMethod::NumArgRegisters(shorty));
   1392 
   1393   // We don't know the total number of variables yet, so leave a blank and update it later.
   1394   size_t variable_count_offset = expandBufGetLength(pReply);
   1395   expandBufAdd4BE(pReply, 0);
   1396 
   1397   DebugCallbackContext context;
   1398   context.pReply = pReply;
   1399   context.variable_count = 0;
   1400   context.with_generic = with_generic;
   1401 
   1402   mh.GetDexFile().DecodeDebugInfo(code_item, m->IsStatic(), m->GetDexMethodIndex(), NULL,
   1403                                   DebugCallbackContext::Callback, &context);
   1404 
   1405   JDWP::Set4BE(expandBufGetBuffer(pReply) + variable_count_offset, context.variable_count);
   1406 }
   1407 
   1408 JDWP::JdwpError Dbg::GetBytecodes(JDWP::RefTypeId, JDWP::MethodId method_id,
   1409                                   std::vector<uint8_t>& bytecodes)
   1410     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1411   mirror::ArtMethod* m = FromMethodId(method_id);
   1412   if (m == NULL) {
   1413     return JDWP::ERR_INVALID_METHODID;
   1414   }
   1415   MethodHelper mh(m);
   1416   const DexFile::CodeItem* code_item = mh.GetCodeItem();
   1417   size_t byte_count = code_item->insns_size_in_code_units_ * 2;
   1418   const uint8_t* begin = reinterpret_cast<const uint8_t*>(code_item->insns_);
   1419   const uint8_t* end = begin + byte_count;
   1420   for (const uint8_t* p = begin; p != end; ++p) {
   1421     bytecodes.push_back(*p);
   1422   }
   1423   return JDWP::ERR_NONE;
   1424 }
   1425 
   1426 JDWP::JdwpTag Dbg::GetFieldBasicTag(JDWP::FieldId field_id) {
   1427   return BasicTagFromDescriptor(FieldHelper(FromFieldId(field_id)).GetTypeDescriptor());
   1428 }
   1429 
   1430 JDWP::JdwpTag Dbg::GetStaticFieldBasicTag(JDWP::FieldId field_id) {
   1431   return BasicTagFromDescriptor(FieldHelper(FromFieldId(field_id)).GetTypeDescriptor());
   1432 }
   1433 
   1434 static JDWP::JdwpError GetFieldValueImpl(JDWP::RefTypeId ref_type_id, JDWP::ObjectId object_id,
   1435                                          JDWP::FieldId field_id, JDWP::ExpandBuf* pReply,
   1436                                          bool is_static)
   1437     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1438   JDWP::JdwpError status;
   1439   mirror::Class* c = DecodeClass(ref_type_id, status);
   1440   if (ref_type_id != 0 && c == NULL) {
   1441     return status;
   1442   }
   1443 
   1444   mirror::Object* o = gRegistry->Get<mirror::Object*>(object_id);
   1445   if ((!is_static && o == NULL) || o == ObjectRegistry::kInvalidObject) {
   1446     return JDWP::ERR_INVALID_OBJECT;
   1447   }
   1448   mirror::ArtField* f = FromFieldId(field_id);
   1449 
   1450   mirror::Class* receiver_class = c;
   1451   if (receiver_class == NULL && o != NULL) {
   1452     receiver_class = o->GetClass();
   1453   }
   1454   // TODO: should we give up now if receiver_class is NULL?
   1455   if (receiver_class != NULL && !f->GetDeclaringClass()->IsAssignableFrom(receiver_class)) {
   1456     LOG(INFO) << "ERR_INVALID_FIELDID: " << PrettyField(f) << " " << PrettyClass(receiver_class);
   1457     return JDWP::ERR_INVALID_FIELDID;
   1458   }
   1459 
   1460   // The RI only enforces the static/non-static mismatch in one direction.
   1461   // TODO: should we change the tests and check both?
   1462   if (is_static) {
   1463     if (!f->IsStatic()) {
   1464       return JDWP::ERR_INVALID_FIELDID;
   1465     }
   1466   } else {
   1467     if (f->IsStatic()) {
   1468       LOG(WARNING) << "Ignoring non-NULL receiver for ObjectReference.SetValues on static field " << PrettyField(f);
   1469     }
   1470   }
   1471   if (f->IsStatic()) {
   1472     o = f->GetDeclaringClass();
   1473   }
   1474 
   1475   JDWP::JdwpTag tag = BasicTagFromDescriptor(FieldHelper(f).GetTypeDescriptor());
   1476 
   1477   if (IsPrimitiveTag(tag)) {
   1478     expandBufAdd1(pReply, tag);
   1479     if (tag == JDWP::JT_BOOLEAN || tag == JDWP::JT_BYTE) {
   1480       expandBufAdd1(pReply, f->Get32(o));
   1481     } else if (tag == JDWP::JT_CHAR || tag == JDWP::JT_SHORT) {
   1482       expandBufAdd2BE(pReply, f->Get32(o));
   1483     } else if (tag == JDWP::JT_FLOAT || tag == JDWP::JT_INT) {
   1484       expandBufAdd4BE(pReply, f->Get32(o));
   1485     } else if (tag == JDWP::JT_DOUBLE || tag == JDWP::JT_LONG) {
   1486       expandBufAdd8BE(pReply, f->Get64(o));
   1487     } else {
   1488       LOG(FATAL) << "Unknown tag: " << tag;
   1489     }
   1490   } else {
   1491     mirror::Object* value = f->GetObject(o);
   1492     expandBufAdd1(pReply, TagFromObject(value));
   1493     expandBufAddObjectId(pReply, gRegistry->Add(value));
   1494   }
   1495   return JDWP::ERR_NONE;
   1496 }
   1497 
   1498 JDWP::JdwpError Dbg::GetFieldValue(JDWP::ObjectId object_id, JDWP::FieldId field_id,
   1499                                    JDWP::ExpandBuf* pReply) {
   1500   return GetFieldValueImpl(0, object_id, field_id, pReply, false);
   1501 }
   1502 
   1503 JDWP::JdwpError Dbg::GetStaticFieldValue(JDWP::RefTypeId ref_type_id, JDWP::FieldId field_id, JDWP::ExpandBuf* pReply) {
   1504   return GetFieldValueImpl(ref_type_id, 0, field_id, pReply, true);
   1505 }
   1506 
   1507 static JDWP::JdwpError SetFieldValueImpl(JDWP::ObjectId object_id, JDWP::FieldId field_id,
   1508                                          uint64_t value, int width, bool is_static)
   1509     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1510   mirror::Object* o = gRegistry->Get<mirror::Object*>(object_id);
   1511   if ((!is_static && o == NULL) || o == ObjectRegistry::kInvalidObject) {
   1512     return JDWP::ERR_INVALID_OBJECT;
   1513   }
   1514   mirror::ArtField* f = FromFieldId(field_id);
   1515 
   1516   // The RI only enforces the static/non-static mismatch in one direction.
   1517   // TODO: should we change the tests and check both?
   1518   if (is_static) {
   1519     if (!f->IsStatic()) {
   1520       return JDWP::ERR_INVALID_FIELDID;
   1521     }
   1522   } else {
   1523     if (f->IsStatic()) {
   1524       LOG(WARNING) << "Ignoring non-NULL receiver for ObjectReference.SetValues on static field " << PrettyField(f);
   1525     }
   1526   }
   1527   if (f->IsStatic()) {
   1528     o = f->GetDeclaringClass();
   1529   }
   1530 
   1531   JDWP::JdwpTag tag = BasicTagFromDescriptor(FieldHelper(f).GetTypeDescriptor());
   1532 
   1533   if (IsPrimitiveTag(tag)) {
   1534     if (tag == JDWP::JT_DOUBLE || tag == JDWP::JT_LONG) {
   1535       CHECK_EQ(width, 8);
   1536       f->Set64(o, value);
   1537     } else {
   1538       CHECK_LE(width, 4);
   1539       f->Set32(o, value);
   1540     }
   1541   } else {
   1542     mirror::Object* v = gRegistry->Get<mirror::Object*>(value);
   1543     if (v == ObjectRegistry::kInvalidObject) {
   1544       return JDWP::ERR_INVALID_OBJECT;
   1545     }
   1546     if (v != NULL) {
   1547       mirror::Class* field_type = FieldHelper(f).GetType();
   1548       if (!field_type->IsAssignableFrom(v->GetClass())) {
   1549         return JDWP::ERR_INVALID_OBJECT;
   1550       }
   1551     }
   1552     f->SetObject(o, v);
   1553   }
   1554 
   1555   return JDWP::ERR_NONE;
   1556 }
   1557 
   1558 JDWP::JdwpError Dbg::SetFieldValue(JDWP::ObjectId object_id, JDWP::FieldId field_id, uint64_t value,
   1559                                    int width) {
   1560   return SetFieldValueImpl(object_id, field_id, value, width, false);
   1561 }
   1562 
   1563 JDWP::JdwpError Dbg::SetStaticFieldValue(JDWP::FieldId field_id, uint64_t value, int width) {
   1564   return SetFieldValueImpl(0, field_id, value, width, true);
   1565 }
   1566 
   1567 std::string Dbg::StringToUtf8(JDWP::ObjectId string_id) {
   1568   mirror::String* s = gRegistry->Get<mirror::String*>(string_id);
   1569   return s->ToModifiedUtf8();
   1570 }
   1571 
   1572 JDWP::JdwpError Dbg::GetThreadName(JDWP::ObjectId thread_id, std::string& name) {
   1573   ScopedObjectAccessUnchecked soa(Thread::Current());
   1574   MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
   1575   Thread* thread;
   1576   JDWP::JdwpError error = DecodeThread(soa, thread_id, thread);
   1577   if (error != JDWP::ERR_NONE && error != JDWP::ERR_THREAD_NOT_ALIVE) {
   1578     return error;
   1579   }
   1580 
   1581   // We still need to report the zombie threads' names, so we can't just call Thread::GetThreadName.
   1582   mirror::Object* thread_object = gRegistry->Get<mirror::Object*>(thread_id);
   1583   mirror::ArtField* java_lang_Thread_name_field =
   1584       soa.DecodeField(WellKnownClasses::java_lang_Thread_name);
   1585   mirror::String* s =
   1586       reinterpret_cast<mirror::String*>(java_lang_Thread_name_field->GetObject(thread_object));
   1587   if (s != NULL) {
   1588     name = s->ToModifiedUtf8();
   1589   }
   1590   return JDWP::ERR_NONE;
   1591 }
   1592 
   1593 JDWP::JdwpError Dbg::GetThreadGroup(JDWP::ObjectId thread_id, JDWP::ExpandBuf* pReply) {
   1594   ScopedObjectAccess soa(Thread::Current());
   1595   mirror::Object* thread_object = gRegistry->Get<mirror::Object*>(thread_id);
   1596   if (thread_object == ObjectRegistry::kInvalidObject) {
   1597     return JDWP::ERR_INVALID_OBJECT;
   1598   }
   1599 
   1600   // Okay, so it's an object, but is it actually a thread?
   1601   MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
   1602   Thread* thread;
   1603   JDWP::JdwpError error = DecodeThread(soa, thread_id, thread);
   1604   if (error == JDWP::ERR_THREAD_NOT_ALIVE) {
   1605     // Zombie threads are in the null group.
   1606     expandBufAddObjectId(pReply, JDWP::ObjectId(0));
   1607     return JDWP::ERR_NONE;
   1608   }
   1609   if (error != JDWP::ERR_NONE) {
   1610     return error;
   1611   }
   1612 
   1613   mirror::Class* c = Runtime::Current()->GetClassLinker()->FindSystemClass("Ljava/lang/Thread;");
   1614   CHECK(c != NULL);
   1615   mirror::ArtField* f = c->FindInstanceField("group", "Ljava/lang/ThreadGroup;");
   1616   CHECK(f != NULL);
   1617   mirror::Object* group = f->GetObject(thread_object);
   1618   CHECK(group != NULL);
   1619   JDWP::ObjectId thread_group_id = gRegistry->Add(group);
   1620 
   1621   expandBufAddObjectId(pReply, thread_group_id);
   1622   return JDWP::ERR_NONE;
   1623 }
   1624 
   1625 std::string Dbg::GetThreadGroupName(JDWP::ObjectId thread_group_id) {
   1626   ScopedObjectAccess soa(Thread::Current());
   1627   mirror::Object* thread_group = gRegistry->Get<mirror::Object*>(thread_group_id);
   1628   CHECK(thread_group != NULL);
   1629 
   1630   mirror::Class* c = Runtime::Current()->GetClassLinker()->FindSystemClass("Ljava/lang/ThreadGroup;");
   1631   CHECK(c != NULL);
   1632   mirror::ArtField* f = c->FindInstanceField("name", "Ljava/lang/String;");
   1633   CHECK(f != NULL);
   1634   mirror::String* s = reinterpret_cast<mirror::String*>(f->GetObject(thread_group));
   1635   return s->ToModifiedUtf8();
   1636 }
   1637 
   1638 JDWP::ObjectId Dbg::GetThreadGroupParent(JDWP::ObjectId thread_group_id) {
   1639   mirror::Object* thread_group = gRegistry->Get<mirror::Object*>(thread_group_id);
   1640   CHECK(thread_group != NULL);
   1641 
   1642   mirror::Class* c = Runtime::Current()->GetClassLinker()->FindSystemClass("Ljava/lang/ThreadGroup;");
   1643   CHECK(c != NULL);
   1644   mirror::ArtField* f = c->FindInstanceField("parent", "Ljava/lang/ThreadGroup;");
   1645   CHECK(f != NULL);
   1646   mirror::Object* parent = f->GetObject(thread_group);
   1647   return gRegistry->Add(parent);
   1648 }
   1649 
   1650 JDWP::ObjectId Dbg::GetSystemThreadGroupId() {
   1651   ScopedObjectAccessUnchecked soa(Thread::Current());
   1652   mirror::ArtField* f = soa.DecodeField(WellKnownClasses::java_lang_ThreadGroup_systemThreadGroup);
   1653   mirror::Object* group = f->GetObject(f->GetDeclaringClass());
   1654   return gRegistry->Add(group);
   1655 }
   1656 
   1657 JDWP::ObjectId Dbg::GetMainThreadGroupId() {
   1658   ScopedObjectAccess soa(Thread::Current());
   1659   mirror::ArtField* f = soa.DecodeField(WellKnownClasses::java_lang_ThreadGroup_mainThreadGroup);
   1660   mirror::Object* group = f->GetObject(f->GetDeclaringClass());
   1661   return gRegistry->Add(group);
   1662 }
   1663 
   1664 JDWP::JdwpThreadStatus Dbg::ToJdwpThreadStatus(ThreadState state) {
   1665   switch (state) {
   1666     case kBlocked:
   1667       return JDWP::TS_MONITOR;
   1668     case kNative:
   1669     case kRunnable:
   1670     case kSuspended:
   1671       return JDWP::TS_RUNNING;
   1672     case kSleeping:
   1673       return JDWP::TS_SLEEPING;
   1674     case kStarting:
   1675     case kTerminated:
   1676       return JDWP::TS_ZOMBIE;
   1677     case kTimedWaiting:
   1678     case kWaitingForDebuggerSend:
   1679     case kWaitingForDebuggerSuspension:
   1680     case kWaitingForDebuggerToAttach:
   1681     case kWaitingForGcToComplete:
   1682     case kWaitingForCheckPointsToRun:
   1683     case kWaitingForJniOnLoad:
   1684     case kWaitingForSignalCatcherOutput:
   1685     case kWaitingInMainDebuggerLoop:
   1686     case kWaitingInMainSignalCatcherLoop:
   1687     case kWaitingPerformingGc:
   1688     case kWaiting:
   1689       return JDWP::TS_WAIT;
   1690       // Don't add a 'default' here so the compiler can spot incompatible enum changes.
   1691   }
   1692   LOG(FATAL) << "Unknown thread state: " << state;
   1693   return JDWP::TS_ZOMBIE;
   1694 }
   1695 
   1696 JDWP::JdwpError Dbg::GetThreadStatus(JDWP::ObjectId thread_id, JDWP::JdwpThreadStatus* pThreadStatus, JDWP::JdwpSuspendStatus* pSuspendStatus) {
   1697   ScopedObjectAccess soa(Thread::Current());
   1698 
   1699   *pSuspendStatus = JDWP::SUSPEND_STATUS_NOT_SUSPENDED;
   1700 
   1701   MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
   1702   Thread* thread;
   1703   JDWP::JdwpError error = DecodeThread(soa, thread_id, thread);
   1704   if (error != JDWP::ERR_NONE) {
   1705     if (error == JDWP::ERR_THREAD_NOT_ALIVE) {
   1706       *pThreadStatus = JDWP::TS_ZOMBIE;
   1707       return JDWP::ERR_NONE;
   1708     }
   1709     return error;
   1710   }
   1711 
   1712   if (IsSuspendedForDebugger(soa, thread)) {
   1713     *pSuspendStatus = JDWP::SUSPEND_STATUS_SUSPENDED;
   1714   }
   1715 
   1716   *pThreadStatus = ToJdwpThreadStatus(thread->GetState());
   1717   return JDWP::ERR_NONE;
   1718 }
   1719 
   1720 JDWP::JdwpError Dbg::GetThreadDebugSuspendCount(JDWP::ObjectId thread_id, JDWP::ExpandBuf* pReply) {
   1721   ScopedObjectAccess soa(Thread::Current());
   1722   MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
   1723   Thread* thread;
   1724   JDWP::JdwpError error = DecodeThread(soa, thread_id, thread);
   1725   if (error != JDWP::ERR_NONE) {
   1726     return error;
   1727   }
   1728   MutexLock mu2(soa.Self(), *Locks::thread_suspend_count_lock_);
   1729   expandBufAdd4BE(pReply, thread->GetDebugSuspendCount());
   1730   return JDWP::ERR_NONE;
   1731 }
   1732 
   1733 JDWP::JdwpError Dbg::Interrupt(JDWP::ObjectId thread_id) {
   1734   ScopedObjectAccess soa(Thread::Current());
   1735   MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
   1736   Thread* thread;
   1737   JDWP::JdwpError error = DecodeThread(soa, thread_id, thread);
   1738   if (error != JDWP::ERR_NONE) {
   1739     return error;
   1740   }
   1741   thread->Interrupt();
   1742   return JDWP::ERR_NONE;
   1743 }
   1744 
   1745 void Dbg::GetThreads(JDWP::ObjectId thread_group_id, std::vector<JDWP::ObjectId>& thread_ids) {
   1746   class ThreadListVisitor {
   1747    public:
   1748     ThreadListVisitor(const ScopedObjectAccessUnchecked& soa, mirror::Object* desired_thread_group,
   1749                       std::vector<JDWP::ObjectId>& thread_ids)
   1750         SHARED_LOCKS_REQUIRED(Locks::mutator_lock_)
   1751         : soa_(soa), desired_thread_group_(desired_thread_group), thread_ids_(thread_ids) {}
   1752 
   1753     static void Visit(Thread* t, void* arg) {
   1754       reinterpret_cast<ThreadListVisitor*>(arg)->Visit(t);
   1755     }
   1756 
   1757     // TODO: Enable annotalysis. We know lock is held in constructor, but abstraction confuses
   1758     // annotalysis.
   1759     void Visit(Thread* t) NO_THREAD_SAFETY_ANALYSIS {
   1760       if (t == Dbg::GetDebugThread()) {
   1761         // Skip the JDWP thread. Some debuggers get bent out of shape when they can't suspend and
   1762         // query all threads, so it's easier if we just don't tell them about this thread.
   1763         return;
   1764       }
   1765       mirror::Object* peer = t->GetPeer();
   1766       if (IsInDesiredThreadGroup(peer)) {
   1767         thread_ids_.push_back(gRegistry->Add(peer));
   1768       }
   1769     }
   1770 
   1771    private:
   1772     bool IsInDesiredThreadGroup(mirror::Object* peer)
   1773         SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1774       // peer might be NULL if the thread is still starting up.
   1775       if (peer == NULL) {
   1776         // We can't tell the debugger about this thread yet.
   1777         // TODO: if we identified threads to the debugger by their Thread*
   1778         // rather than their peer's mirror::Object*, we could fix this.
   1779         // Doing so might help us report ZOMBIE threads too.
   1780         return false;
   1781       }
   1782       // Do we want threads from all thread groups?
   1783       if (desired_thread_group_ == NULL) {
   1784         return true;
   1785       }
   1786       mirror::Object* group = soa_.DecodeField(WellKnownClasses::java_lang_Thread_group)->GetObject(peer);
   1787       return (group == desired_thread_group_);
   1788     }
   1789 
   1790     const ScopedObjectAccessUnchecked& soa_;
   1791     mirror::Object* const desired_thread_group_;
   1792     std::vector<JDWP::ObjectId>& thread_ids_;
   1793   };
   1794 
   1795   ScopedObjectAccessUnchecked soa(Thread::Current());
   1796   mirror::Object* thread_group = gRegistry->Get<mirror::Object*>(thread_group_id);
   1797   ThreadListVisitor tlv(soa, thread_group, thread_ids);
   1798   MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
   1799   Runtime::Current()->GetThreadList()->ForEach(ThreadListVisitor::Visit, &tlv);
   1800 }
   1801 
   1802 void Dbg::GetChildThreadGroups(JDWP::ObjectId thread_group_id, std::vector<JDWP::ObjectId>& child_thread_group_ids) {
   1803   ScopedObjectAccess soa(Thread::Current());
   1804   mirror::Object* thread_group = gRegistry->Get<mirror::Object*>(thread_group_id);
   1805 
   1806   // Get the ArrayList<ThreadGroup> "groups" out of this thread group...
   1807   mirror::ArtField* groups_field = thread_group->GetClass()->FindInstanceField("groups", "Ljava/util/List;");
   1808   mirror::Object* groups_array_list = groups_field->GetObject(thread_group);
   1809 
   1810   // Get the array and size out of the ArrayList<ThreadGroup>...
   1811   mirror::ArtField* array_field = groups_array_list->GetClass()->FindInstanceField("array", "[Ljava/lang/Object;");
   1812   mirror::ArtField* size_field = groups_array_list->GetClass()->FindInstanceField("size", "I");
   1813   mirror::ObjectArray<mirror::Object>* groups_array =
   1814       array_field->GetObject(groups_array_list)->AsObjectArray<mirror::Object>();
   1815   const int32_t size = size_field->GetInt(groups_array_list);
   1816 
   1817   // Copy the first 'size' elements out of the array into the result.
   1818   for (int32_t i = 0; i < size; ++i) {
   1819     child_thread_group_ids.push_back(gRegistry->Add(groups_array->Get(i)));
   1820   }
   1821 }
   1822 
   1823 static int GetStackDepth(Thread* thread)
   1824     SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   1825   struct CountStackDepthVisitor : public StackVisitor {
   1826     explicit CountStackDepthVisitor(Thread* thread)
   1827         : StackVisitor(thread, NULL), depth(0) {}
   1828 
   1829     // TODO: Enable annotalysis. We know lock is held in constructor, but abstraction confuses
   1830     // annotalysis.
   1831     bool VisitFrame() NO_THREAD_SAFETY_ANALYSIS {
   1832       if (!GetMethod()->IsRuntimeMethod()) {
   1833         ++depth;
   1834       }
   1835       return true;
   1836     }
   1837     size_t depth;
   1838   };
   1839 
   1840   CountStackDepthVisitor visitor(thread);
   1841   visitor.WalkStack();
   1842   return visitor.depth;
   1843 }
   1844 
   1845 JDWP::JdwpError Dbg::GetThreadFrameCount(JDWP::ObjectId thread_id, size_t& result) {
   1846   ScopedObjectAccess soa(Thread::Current());
   1847   MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
   1848   Thread* thread;
   1849   JDWP::JdwpError error = DecodeThread(soa, thread_id, thread);
   1850   if (error != JDWP::ERR_NONE) {
   1851     return error;
   1852   }
   1853   if (!IsSuspendedForDebugger(soa, thread)) {
   1854     return JDWP::ERR_THREAD_NOT_SUSPENDED;
   1855   }
   1856   result = GetStackDepth(thread);
   1857   return JDWP::ERR_NONE;
   1858 }
   1859 
   1860 JDWP::JdwpError Dbg::GetThreadFrames(JDWP::ObjectId thread_id, size_t start_frame,
   1861                                      size_t frame_count, JDWP::ExpandBuf* buf) {
   1862   class GetFrameVisitor : public StackVisitor {
   1863    public:
   1864     GetFrameVisitor(Thread* thread, size_t start_frame, size_t frame_count, JDWP::ExpandBuf* buf)
   1865         SHARED_LOCKS_REQUIRED(Locks::mutator_lock_)
   1866         : StackVisitor(thread, NULL), depth_(0),
   1867           start_frame_(start_frame), frame_count_(frame_count), buf_(buf) {
   1868       expandBufAdd4BE(buf_, frame_count_);
   1869     }
   1870 
   1871     // TODO: Enable annotalysis. We know lock is held in constructor, but abstraction confuses
   1872     // annotalysis.
   1873     virtual bool VisitFrame() NO_THREAD_SAFETY_ANALYSIS {
   1874       if (GetMethod()->IsRuntimeMethod()) {
   1875         return true;  // The debugger can't do anything useful with a frame that has no Method*.
   1876       }
   1877       if (depth_ >= start_frame_ + frame_count_) {
   1878         return false;
   1879       }
   1880       if (depth_ >= start_frame_) {
   1881         JDWP::FrameId frame_id(GetFrameId());
   1882         JDWP::JdwpLocation location;
   1883         SetLocation(location, GetMethod(), GetDexPc());
   1884         VLOG(jdwp) << StringPrintf("    Frame %3zd: id=%3lld ", depth_, frame_id) << location;
   1885         expandBufAdd8BE(buf_, frame_id);
   1886         expandBufAddLocation(buf_, location);
   1887       }
   1888       ++depth_;
   1889       return true;
   1890     }
   1891 
   1892    private:
   1893     size_t depth_;
   1894     const size_t start_frame_;
   1895     const size_t frame_count_;
   1896     JDWP::ExpandBuf* buf_;
   1897   };
   1898 
   1899   ScopedObjectAccessUnchecked soa(Thread::Current());
   1900   MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
   1901   Thread* thread;
   1902   JDWP::JdwpError error = DecodeThread(soa, thread_id, thread);
   1903   if (error != JDWP::ERR_NONE) {
   1904     return error;
   1905   }
   1906   if (!IsSuspendedForDebugger(soa, thread)) {
   1907     return JDWP::ERR_THREAD_NOT_SUSPENDED;
   1908   }
   1909   GetFrameVisitor visitor(thread, start_frame, frame_count, buf);
   1910   visitor.WalkStack();
   1911   return JDWP::ERR_NONE;
   1912 }
   1913 
   1914 JDWP::ObjectId Dbg::GetThreadSelfId() {
   1915   ScopedObjectAccessUnchecked soa(Thread::Current());
   1916   return gRegistry->Add(soa.Self()->GetPeer());
   1917 }
   1918 
   1919 void Dbg::SuspendVM() {
   1920   Runtime::Current()->GetThreadList()->SuspendAllForDebugger();
   1921 }
   1922 
   1923 void Dbg::ResumeVM() {
   1924   Runtime::Current()->GetThreadList()->UndoDebuggerSuspensions();
   1925 }
   1926 
   1927 JDWP::JdwpError Dbg::SuspendThread(JDWP::ObjectId thread_id, bool request_suspension) {
   1928   ScopedLocalRef<jobject> peer(Thread::Current()->GetJniEnv(), NULL);
   1929   {
   1930     ScopedObjectAccess soa(Thread::Current());
   1931     peer.reset(soa.AddLocalReference<jobject>(gRegistry->Get<mirror::Object*>(thread_id)));
   1932   }
   1933   if (peer.get() == NULL) {
   1934     return JDWP::ERR_THREAD_NOT_ALIVE;
   1935   }
   1936   // Suspend thread to build stack trace.
   1937   bool timed_out;
   1938   Thread* thread = Thread::SuspendForDebugger(peer.get(), request_suspension, &timed_out);
   1939   if (thread != NULL) {
   1940     return JDWP::ERR_NONE;
   1941   } else if (timed_out) {
   1942     return JDWP::ERR_INTERNAL;
   1943   } else {
   1944     return JDWP::ERR_THREAD_NOT_ALIVE;
   1945   }
   1946 }
   1947 
   1948 void Dbg::ResumeThread(JDWP::ObjectId thread_id) {
   1949   ScopedObjectAccessUnchecked soa(Thread::Current());
   1950   mirror::Object* peer = gRegistry->Get<mirror::Object*>(thread_id);
   1951   Thread* thread;
   1952   {
   1953     MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
   1954     thread = Thread::FromManagedThread(soa, peer);
   1955   }
   1956   if (thread == NULL) {
   1957     LOG(WARNING) << "No such thread for resume: " << peer;
   1958     return;
   1959   }
   1960   bool needs_resume;
   1961   {
   1962     MutexLock mu2(soa.Self(), *Locks::thread_suspend_count_lock_);
   1963     needs_resume = thread->GetSuspendCount() > 0;
   1964   }
   1965   if (needs_resume) {
   1966     Runtime::Current()->GetThreadList()->Resume(thread, true);
   1967   }
   1968 }
   1969 
   1970 void Dbg::SuspendSelf() {
   1971   Runtime::Current()->GetThreadList()->SuspendSelfForDebugger();
   1972 }
   1973 
   1974 struct GetThisVisitor : public StackVisitor {
   1975   GetThisVisitor(Thread* thread, Context* context, JDWP::FrameId frame_id)
   1976       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_)
   1977       : StackVisitor(thread, context), this_object(NULL), frame_id(frame_id) {}
   1978 
   1979   // TODO: Enable annotalysis. We know lock is held in constructor, but abstraction confuses
   1980   // annotalysis.
   1981   virtual bool VisitFrame() NO_THREAD_SAFETY_ANALYSIS {
   1982     if (frame_id != GetFrameId()) {
   1983       return true;  // continue
   1984     } else {
   1985       this_object = GetThisObject();
   1986       return false;
   1987     }
   1988   }
   1989 
   1990   mirror::Object* this_object;
   1991   JDWP::FrameId frame_id;
   1992 };
   1993 
   1994 JDWP::JdwpError Dbg::GetThisObject(JDWP::ObjectId thread_id, JDWP::FrameId frame_id,
   1995                                    JDWP::ObjectId* result) {
   1996   ScopedObjectAccessUnchecked soa(Thread::Current());
   1997   Thread* thread;
   1998   {
   1999     MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
   2000     JDWP::JdwpError error = DecodeThread(soa, thread_id, thread);
   2001     if (error != JDWP::ERR_NONE) {
   2002       return error;
   2003     }
   2004     if (!IsSuspendedForDebugger(soa, thread)) {
   2005       return JDWP::ERR_THREAD_NOT_SUSPENDED;
   2006     }
   2007   }
   2008   UniquePtr<Context> context(Context::Create());
   2009   GetThisVisitor visitor(thread, context.get(), frame_id);
   2010   visitor.WalkStack();
   2011   *result = gRegistry->Add(visitor.this_object);
   2012   return JDWP::ERR_NONE;
   2013 }
   2014 
   2015 void Dbg::GetLocalValue(JDWP::ObjectId thread_id, JDWP::FrameId frame_id, int slot, JDWP::JdwpTag tag,
   2016                         uint8_t* buf, size_t width) {
   2017   struct GetLocalVisitor : public StackVisitor {
   2018     GetLocalVisitor(Thread* thread, Context* context, JDWP::FrameId frame_id, int slot,
   2019                     JDWP::JdwpTag tag, uint8_t* buf, size_t width)
   2020         SHARED_LOCKS_REQUIRED(Locks::mutator_lock_)
   2021         : StackVisitor(thread, context), frame_id_(frame_id), slot_(slot), tag_(tag),
   2022           buf_(buf), width_(width) {}
   2023 
   2024     // TODO: Enable annotalysis. We know lock is held in constructor, but abstraction confuses
   2025     // annotalysis.
   2026     bool VisitFrame() NO_THREAD_SAFETY_ANALYSIS {
   2027       if (GetFrameId() != frame_id_) {
   2028         return true;  // Not our frame, carry on.
   2029       }
   2030       // TODO: check that the tag is compatible with the actual type of the slot!
   2031       mirror::ArtMethod* m = GetMethod();
   2032       uint16_t reg = DemangleSlot(slot_, m);
   2033 
   2034       switch (tag_) {
   2035       case JDWP::JT_BOOLEAN:
   2036         {
   2037           CHECK_EQ(width_, 1U);
   2038           uint32_t intVal = GetVReg(m, reg, kIntVReg);
   2039           VLOG(jdwp) << "get boolean local " << reg << " = " << intVal;
   2040           JDWP::Set1(buf_+1, intVal != 0);
   2041         }
   2042         break;
   2043       case JDWP::JT_BYTE:
   2044         {
   2045           CHECK_EQ(width_, 1U);
   2046           uint32_t intVal = GetVReg(m, reg, kIntVReg);
   2047           VLOG(jdwp) << "get byte local " << reg << " = " << intVal;
   2048           JDWP::Set1(buf_+1, intVal);
   2049         }
   2050         break;
   2051       case JDWP::JT_SHORT:
   2052       case JDWP::JT_CHAR:
   2053         {
   2054           CHECK_EQ(width_, 2U);
   2055           uint32_t intVal = GetVReg(m, reg, kIntVReg);
   2056           VLOG(jdwp) << "get short/char local " << reg << " = " << intVal;
   2057           JDWP::Set2BE(buf_+1, intVal);
   2058         }
   2059         break;
   2060       case JDWP::JT_INT:
   2061         {
   2062           CHECK_EQ(width_, 4U);
   2063           uint32_t intVal = GetVReg(m, reg, kIntVReg);
   2064           VLOG(jdwp) << "get int local " << reg << " = " << intVal;
   2065           JDWP::Set4BE(buf_+1, intVal);
   2066         }
   2067         break;
   2068       case JDWP::JT_FLOAT:
   2069         {
   2070           CHECK_EQ(width_, 4U);
   2071           uint32_t intVal = GetVReg(m, reg, kFloatVReg);
   2072           VLOG(jdwp) << "get int/float local " << reg << " = " << intVal;
   2073           JDWP::Set4BE(buf_+1, intVal);
   2074         }
   2075         break;
   2076       case JDWP::JT_ARRAY:
   2077         {
   2078           CHECK_EQ(width_, sizeof(JDWP::ObjectId));
   2079           mirror::Object* o = reinterpret_cast<mirror::Object*>(GetVReg(m, reg, kReferenceVReg));
   2080           VLOG(jdwp) << "get array local " << reg << " = " << o;
   2081           if (!Runtime::Current()->GetHeap()->IsHeapAddress(o)) {
   2082             LOG(FATAL) << "Register " << reg << " expected to hold array: " << o;
   2083           }
   2084           JDWP::SetObjectId(buf_+1, gRegistry->Add(o));
   2085         }
   2086         break;
   2087       case JDWP::JT_CLASS_LOADER:
   2088       case JDWP::JT_CLASS_OBJECT:
   2089       case JDWP::JT_OBJECT:
   2090       case JDWP::JT_STRING:
   2091       case JDWP::JT_THREAD:
   2092       case JDWP::JT_THREAD_GROUP:
   2093         {
   2094           CHECK_EQ(width_, sizeof(JDWP::ObjectId));
   2095           mirror::Object* o = reinterpret_cast<mirror::Object*>(GetVReg(m, reg, kReferenceVReg));
   2096           VLOG(jdwp) << "get object local " << reg << " = " << o;
   2097           if (!Runtime::Current()->GetHeap()->IsHeapAddress(o)) {
   2098             LOG(FATAL) << "Register " << reg << " expected to hold object: " << o;
   2099           }
   2100           tag_ = TagFromObject(o);
   2101           JDWP::SetObjectId(buf_+1, gRegistry->Add(o));
   2102         }
   2103         break;
   2104       case JDWP::JT_DOUBLE:
   2105         {
   2106           CHECK_EQ(width_, 8U);
   2107           uint32_t lo = GetVReg(m, reg, kDoubleLoVReg);
   2108           uint64_t hi = GetVReg(m, reg + 1, kDoubleHiVReg);
   2109           uint64_t longVal = (hi << 32) | lo;
   2110           VLOG(jdwp) << "get double/long local " << hi << ":" << lo << " = " << longVal;
   2111           JDWP::Set8BE(buf_+1, longVal);
   2112         }
   2113         break;
   2114       case JDWP::JT_LONG:
   2115         {
   2116           CHECK_EQ(width_, 8U);
   2117           uint32_t lo = GetVReg(m, reg, kLongLoVReg);
   2118           uint64_t hi = GetVReg(m, reg + 1, kLongHiVReg);
   2119           uint64_t longVal = (hi << 32) | lo;
   2120           VLOG(jdwp) << "get double/long local " << hi << ":" << lo << " = " << longVal;
   2121           JDWP::Set8BE(buf_+1, longVal);
   2122         }
   2123         break;
   2124       default:
   2125         LOG(FATAL) << "Unknown tag " << tag_;
   2126         break;
   2127       }
   2128 
   2129       // Prepend tag, which may have been updated.
   2130       JDWP::Set1(buf_, tag_);
   2131       return false;
   2132     }
   2133 
   2134     const JDWP::FrameId frame_id_;
   2135     const int slot_;
   2136     JDWP::JdwpTag tag_;
   2137     uint8_t* const buf_;
   2138     const size_t width_;
   2139   };
   2140 
   2141   ScopedObjectAccessUnchecked soa(Thread::Current());
   2142   MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
   2143   Thread* thread;
   2144   JDWP::JdwpError error = DecodeThread(soa, thread_id, thread);
   2145   if (error != JDWP::ERR_NONE) {
   2146     return;
   2147   }
   2148   UniquePtr<Context> context(Context::Create());
   2149   GetLocalVisitor visitor(thread, context.get(), frame_id, slot, tag, buf, width);
   2150   visitor.WalkStack();
   2151 }
   2152 
   2153 void Dbg::SetLocalValue(JDWP::ObjectId thread_id, JDWP::FrameId frame_id, int slot, JDWP::JdwpTag tag,
   2154                         uint64_t value, size_t width) {
   2155   struct SetLocalVisitor : public StackVisitor {
   2156     SetLocalVisitor(Thread* thread, Context* context,
   2157                     JDWP::FrameId frame_id, int slot, JDWP::JdwpTag tag, uint64_t value,
   2158                     size_t width)
   2159         SHARED_LOCKS_REQUIRED(Locks::mutator_lock_)
   2160         : StackVisitor(thread, context),
   2161           frame_id_(frame_id), slot_(slot), tag_(tag), value_(value), width_(width) {}
   2162 
   2163     // TODO: Enable annotalysis. We know lock is held in constructor, but abstraction confuses
   2164     // annotalysis.
   2165     bool VisitFrame() NO_THREAD_SAFETY_ANALYSIS {
   2166       if (GetFrameId() != frame_id_) {
   2167         return true;  // Not our frame, carry on.
   2168       }
   2169       // TODO: check that the tag is compatible with the actual type of the slot!
   2170       mirror::ArtMethod* m = GetMethod();
   2171       uint16_t reg = DemangleSlot(slot_, m);
   2172 
   2173       switch (tag_) {
   2174         case JDWP::JT_BOOLEAN:
   2175         case JDWP::JT_BYTE:
   2176           CHECK_EQ(width_, 1U);
   2177           SetVReg(m, reg, static_cast<uint32_t>(value_), kIntVReg);
   2178           break;
   2179         case JDWP::JT_SHORT:
   2180         case JDWP::JT_CHAR:
   2181           CHECK_EQ(width_, 2U);
   2182           SetVReg(m, reg, static_cast<uint32_t>(value_), kIntVReg);
   2183           break;
   2184         case JDWP::JT_INT:
   2185           CHECK_EQ(width_, 4U);
   2186           SetVReg(m, reg, static_cast<uint32_t>(value_), kIntVReg);
   2187           break;
   2188         case JDWP::JT_FLOAT:
   2189           CHECK_EQ(width_, 4U);
   2190           SetVReg(m, reg, static_cast<uint32_t>(value_), kFloatVReg);
   2191           break;
   2192         case JDWP::JT_ARRAY:
   2193         case JDWP::JT_OBJECT:
   2194         case JDWP::JT_STRING:
   2195         {
   2196           CHECK_EQ(width_, sizeof(JDWP::ObjectId));
   2197           mirror::Object* o = gRegistry->Get<mirror::Object*>(static_cast<JDWP::ObjectId>(value_));
   2198           if (o == ObjectRegistry::kInvalidObject) {
   2199             UNIMPLEMENTED(FATAL) << "return an error code when given an invalid object to store";
   2200           }
   2201           SetVReg(m, reg, static_cast<uint32_t>(reinterpret_cast<uintptr_t>(o)), kReferenceVReg);
   2202         }
   2203         break;
   2204         case JDWP::JT_DOUBLE:
   2205           CHECK_EQ(width_, 8U);
   2206           SetVReg(m, reg, static_cast<uint32_t>(value_), kDoubleLoVReg);
   2207           SetVReg(m, reg + 1, static_cast<uint32_t>(value_ >> 32), kDoubleHiVReg);
   2208           break;
   2209         case JDWP::JT_LONG:
   2210           CHECK_EQ(width_, 8U);
   2211           SetVReg(m, reg, static_cast<uint32_t>(value_), kLongLoVReg);
   2212           SetVReg(m, reg + 1, static_cast<uint32_t>(value_ >> 32), kLongHiVReg);
   2213           break;
   2214         default:
   2215           LOG(FATAL) << "Unknown tag " << tag_;
   2216           break;
   2217       }
   2218       return false;
   2219     }
   2220 
   2221     const JDWP::FrameId frame_id_;
   2222     const int slot_;
   2223     const JDWP::JdwpTag tag_;
   2224     const uint64_t value_;
   2225     const size_t width_;
   2226   };
   2227 
   2228   ScopedObjectAccessUnchecked soa(Thread::Current());
   2229   MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
   2230   Thread* thread;
   2231   JDWP::JdwpError error = DecodeThread(soa, thread_id, thread);
   2232   if (error != JDWP::ERR_NONE) {
   2233     return;
   2234   }
   2235   UniquePtr<Context> context(Context::Create());
   2236   SetLocalVisitor visitor(thread, context.get(), frame_id, slot, tag, value, width);
   2237   visitor.WalkStack();
   2238 }
   2239 
   2240 void Dbg::PostLocationEvent(const mirror::ArtMethod* m, int dex_pc,
   2241                             mirror::Object* this_object, int event_flags) {
   2242   mirror::Class* c = m->GetDeclaringClass();
   2243 
   2244   JDWP::JdwpLocation location;
   2245   location.type_tag = c->IsInterface() ? JDWP::TT_INTERFACE : JDWP::TT_CLASS;
   2246   location.class_id = gRegistry->AddRefType(c);
   2247   location.method_id = ToMethodId(m);
   2248   location.dex_pc = m->IsNative() ? -1 : dex_pc;
   2249 
   2250   // If 'this_object' isn't already in the registry, we know that we're not looking for it,
   2251   // so there's no point adding it to the registry and burning through ids.
   2252   JDWP::ObjectId this_id = 0;
   2253   if (gRegistry->Contains(this_object)) {
   2254     this_id = gRegistry->Add(this_object);
   2255   }
   2256   gJdwpState->PostLocationEvent(&location, this_id, event_flags);
   2257 }
   2258 
   2259 void Dbg::PostException(Thread* thread, const ThrowLocation& throw_location,
   2260                         mirror::ArtMethod* catch_method,
   2261                         uint32_t catch_dex_pc, mirror::Throwable* exception_object) {
   2262   if (!IsDebuggerActive()) {
   2263     return;
   2264   }
   2265 
   2266   JDWP::JdwpLocation jdwp_throw_location;
   2267   SetLocation(jdwp_throw_location, throw_location.GetMethod(), throw_location.GetDexPc());
   2268   JDWP::JdwpLocation catch_location;
   2269   SetLocation(catch_location, catch_method, catch_dex_pc);
   2270 
   2271   // We need 'this' for InstanceOnly filters.
   2272   JDWP::ObjectId this_id = gRegistry->Add(throw_location.GetThis());
   2273   JDWP::ObjectId exception_id = gRegistry->Add(exception_object);
   2274   JDWP::RefTypeId exception_class_id = gRegistry->AddRefType(exception_object->GetClass());
   2275 
   2276   gJdwpState->PostException(&jdwp_throw_location, exception_id, exception_class_id, &catch_location,
   2277                             this_id);
   2278 }
   2279 
   2280 void Dbg::PostClassPrepare(mirror::Class* c) {
   2281   if (!IsDebuggerActive()) {
   2282     return;
   2283   }
   2284 
   2285   // OLD-TODO - we currently always send both "verified" and "prepared" since
   2286   // debuggers seem to like that.  There might be some advantage to honesty,
   2287   // since the class may not yet be verified.
   2288   int state = JDWP::CS_VERIFIED | JDWP::CS_PREPARED;
   2289   JDWP::JdwpTypeTag tag = c->IsInterface() ? JDWP::TT_INTERFACE : JDWP::TT_CLASS;
   2290   gJdwpState->PostClassPrepare(tag, gRegistry->Add(c), ClassHelper(c).GetDescriptor(), state);
   2291 }
   2292 
   2293 void Dbg::UpdateDebugger(Thread* thread, mirror::Object* this_object,
   2294                          const mirror::ArtMethod* m, uint32_t dex_pc) {
   2295   if (!IsDebuggerActive() || dex_pc == static_cast<uint32_t>(-2) /* fake method exit */) {
   2296     return;
   2297   }
   2298 
   2299   int event_flags = 0;
   2300 
   2301   if (IsBreakpoint(m, dex_pc)) {
   2302     event_flags |= kBreakpoint;
   2303   }
   2304 
   2305   {
   2306     // If the debugger is single-stepping one of our threads, check to
   2307     // see if we're that thread and we've reached a step point.
   2308     MutexLock mu(Thread::Current(), *Locks::breakpoint_lock_);
   2309     if (gSingleStepControl.is_active && gSingleStepControl.thread == thread) {
   2310       CHECK(!m->IsNative());
   2311       if (gSingleStepControl.step_depth == JDWP::SD_INTO) {
   2312         // Step into method calls.  We break when the line number
   2313         // or method pointer changes.  If we're in SS_MIN mode, we
   2314         // always stop.
   2315         if (gSingleStepControl.method != m) {
   2316           event_flags |= kSingleStep;
   2317           VLOG(jdwp) << "SS new method";
   2318         } else if (gSingleStepControl.step_size == JDWP::SS_MIN) {
   2319           event_flags |= kSingleStep;
   2320           VLOG(jdwp) << "SS new instruction";
   2321         } else if (gSingleStepControl.dex_pcs.find(dex_pc) == gSingleStepControl.dex_pcs.end()) {
   2322           event_flags |= kSingleStep;
   2323           VLOG(jdwp) << "SS new line";
   2324         }
   2325       } else if (gSingleStepControl.step_depth == JDWP::SD_OVER) {
   2326         // Step over method calls.  We break when the line number is
   2327         // different and the frame depth is <= the original frame
   2328         // depth.  (We can't just compare on the method, because we
   2329         // might get unrolled past it by an exception, and it's tricky
   2330         // to identify recursion.)
   2331 
   2332         int stack_depth = GetStackDepth(thread);
   2333 
   2334         if (stack_depth < gSingleStepControl.stack_depth) {
   2335           // popped up one or more frames, always trigger
   2336           event_flags |= kSingleStep;
   2337           VLOG(jdwp) << "SS method pop";
   2338         } else if (stack_depth == gSingleStepControl.stack_depth) {
   2339           // same depth, see if we moved
   2340           if (gSingleStepControl.step_size == JDWP::SS_MIN) {
   2341             event_flags |= kSingleStep;
   2342             VLOG(jdwp) << "SS new instruction";
   2343           } else if (gSingleStepControl.dex_pcs.find(dex_pc) == gSingleStepControl.dex_pcs.end()) {
   2344             event_flags |= kSingleStep;
   2345             VLOG(jdwp) << "SS new line";
   2346           }
   2347         }
   2348       } else {
   2349         CHECK_EQ(gSingleStepControl.step_depth, JDWP::SD_OUT);
   2350         // Return from the current method.  We break when the frame
   2351         // depth pops up.
   2352 
   2353         // This differs from the "method exit" break in that it stops
   2354         // with the PC at the next instruction in the returned-to
   2355         // function, rather than the end of the returning function.
   2356 
   2357         int stack_depth = GetStackDepth(thread);
   2358         if (stack_depth < gSingleStepControl.stack_depth) {
   2359           event_flags |= kSingleStep;
   2360           VLOG(jdwp) << "SS method pop";
   2361         }
   2362       }
   2363     }
   2364   }
   2365 
   2366   // If there's something interesting going on, see if it matches one
   2367   // of the debugger filters.
   2368   if (event_flags != 0) {
   2369     Dbg::PostLocationEvent(m, dex_pc, this_object, event_flags);
   2370   }
   2371 }
   2372 
   2373 void Dbg::WatchLocation(const JDWP::JdwpLocation* location) {
   2374   MutexLock mu(Thread::Current(), *Locks::breakpoint_lock_);
   2375   mirror::ArtMethod* m = FromMethodId(location->method_id);
   2376   gBreakpoints.push_back(Breakpoint(m, location->dex_pc));
   2377   VLOG(jdwp) << "Set breakpoint #" << (gBreakpoints.size() - 1) << ": " << gBreakpoints[gBreakpoints.size() - 1];
   2378 }
   2379 
   2380 void Dbg::UnwatchLocation(const JDWP::JdwpLocation* location) {
   2381   MutexLock mu(Thread::Current(), *Locks::breakpoint_lock_);
   2382   mirror::ArtMethod* m = FromMethodId(location->method_id);
   2383   for (size_t i = 0; i < gBreakpoints.size(); ++i) {
   2384     if (gBreakpoints[i].method == m && gBreakpoints[i].dex_pc == location->dex_pc) {
   2385       VLOG(jdwp) << "Removed breakpoint #" << i << ": " << gBreakpoints[i];
   2386       gBreakpoints.erase(gBreakpoints.begin() + i);
   2387       return;
   2388     }
   2389   }
   2390 }
   2391 
   2392 // Scoped utility class to suspend a thread so that we may do tasks such as walk its stack. Doesn't
   2393 // cause suspension if the thread is the current thread.
   2394 class ScopedThreadSuspension {
   2395  public:
   2396   ScopedThreadSuspension(Thread* self, JDWP::ObjectId thread_id)
   2397       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) :
   2398       thread_(NULL),
   2399       error_(JDWP::ERR_NONE),
   2400       self_suspend_(false),
   2401       other_suspend_(false) {
   2402     ScopedObjectAccessUnchecked soa(self);
   2403     {
   2404       MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
   2405       error_ = DecodeThread(soa, thread_id, thread_);
   2406     }
   2407     if (error_ == JDWP::ERR_NONE) {
   2408       if (thread_ == soa.Self()) {
   2409         self_suspend_ = true;
   2410       } else {
   2411         soa.Self()->TransitionFromRunnableToSuspended(kWaitingForDebuggerSuspension);
   2412         jobject thread_peer = gRegistry->GetJObject(thread_id);
   2413         bool timed_out;
   2414         Thread* suspended_thread = Thread::SuspendForDebugger(thread_peer, true, &timed_out);
   2415         CHECK_EQ(soa.Self()->TransitionFromSuspendedToRunnable(), kWaitingForDebuggerSuspension);
   2416         if (suspended_thread == NULL) {
   2417           // Thread terminated from under us while suspending.
   2418           error_ = JDWP::ERR_INVALID_THREAD;
   2419         } else {
   2420           CHECK_EQ(suspended_thread, thread_);
   2421           other_suspend_ = true;
   2422         }
   2423       }
   2424     }
   2425   }
   2426 
   2427   Thread* GetThread() const {
   2428     return thread_;
   2429   }
   2430 
   2431   JDWP::JdwpError GetError() const {
   2432     return error_;
   2433   }
   2434 
   2435   ~ScopedThreadSuspension() {
   2436     if (other_suspend_) {
   2437       Runtime::Current()->GetThreadList()->Resume(thread_, true);
   2438     }
   2439   }
   2440 
   2441  private:
   2442   Thread* thread_;
   2443   JDWP::JdwpError error_;
   2444   bool self_suspend_;
   2445   bool other_suspend_;
   2446 };
   2447 
   2448 JDWP::JdwpError Dbg::ConfigureStep(JDWP::ObjectId thread_id, JDWP::JdwpStepSize step_size,
   2449                                    JDWP::JdwpStepDepth step_depth) {
   2450   Thread* self = Thread::Current();
   2451   ScopedThreadSuspension sts(self, thread_id);
   2452   if (sts.GetError() != JDWP::ERR_NONE) {
   2453     return sts.GetError();
   2454   }
   2455 
   2456   MutexLock mu2(self, *Locks::breakpoint_lock_);
   2457   // TODO: there's no theoretical reason why we couldn't support single-stepping
   2458   // of multiple threads at once, but we never did so historically.
   2459   if (gSingleStepControl.thread != NULL && sts.GetThread() != gSingleStepControl.thread) {
   2460     LOG(WARNING) << "single-step already active for " << *gSingleStepControl.thread
   2461                  << "; switching to " << *sts.GetThread();
   2462   }
   2463 
   2464   //
   2465   // Work out what Method* we're in, the current line number, and how deep the stack currently
   2466   // is for step-out.
   2467   //
   2468 
   2469   struct SingleStepStackVisitor : public StackVisitor {
   2470     explicit SingleStepStackVisitor(Thread* thread)
   2471         EXCLUSIVE_LOCKS_REQUIRED(Locks::breakpoint_lock_)
   2472         SHARED_LOCKS_REQUIRED(Locks::mutator_lock_)
   2473         : StackVisitor(thread, NULL) {
   2474       gSingleStepControl.method = NULL;
   2475       gSingleStepControl.stack_depth = 0;
   2476     }
   2477 
   2478     // TODO: Enable annotalysis. We know lock is held in constructor, but abstraction confuses
   2479     // annotalysis.
   2480     bool VisitFrame() NO_THREAD_SAFETY_ANALYSIS {
   2481       Locks::breakpoint_lock_->AssertHeld(Thread::Current());
   2482       const mirror::ArtMethod* m = GetMethod();
   2483       if (!m->IsRuntimeMethod()) {
   2484         ++gSingleStepControl.stack_depth;
   2485         if (gSingleStepControl.method == NULL) {
   2486           const mirror::DexCache* dex_cache = m->GetDeclaringClass()->GetDexCache();
   2487           gSingleStepControl.method = m;
   2488           gSingleStepControl.line_number = -1;
   2489           if (dex_cache != NULL) {
   2490             const DexFile& dex_file = *dex_cache->GetDexFile();
   2491             gSingleStepControl.line_number = dex_file.GetLineNumFromPC(m, GetDexPc());
   2492           }
   2493         }
   2494       }
   2495       return true;
   2496     }
   2497   };
   2498 
   2499   SingleStepStackVisitor visitor(sts.GetThread());
   2500   visitor.WalkStack();
   2501 
   2502   //
   2503   // Find the dex_pc values that correspond to the current line, for line-based single-stepping.
   2504   //
   2505 
   2506   struct DebugCallbackContext {
   2507     DebugCallbackContext() EXCLUSIVE_LOCKS_REQUIRED(Locks::breakpoint_lock_) {
   2508       last_pc_valid = false;
   2509       last_pc = 0;
   2510     }
   2511 
   2512     // TODO: Enable annotalysis. We know lock is held in constructor, but abstraction confuses
   2513     // annotalysis.
   2514     static bool Callback(void* raw_context, uint32_t address, uint32_t line_number) NO_THREAD_SAFETY_ANALYSIS {
   2515       Locks::breakpoint_lock_->AssertHeld(Thread::Current());
   2516       DebugCallbackContext* context = reinterpret_cast<DebugCallbackContext*>(raw_context);
   2517       if (static_cast<int32_t>(line_number) == gSingleStepControl.line_number) {
   2518         if (!context->last_pc_valid) {
   2519           // Everything from this address until the next line change is ours.
   2520           context->last_pc = address;
   2521           context->last_pc_valid = true;
   2522         }
   2523         // Otherwise, if we're already in a valid range for this line,
   2524         // just keep going (shouldn't really happen)...
   2525       } else if (context->last_pc_valid) {  // and the line number is new
   2526         // Add everything from the last entry up until here to the set
   2527         for (uint32_t dex_pc = context->last_pc; dex_pc < address; ++dex_pc) {
   2528           gSingleStepControl.dex_pcs.insert(dex_pc);
   2529         }
   2530         context->last_pc_valid = false;
   2531       }
   2532       return false;  // There may be multiple entries for any given line.
   2533     }
   2534 
   2535     // TODO: Enable annotalysis. We know lock is held in constructor, but abstraction confuses
   2536     // annotalysis.
   2537     ~DebugCallbackContext() NO_THREAD_SAFETY_ANALYSIS {
   2538       Locks::breakpoint_lock_->AssertHeld(Thread::Current());
   2539       // If the line number was the last in the position table...
   2540       if (last_pc_valid) {
   2541         size_t end = MethodHelper(gSingleStepControl.method).GetCodeItem()->insns_size_in_code_units_;
   2542         for (uint32_t dex_pc = last_pc; dex_pc < end; ++dex_pc) {
   2543           gSingleStepControl.dex_pcs.insert(dex_pc);
   2544         }
   2545       }
   2546     }
   2547 
   2548     bool last_pc_valid;
   2549     uint32_t last_pc;
   2550   };
   2551   gSingleStepControl.dex_pcs.clear();
   2552   const mirror::ArtMethod* m = gSingleStepControl.method;
   2553   if (m->IsNative()) {
   2554     gSingleStepControl.line_number = -1;
   2555   } else {
   2556     DebugCallbackContext context;
   2557     MethodHelper mh(m);
   2558     mh.GetDexFile().DecodeDebugInfo(mh.GetCodeItem(), m->IsStatic(), m->GetDexMethodIndex(),
   2559                                     DebugCallbackContext::Callback, NULL, &context);
   2560   }
   2561 
   2562   //
   2563   // Everything else...
   2564   //
   2565 
   2566   gSingleStepControl.thread = sts.GetThread();
   2567   gSingleStepControl.step_size = step_size;
   2568   gSingleStepControl.step_depth = step_depth;
   2569   gSingleStepControl.is_active = true;
   2570 
   2571   if (VLOG_IS_ON(jdwp)) {
   2572     VLOG(jdwp) << "Single-step thread: " << *gSingleStepControl.thread;
   2573     VLOG(jdwp) << "Single-step step size: " << gSingleStepControl.step_size;
   2574     VLOG(jdwp) << "Single-step step depth: " << gSingleStepControl.step_depth;
   2575     VLOG(jdwp) << "Single-step current method: " << PrettyMethod(gSingleStepControl.method);
   2576     VLOG(jdwp) << "Single-step current line: " << gSingleStepControl.line_number;
   2577     VLOG(jdwp) << "Single-step current stack depth: " << gSingleStepControl.stack_depth;
   2578     VLOG(jdwp) << "Single-step dex_pc values:";
   2579     for (std::set<uint32_t>::iterator it = gSingleStepControl.dex_pcs.begin() ; it != gSingleStepControl.dex_pcs.end(); ++it) {
   2580       VLOG(jdwp) << StringPrintf(" %#x", *it);
   2581     }
   2582   }
   2583 
   2584   return JDWP::ERR_NONE;
   2585 }
   2586 
   2587 void Dbg::UnconfigureStep(JDWP::ObjectId /*thread_id*/) {
   2588   MutexLock mu(Thread::Current(), *Locks::breakpoint_lock_);
   2589 
   2590   gSingleStepControl.is_active = false;
   2591   gSingleStepControl.thread = NULL;
   2592   gSingleStepControl.dex_pcs.clear();
   2593 }
   2594 
   2595 static char JdwpTagToShortyChar(JDWP::JdwpTag tag) {
   2596   switch (tag) {
   2597     default:
   2598       LOG(FATAL) << "unknown JDWP tag: " << PrintableChar(tag);
   2599 
   2600     // Primitives.
   2601     case JDWP::JT_BYTE:    return 'B';
   2602     case JDWP::JT_CHAR:    return 'C';
   2603     case JDWP::JT_FLOAT:   return 'F';
   2604     case JDWP::JT_DOUBLE:  return 'D';
   2605     case JDWP::JT_INT:     return 'I';
   2606     case JDWP::JT_LONG:    return 'J';
   2607     case JDWP::JT_SHORT:   return 'S';
   2608     case JDWP::JT_VOID:    return 'V';
   2609     case JDWP::JT_BOOLEAN: return 'Z';
   2610 
   2611     // Reference types.
   2612     case JDWP::JT_ARRAY:
   2613     case JDWP::JT_OBJECT:
   2614     case JDWP::JT_STRING:
   2615     case JDWP::JT_THREAD:
   2616     case JDWP::JT_THREAD_GROUP:
   2617     case JDWP::JT_CLASS_LOADER:
   2618     case JDWP::JT_CLASS_OBJECT:
   2619       return 'L';
   2620   }
   2621 }
   2622 
   2623 JDWP::JdwpError Dbg::InvokeMethod(JDWP::ObjectId thread_id, JDWP::ObjectId object_id,
   2624                                   JDWP::RefTypeId class_id, JDWP::MethodId method_id,
   2625                                   uint32_t arg_count, uint64_t* arg_values,
   2626                                   JDWP::JdwpTag* arg_types, uint32_t options,
   2627                                   JDWP::JdwpTag* pResultTag, uint64_t* pResultValue,
   2628                                   JDWP::ObjectId* pExceptionId) {
   2629   ThreadList* thread_list = Runtime::Current()->GetThreadList();
   2630 
   2631   Thread* targetThread = NULL;
   2632   DebugInvokeReq* req = NULL;
   2633   Thread* self = Thread::Current();
   2634   {
   2635     ScopedObjectAccessUnchecked soa(self);
   2636     MutexLock mu(soa.Self(), *Locks::thread_list_lock_);
   2637     JDWP::JdwpError error = DecodeThread(soa, thread_id, targetThread);
   2638     if (error != JDWP::ERR_NONE) {
   2639       LOG(ERROR) << "InvokeMethod request for invalid thread id " << thread_id;
   2640       return error;
   2641     }
   2642     req = targetThread->GetInvokeReq();
   2643     if (!req->ready) {
   2644       LOG(ERROR) << "InvokeMethod request for thread not stopped by event: " << *targetThread;
   2645       return JDWP::ERR_INVALID_THREAD;
   2646     }
   2647 
   2648     /*
   2649      * We currently have a bug where we don't successfully resume the
   2650      * target thread if the suspend count is too deep.  We're expected to
   2651      * require one "resume" for each "suspend", but when asked to execute
   2652      * a method we have to resume fully and then re-suspend it back to the
   2653      * same level.  (The easiest way to cause this is to type "suspend"
   2654      * multiple times in jdb.)
   2655      *
   2656      * It's unclear what this means when the event specifies "resume all"
   2657      * and some threads are suspended more deeply than others.  This is
   2658      * a rare problem, so for now we just prevent it from hanging forever
   2659      * by rejecting the method invocation request.  Without this, we will
   2660      * be stuck waiting on a suspended thread.
   2661      */
   2662     int suspend_count;
   2663     {
   2664       MutexLock mu2(soa.Self(), *Locks::thread_suspend_count_lock_);
   2665       suspend_count = targetThread->GetSuspendCount();
   2666     }
   2667     if (suspend_count > 1) {
   2668       LOG(ERROR) << *targetThread << " suspend count too deep for method invocation: " << suspend_count;
   2669       return JDWP::ERR_THREAD_SUSPENDED;  // Probably not expected here.
   2670     }
   2671 
   2672     JDWP::JdwpError status;
   2673     mirror::Object* receiver = gRegistry->Get<mirror::Object*>(object_id);
   2674     if (receiver == ObjectRegistry::kInvalidObject) {
   2675       return JDWP::ERR_INVALID_OBJECT;
   2676     }
   2677 
   2678     mirror::Object* thread = gRegistry->Get<mirror::Object*>(thread_id);
   2679     if (thread == ObjectRegistry::kInvalidObject) {
   2680       return JDWP::ERR_INVALID_OBJECT;
   2681     }
   2682     // TODO: check that 'thread' is actually a java.lang.Thread!
   2683 
   2684     mirror::Class* c = DecodeClass(class_id, status);
   2685     if (c == NULL) {
   2686       return status;
   2687     }
   2688 
   2689     mirror::ArtMethod* m = FromMethodId(method_id);
   2690     if (m->IsStatic() != (receiver == NULL)) {
   2691       return JDWP::ERR_INVALID_METHODID;
   2692     }
   2693     if (m->IsStatic()) {
   2694       if (m->GetDeclaringClass() != c) {
   2695         return JDWP::ERR_INVALID_METHODID;
   2696       }
   2697     } else {
   2698       if (!m->GetDeclaringClass()->IsAssignableFrom(c)) {
   2699         return JDWP::ERR_INVALID_METHODID;
   2700       }
   2701     }
   2702 
   2703     // Check the argument list matches the method.
   2704     MethodHelper mh(m);
   2705     if (mh.GetShortyLength() - 1 != arg_count) {
   2706       return JDWP::ERR_ILLEGAL_ARGUMENT;
   2707     }
   2708     const char* shorty = mh.GetShorty();
   2709     const DexFile::TypeList* types = mh.GetParameterTypeList();
   2710     for (size_t i = 0; i < arg_count; ++i) {
   2711       if (shorty[i + 1] != JdwpTagToShortyChar(arg_types[i])) {
   2712         return JDWP::ERR_ILLEGAL_ARGUMENT;
   2713       }
   2714 
   2715       if (shorty[i + 1] == 'L') {
   2716         // Did we really get an argument of an appropriate reference type?
   2717         mirror::Class* parameter_type = mh.GetClassFromTypeIdx(types->GetTypeItem(i).type_idx_);
   2718         mirror::Object* argument = gRegistry->Get<mirror::Object*>(arg_values[i]);
   2719         if (argument == ObjectRegistry::kInvalidObject) {
   2720           return JDWP::ERR_INVALID_OBJECT;
   2721         }
   2722         if (!argument->InstanceOf(parameter_type)) {
   2723           return JDWP::ERR_ILLEGAL_ARGUMENT;
   2724         }
   2725 
   2726         // Turn the on-the-wire ObjectId into a jobject.
   2727         jvalue& v = reinterpret_cast<jvalue&>(arg_values[i]);
   2728         v.l = gRegistry->GetJObject(arg_values[i]);
   2729       }
   2730     }
   2731 
   2732     req->receiver_ = receiver;
   2733     req->thread_ = thread;
   2734     req->class_ = c;
   2735     req->method_ = m;
   2736     req->arg_count_ = arg_count;
   2737     req->arg_values_ = arg_values;
   2738     req->options_ = options;
   2739     req->invoke_needed_ = true;
   2740   }
   2741 
   2742   // The fact that we've released the thread list lock is a bit risky --- if the thread goes
   2743   // away we're sitting high and dry -- but we must release this before the ResumeAllThreads
   2744   // call, and it's unwise to hold it during WaitForSuspend.
   2745 
   2746   {
   2747     /*
   2748      * We change our (JDWP thread) status, which should be THREAD_RUNNING,
   2749      * so we can suspend for a GC if the invoke request causes us to
   2750      * run out of memory.  It's also a good idea to change it before locking
   2751      * the invokeReq mutex, although that should never be held for long.
   2752      */
   2753     self->TransitionFromRunnableToSuspended(kWaitingForDebuggerSend);
   2754 
   2755     VLOG(jdwp) << "    Transferring control to event thread";
   2756     {
   2757       MutexLock mu(self, req->lock_);
   2758 
   2759       if ((options & JDWP::INVOKE_SINGLE_THREADED) == 0) {
   2760         VLOG(jdwp) << "      Resuming all threads";
   2761         thread_list->UndoDebuggerSuspensions();
   2762       } else {
   2763         VLOG(jdwp) << "      Resuming event thread only";
   2764         thread_list->Resume(targetThread, true);
   2765       }
   2766 
   2767       // Wait for the request to finish executing.
   2768       while (req->invoke_needed_) {
   2769         req->cond_.Wait(self);
   2770       }
   2771     }
   2772     VLOG(jdwp) << "    Control has returned from event thread";
   2773 
   2774     /* wait for thread to re-suspend itself */
   2775     SuspendThread(thread_id, false /* request_suspension */);
   2776     self->TransitionFromSuspendedToRunnable();
   2777   }
   2778 
   2779   /*
   2780    * Suspend the threads.  We waited for the target thread to suspend
   2781    * itself, so all we need to do is suspend the others.
   2782    *
   2783    * The suspendAllThreads() call will double-suspend the event thread,
   2784    * so we want to resume the target thread once to keep the books straight.
   2785    */
   2786   if ((options & JDWP::INVOKE_SINGLE_THREADED) == 0) {
   2787     self->TransitionFromRunnableToSuspended(kWaitingForDebuggerSuspension);
   2788     VLOG(jdwp) << "      Suspending all threads";
   2789     thread_list->SuspendAllForDebugger();
   2790     self->TransitionFromSuspendedToRunnable();
   2791     VLOG(jdwp) << "      Resuming event thread to balance the count";
   2792     thread_list->Resume(targetThread, true);
   2793   }
   2794 
   2795   // Copy the result.
   2796   *pResultTag = req->result_tag;
   2797   if (IsPrimitiveTag(req->result_tag)) {
   2798     *pResultValue = req->result_value.GetJ();
   2799   } else {
   2800     *pResultValue = gRegistry->Add(req->result_value.GetL());
   2801   }
   2802   *pExceptionId = req->exception;
   2803   return req->error;
   2804 }
   2805 
   2806 void Dbg::ExecuteMethod(DebugInvokeReq* pReq) {
   2807   ScopedObjectAccess soa(Thread::Current());
   2808 
   2809   // We can be called while an exception is pending. We need
   2810   // to preserve that across the method invocation.
   2811   SirtRef<mirror::Object> old_throw_this_object(soa.Self(), NULL);
   2812   SirtRef<mirror::ArtMethod> old_throw_method(soa.Self(), NULL);
   2813   SirtRef<mirror::Throwable> old_exception(soa.Self(), NULL);
   2814   uint32_t old_throw_dex_pc;
   2815   {
   2816     ThrowLocation old_throw_location;
   2817     mirror::Throwable* old_exception_obj = soa.Self()->GetException(&old_throw_location);
   2818     old_throw_this_object.reset(old_throw_location.GetThis());
   2819     old_throw_method.reset(old_throw_location.GetMethod());
   2820     old_exception.reset(old_exception_obj);
   2821     old_throw_dex_pc = old_throw_location.GetDexPc();
   2822     soa.Self()->ClearException();
   2823   }
   2824 
   2825   // Translate the method through the vtable, unless the debugger wants to suppress it.
   2826   mirror::ArtMethod* m = pReq->method_;
   2827   if ((pReq->options_ & JDWP::INVOKE_NONVIRTUAL) == 0 && pReq->receiver_ != NULL) {
   2828     mirror::ArtMethod* actual_method = pReq->class_->FindVirtualMethodForVirtualOrInterface(pReq->method_);
   2829     if (actual_method != m) {
   2830       VLOG(jdwp) << "ExecuteMethod translated " << PrettyMethod(m) << " to " << PrettyMethod(actual_method);
   2831       m = actual_method;
   2832     }
   2833   }
   2834   VLOG(jdwp) << "ExecuteMethod " << PrettyMethod(m)
   2835              << " receiver=" << pReq->receiver_
   2836              << " arg_count=" << pReq->arg_count_;
   2837   CHECK(m != NULL);
   2838 
   2839   CHECK_EQ(sizeof(jvalue), sizeof(uint64_t));
   2840 
   2841   MethodHelper mh(m);
   2842   ArgArray arg_array(mh.GetShorty(), mh.GetShortyLength());
   2843   arg_array.BuildArgArray(soa, pReq->receiver_, reinterpret_cast<jvalue*>(pReq->arg_values_));
   2844   InvokeWithArgArray(soa, m, &arg_array, &pReq->result_value, mh.GetShorty()[0]);
   2845 
   2846   mirror::Throwable* exception = soa.Self()->GetException(NULL);
   2847   soa.Self()->ClearException();
   2848   pReq->exception = gRegistry->Add(exception);
   2849   pReq->result_tag = BasicTagFromDescriptor(MethodHelper(m).GetShorty());
   2850   if (pReq->exception != 0) {
   2851     VLOG(jdwp) << "  JDWP invocation returning with exception=" << exception
   2852         << " " << exception->Dump();
   2853     pReq->result_value.SetJ(0);
   2854   } else if (pReq->result_tag == JDWP::JT_OBJECT) {
   2855     /* if no exception thrown, examine object result more closely */
   2856     JDWP::JdwpTag new_tag = TagFromObject(pReq->result_value.GetL());
   2857     if (new_tag != pReq->result_tag) {
   2858       VLOG(jdwp) << "  JDWP promoted result from " << pReq->result_tag << " to " << new_tag;
   2859       pReq->result_tag = new_tag;
   2860     }
   2861 
   2862     /*
   2863      * Register the object.  We don't actually need an ObjectId yet,
   2864      * but we do need to be sure that the GC won't move or discard the
   2865      * object when we switch out of RUNNING.  The ObjectId conversion
   2866      * will add the object to the "do not touch" list.
   2867      *
   2868      * We can't use the "tracked allocation" mechanism here because
   2869      * the object is going to be handed off to a different thread.
   2870      */
   2871     gRegistry->Add(pReq->result_value.GetL());
   2872   }
   2873 
   2874   if (old_exception.get() != NULL) {
   2875     ThrowLocation gc_safe_throw_location(old_throw_this_object.get(), old_throw_method.get(),
   2876                                          old_throw_dex_pc);
   2877     soa.Self()->SetException(gc_safe_throw_location, old_exception.get());
   2878   }
   2879 }
   2880 
   2881 /*
   2882  * "request" contains a full JDWP packet, possibly with multiple chunks.  We
   2883  * need to process each, accumulate the replies, and ship the whole thing
   2884  * back.
   2885  *
   2886  * Returns "true" if we have a reply.  The reply buffer is newly allocated,
   2887  * and includes the chunk type/length, followed by the data.
   2888  *
   2889  * OLD-TODO: we currently assume that the request and reply include a single
   2890  * chunk.  If this becomes inconvenient we will need to adapt.
   2891  */
   2892 bool Dbg::DdmHandlePacket(JDWP::Request& request, uint8_t** pReplyBuf, int* pReplyLen) {
   2893   Thread* self = Thread::Current();
   2894   JNIEnv* env = self->GetJniEnv();
   2895 
   2896   uint32_t type = request.ReadUnsigned32("type");
   2897   uint32_t length = request.ReadUnsigned32("length");
   2898 
   2899   // Create a byte[] corresponding to 'request'.
   2900   size_t request_length = request.size();
   2901   ScopedLocalRef<jbyteArray> dataArray(env, env->NewByteArray(request_length));
   2902   if (dataArray.get() == NULL) {
   2903     LOG(WARNING) << "byte[] allocation failed: " << request_length;
   2904     env->ExceptionClear();
   2905     return false;
   2906   }
   2907   env->SetByteArrayRegion(dataArray.get(), 0, request_length, reinterpret_cast<const jbyte*>(request.data()));
   2908   request.Skip(request_length);
   2909 
   2910   // Run through and find all chunks.  [Currently just find the first.]
   2911   ScopedByteArrayRO contents(env, dataArray.get());
   2912   if (length != request_length) {
   2913     LOG(WARNING) << StringPrintf("bad chunk found (len=%u pktLen=%d)", length, request_length);
   2914     return false;
   2915   }
   2916 
   2917   // Call "private static Chunk dispatch(int type, byte[] data, int offset, int length)".
   2918   ScopedLocalRef<jobject> chunk(env, env->CallStaticObjectMethod(WellKnownClasses::org_apache_harmony_dalvik_ddmc_DdmServer,
   2919                                                                  WellKnownClasses::org_apache_harmony_dalvik_ddmc_DdmServer_dispatch,
   2920                                                                  type, dataArray.get(), 0, length));
   2921   if (env->ExceptionCheck()) {
   2922     LOG(INFO) << StringPrintf("Exception thrown by dispatcher for 0x%08x", type);
   2923     env->ExceptionDescribe();
   2924     env->ExceptionClear();
   2925     return false;
   2926   }
   2927 
   2928   if (chunk.get() == NULL) {
   2929     return false;
   2930   }
   2931 
   2932   /*
   2933    * Pull the pieces out of the chunk.  We copy the results into a
   2934    * newly-allocated buffer that the caller can free.  We don't want to
   2935    * continue using the Chunk object because nothing has a reference to it.
   2936    *
   2937    * We could avoid this by returning type/data/offset/length and having
   2938    * the caller be aware of the object lifetime issues, but that
   2939    * integrates the JDWP code more tightly into the rest of the runtime, and doesn't work
   2940    * if we have responses for multiple chunks.
   2941    *
   2942    * So we're pretty much stuck with copying data around multiple times.
   2943    */
   2944   ScopedLocalRef<jbyteArray> replyData(env, reinterpret_cast<jbyteArray>(env->GetObjectField(chunk.get(), WellKnownClasses::org_apache_harmony_dalvik_ddmc_Chunk_data)));
   2945   jint offset = env->GetIntField(chunk.get(), WellKnownClasses::org_apache_harmony_dalvik_ddmc_Chunk_offset);
   2946   length = env->GetIntField(chunk.get(), WellKnownClasses::org_apache_harmony_dalvik_ddmc_Chunk_length);
   2947   type = env->GetIntField(chunk.get(), WellKnownClasses::org_apache_harmony_dalvik_ddmc_Chunk_type);
   2948 
   2949   VLOG(jdwp) << StringPrintf("DDM reply: type=0x%08x data=%p offset=%d length=%d", type, replyData.get(), offset, length);
   2950   if (length == 0 || replyData.get() == NULL) {
   2951     return false;
   2952   }
   2953 
   2954   const int kChunkHdrLen = 8;
   2955   uint8_t* reply = new uint8_t[length + kChunkHdrLen];
   2956   if (reply == NULL) {
   2957     LOG(WARNING) << "malloc failed: " << (length + kChunkHdrLen);
   2958     return false;
   2959   }
   2960   JDWP::Set4BE(reply + 0, type);
   2961   JDWP::Set4BE(reply + 4, length);
   2962   env->GetByteArrayRegion(replyData.get(), offset, length, reinterpret_cast<jbyte*>(reply + kChunkHdrLen));
   2963 
   2964   *pReplyBuf = reply;
   2965   *pReplyLen = length + kChunkHdrLen;
   2966 
   2967   VLOG(jdwp) << StringPrintf("dvmHandleDdm returning type=%.4s %p len=%d", reinterpret_cast<char*>(reply), reply, length);
   2968   return true;
   2969 }
   2970 
   2971 void Dbg::DdmBroadcast(bool connect) {
   2972   VLOG(jdwp) << "Broadcasting DDM " << (connect ? "connect" : "disconnect") << "...";
   2973 
   2974   Thread* self = Thread::Current();
   2975   if (self->GetState() != kRunnable) {
   2976     LOG(ERROR) << "DDM broadcast in thread state " << self->GetState();
   2977     /* try anyway? */
   2978   }
   2979 
   2980   JNIEnv* env = self->GetJniEnv();
   2981   jint event = connect ? 1 /*DdmServer.CONNECTED*/ : 2 /*DdmServer.DISCONNECTED*/;
   2982   env->CallStaticVoidMethod(WellKnownClasses::org_apache_harmony_dalvik_ddmc_DdmServer,
   2983                             WellKnownClasses::org_apache_harmony_dalvik_ddmc_DdmServer_broadcast,
   2984                             event);
   2985   if (env->ExceptionCheck()) {
   2986     LOG(ERROR) << "DdmServer.broadcast " << event << " failed";
   2987     env->ExceptionDescribe();
   2988     env->ExceptionClear();
   2989   }
   2990 }
   2991 
   2992 void Dbg::DdmConnected() {
   2993   Dbg::DdmBroadcast(true);
   2994 }
   2995 
   2996 void Dbg::DdmDisconnected() {
   2997   Dbg::DdmBroadcast(false);
   2998   gDdmThreadNotification = false;
   2999 }
   3000 
   3001 /*
   3002  * Send a notification when a thread starts, stops, or changes its name.
   3003  *
   3004  * Because we broadcast the full set of threads when the notifications are
   3005  * first enabled, it's possible for "thread" to be actively executing.
   3006  */
   3007 void Dbg::DdmSendThreadNotification(Thread* t, uint32_t type) {
   3008   if (!gDdmThreadNotification) {
   3009     return;
   3010   }
   3011 
   3012   if (type == CHUNK_TYPE("THDE")) {
   3013     uint8_t buf[4];
   3014     JDWP::Set4BE(&buf[0], t->GetThinLockId());
   3015     Dbg::DdmSendChunk(CHUNK_TYPE("THDE"), 4, buf);
   3016   } else {
   3017     CHECK(type == CHUNK_TYPE("THCR") || type == CHUNK_TYPE("THNM")) << type;
   3018     ScopedObjectAccessUnchecked soa(Thread::Current());
   3019     SirtRef<mirror::String> name(soa.Self(), t->GetThreadName(soa));
   3020     size_t char_count = (name.get() != NULL) ? name->GetLength() : 0;
   3021     const jchar* chars = (name.get() != NULL) ? name->GetCharArray()->GetData() : NULL;
   3022 
   3023     std::vector<uint8_t> bytes;
   3024     JDWP::Append4BE(bytes, t->GetThinLockId());
   3025     JDWP::AppendUtf16BE(bytes, chars, char_count);
   3026     CHECK_EQ(bytes.size(), char_count*2 + sizeof(uint32_t)*2);
   3027     Dbg::DdmSendChunk(type, bytes);
   3028   }
   3029 }
   3030 
   3031 void Dbg::DdmSetThreadNotification(bool enable) {
   3032   // Enable/disable thread notifications.
   3033   gDdmThreadNotification = enable;
   3034   if (enable) {
   3035     // Suspend the VM then post thread start notifications for all threads. Threads attaching will
   3036     // see a suspension in progress and block until that ends. They then post their own start
   3037     // notification.
   3038     SuspendVM();
   3039     std::list<Thread*> threads;
   3040     Thread* self = Thread::Current();
   3041     {
   3042       MutexLock mu(self, *Locks::thread_list_lock_);
   3043       threads = Runtime::Current()->GetThreadList()->GetList();
   3044     }
   3045     {
   3046       ScopedObjectAccess soa(self);
   3047       for (Thread* thread : threads) {
   3048         Dbg::DdmSendThreadNotification(thread, CHUNK_TYPE("THCR"));
   3049       }
   3050     }
   3051     ResumeVM();
   3052   }
   3053 }
   3054 
   3055 void Dbg::PostThreadStartOrStop(Thread* t, uint32_t type) {
   3056   if (IsDebuggerActive()) {
   3057     ScopedObjectAccessUnchecked soa(Thread::Current());
   3058     JDWP::ObjectId id = gRegistry->Add(t->GetPeer());
   3059     gJdwpState->PostThreadChange(id, type == CHUNK_TYPE("THCR"));
   3060   }
   3061   Dbg::DdmSendThreadNotification(t, type);
   3062 }
   3063 
   3064 void Dbg::PostThreadStart(Thread* t) {
   3065   Dbg::PostThreadStartOrStop(t, CHUNK_TYPE("THCR"));
   3066 }
   3067 
   3068 void Dbg::PostThreadDeath(Thread* t) {
   3069   Dbg::PostThreadStartOrStop(t, CHUNK_TYPE("THDE"));
   3070 }
   3071 
   3072 void Dbg::DdmSendChunk(uint32_t type, size_t byte_count, const uint8_t* buf) {
   3073   CHECK(buf != NULL);
   3074   iovec vec[1];
   3075   vec[0].iov_base = reinterpret_cast<void*>(const_cast<uint8_t*>(buf));
   3076   vec[0].iov_len = byte_count;
   3077   Dbg::DdmSendChunkV(type, vec, 1);
   3078 }
   3079 
   3080 void Dbg::DdmSendChunk(uint32_t type, const std::vector<uint8_t>& bytes) {
   3081   DdmSendChunk(type, bytes.size(), &bytes[0]);
   3082 }
   3083 
   3084 void Dbg::DdmSendChunkV(uint32_t type, const iovec* iov, int iov_count) {
   3085   if (gJdwpState == NULL) {
   3086     VLOG(jdwp) << "Debugger thread not active, ignoring DDM send: " << type;
   3087   } else {
   3088     gJdwpState->DdmSendChunkV(type, iov, iov_count);
   3089   }
   3090 }
   3091 
   3092 int Dbg::DdmHandleHpifChunk(HpifWhen when) {
   3093   if (when == HPIF_WHEN_NOW) {
   3094     DdmSendHeapInfo(when);
   3095     return true;
   3096   }
   3097 
   3098   if (when != HPIF_WHEN_NEVER && when != HPIF_WHEN_NEXT_GC && when != HPIF_WHEN_EVERY_GC) {
   3099     LOG(ERROR) << "invalid HpifWhen value: " << static_cast<int>(when);
   3100     return false;
   3101   }
   3102 
   3103   gDdmHpifWhen = when;
   3104   return true;
   3105 }
   3106 
   3107 bool Dbg::DdmHandleHpsgNhsgChunk(Dbg::HpsgWhen when, Dbg::HpsgWhat what, bool native) {
   3108   if (when != HPSG_WHEN_NEVER && when != HPSG_WHEN_EVERY_GC) {
   3109     LOG(ERROR) << "invalid HpsgWhen value: " << static_cast<int>(when);
   3110     return false;
   3111   }
   3112 
   3113   if (what != HPSG_WHAT_MERGED_OBJECTS && what != HPSG_WHAT_DISTINCT_OBJECTS) {
   3114     LOG(ERROR) << "invalid HpsgWhat value: " << static_cast<int>(what);
   3115     return false;
   3116   }
   3117 
   3118   if (native) {
   3119     gDdmNhsgWhen = when;
   3120     gDdmNhsgWhat = what;
   3121   } else {
   3122     gDdmHpsgWhen = when;
   3123     gDdmHpsgWhat = what;
   3124   }
   3125   return true;
   3126 }
   3127 
   3128 void Dbg::DdmSendHeapInfo(HpifWhen reason) {
   3129   // If there's a one-shot 'when', reset it.
   3130   if (reason == gDdmHpifWhen) {
   3131     if (gDdmHpifWhen == HPIF_WHEN_NEXT_GC) {
   3132       gDdmHpifWhen = HPIF_WHEN_NEVER;
   3133     }
   3134   }
   3135 
   3136   /*
   3137    * Chunk HPIF (client --> server)
   3138    *
   3139    * Heap Info. General information about the heap,
   3140    * suitable for a summary display.
   3141    *
   3142    *   [u4]: number of heaps
   3143    *
   3144    *   For each heap:
   3145    *     [u4]: heap ID
   3146    *     [u8]: timestamp in ms since Unix epoch
   3147    *     [u1]: capture reason (same as 'when' value from server)
   3148    *     [u4]: max heap size in bytes (-Xmx)
   3149    *     [u4]: current heap size in bytes
   3150    *     [u4]: current number of bytes allocated
   3151    *     [u4]: current number of objects allocated
   3152    */
   3153   uint8_t heap_count = 1;
   3154   gc::Heap* heap = Runtime::Current()->GetHeap();
   3155   std::vector<uint8_t> bytes;
   3156   JDWP::Append4BE(bytes, heap_count);
   3157   JDWP::Append4BE(bytes, 1);  // Heap id (bogus; we only have one heap).
   3158   JDWP::Append8BE(bytes, MilliTime());
   3159   JDWP::Append1BE(bytes, reason);
   3160   JDWP::Append4BE(bytes, heap->GetMaxMemory());  // Max allowed heap size in bytes.
   3161   JDWP::Append4BE(bytes, heap->GetTotalMemory());  // Current heap size in bytes.
   3162   JDWP::Append4BE(bytes, heap->GetBytesAllocated());
   3163   JDWP::Append4BE(bytes, heap->GetObjectsAllocated());
   3164   CHECK_EQ(bytes.size(), 4U + (heap_count * (4 + 8 + 1 + 4 + 4 + 4 + 4)));
   3165   Dbg::DdmSendChunk(CHUNK_TYPE("HPIF"), bytes);
   3166 }
   3167 
   3168 enum HpsgSolidity {
   3169   SOLIDITY_FREE = 0,
   3170   SOLIDITY_HARD = 1,
   3171   SOLIDITY_SOFT = 2,
   3172   SOLIDITY_WEAK = 3,
   3173   SOLIDITY_PHANTOM = 4,
   3174   SOLIDITY_FINALIZABLE = 5,
   3175   SOLIDITY_SWEEP = 6,
   3176 };
   3177 
   3178 enum HpsgKind {
   3179   KIND_OBJECT = 0,
   3180   KIND_CLASS_OBJECT = 1,
   3181   KIND_ARRAY_1 = 2,
   3182   KIND_ARRAY_2 = 3,
   3183   KIND_ARRAY_4 = 4,
   3184   KIND_ARRAY_8 = 5,
   3185   KIND_UNKNOWN = 6,
   3186   KIND_NATIVE = 7,
   3187 };
   3188 
   3189 #define HPSG_PARTIAL (1<<7)
   3190 #define HPSG_STATE(solidity, kind) ((uint8_t)((((kind) & 0x7) << 3) | ((solidity) & 0x7)))
   3191 
   3192 class HeapChunkContext {
   3193  public:
   3194   // Maximum chunk size.  Obtain this from the formula:
   3195   // (((maximum_heap_size / ALLOCATION_UNIT_SIZE) + 255) / 256) * 2
   3196   HeapChunkContext(bool merge, bool native)
   3197       : buf_(16384 - 16),
   3198         type_(0),
   3199         merge_(merge) {
   3200     Reset();
   3201     if (native) {
   3202       type_ = CHUNK_TYPE("NHSG");
   3203     } else {
   3204       type_ = merge ? CHUNK_TYPE("HPSG") : CHUNK_TYPE("HPSO");
   3205     }
   3206   }
   3207 
   3208   ~HeapChunkContext() {
   3209     if (p_ > &buf_[0]) {
   3210       Flush();
   3211     }
   3212   }
   3213 
   3214   void EnsureHeader(const void* chunk_ptr) {
   3215     if (!needHeader_) {
   3216       return;
   3217     }
   3218 
   3219     // Start a new HPSx chunk.
   3220     JDWP::Write4BE(&p_, 1);  // Heap id (bogus; we only have one heap).
   3221     JDWP::Write1BE(&p_, 8);  // Size of allocation unit, in bytes.
   3222 
   3223     JDWP::Write4BE(&p_, reinterpret_cast<uintptr_t>(chunk_ptr));  // virtual address of segment start.
   3224     JDWP::Write4BE(&p_, 0);  // offset of this piece (relative to the virtual address).
   3225     // [u4]: length of piece, in allocation units
   3226     // We won't know this until we're done, so save the offset and stuff in a dummy value.
   3227     pieceLenField_ = p_;
   3228     JDWP::Write4BE(&p_, 0x55555555);
   3229     needHeader_ = false;
   3230   }
   3231 
   3232   void Flush() SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   3233     if (pieceLenField_ == NULL) {
   3234       // Flush immediately post Reset (maybe back-to-back Flush). Ignore.
   3235       CHECK(needHeader_);
   3236       return;
   3237     }
   3238     // Patch the "length of piece" field.
   3239     CHECK_LE(&buf_[0], pieceLenField_);
   3240     CHECK_LE(pieceLenField_, p_);
   3241     JDWP::Set4BE(pieceLenField_, totalAllocationUnits_);
   3242 
   3243     Dbg::DdmSendChunk(type_, p_ - &buf_[0], &buf_[0]);
   3244     Reset();
   3245   }
   3246 
   3247   static void HeapChunkCallback(void* start, void* end, size_t used_bytes, void* arg)
   3248       SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_,
   3249                             Locks::mutator_lock_) {
   3250     reinterpret_cast<HeapChunkContext*>(arg)->HeapChunkCallback(start, end, used_bytes);
   3251   }
   3252 
   3253  private:
   3254   enum { ALLOCATION_UNIT_SIZE = 8 };
   3255 
   3256   void Reset() {
   3257     p_ = &buf_[0];
   3258     startOfNextMemoryChunk_ = NULL;
   3259     totalAllocationUnits_ = 0;
   3260     needHeader_ = true;
   3261     pieceLenField_ = NULL;
   3262   }
   3263 
   3264   void HeapChunkCallback(void* start, void* /*end*/, size_t used_bytes)
   3265       SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_,
   3266                             Locks::mutator_lock_) {
   3267     // Note: heap call backs cannot manipulate the heap upon which they are crawling, care is taken
   3268     // in the following code not to allocate memory, by ensuring buf_ is of the correct size
   3269     if (used_bytes == 0) {
   3270         if (start == NULL) {
   3271             // Reset for start of new heap.
   3272             startOfNextMemoryChunk_ = NULL;
   3273             Flush();
   3274         }
   3275         // Only process in use memory so that free region information
   3276         // also includes dlmalloc book keeping.
   3277         return;
   3278     }
   3279 
   3280     /* If we're looking at the native heap, we'll just return
   3281      * (SOLIDITY_HARD, KIND_NATIVE) for all allocated chunks
   3282      */
   3283     bool native = type_ == CHUNK_TYPE("NHSG");
   3284 
   3285     if (startOfNextMemoryChunk_ != NULL) {
   3286         // Transmit any pending free memory. Native free memory of
   3287         // over kMaxFreeLen could be because of the use of mmaps, so
   3288         // don't report. If not free memory then start a new segment.
   3289         bool flush = true;
   3290         if (start > startOfNextMemoryChunk_) {
   3291             const size_t kMaxFreeLen = 2 * kPageSize;
   3292             void* freeStart = startOfNextMemoryChunk_;
   3293             void* freeEnd = start;
   3294             size_t freeLen = reinterpret_cast<char*>(freeEnd) - reinterpret_cast<char*>(freeStart);
   3295             if (!native || freeLen < kMaxFreeLen) {
   3296                 AppendChunk(HPSG_STATE(SOLIDITY_FREE, 0), freeStart, freeLen);
   3297                 flush = false;
   3298             }
   3299         }
   3300         if (flush) {
   3301             startOfNextMemoryChunk_ = NULL;
   3302             Flush();
   3303         }
   3304     }
   3305     const mirror::Object* obj = reinterpret_cast<const mirror::Object*>(start);
   3306 
   3307     // Determine the type of this chunk.
   3308     // OLD-TODO: if context.merge, see if this chunk is different from the last chunk.
   3309     // If it's the same, we should combine them.
   3310     uint8_t state = ExamineObject(obj, native);
   3311     // dlmalloc's chunk header is 2 * sizeof(size_t), but if the previous chunk is in use for an
   3312     // allocation then the first sizeof(size_t) may belong to it.
   3313     const size_t dlMallocOverhead = sizeof(size_t);
   3314     AppendChunk(state, start, used_bytes + dlMallocOverhead);
   3315     startOfNextMemoryChunk_ = reinterpret_cast<char*>(start) + used_bytes + dlMallocOverhead;
   3316   }
   3317 
   3318   void AppendChunk(uint8_t state, void* ptr, size_t length)
   3319       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) {
   3320     // Make sure there's enough room left in the buffer.
   3321     // We need to use two bytes for every fractional 256 allocation units used by the chunk plus
   3322     // 17 bytes for any header.
   3323     size_t needed = (((length/ALLOCATION_UNIT_SIZE + 255) / 256) * 2) + 17;
   3324     size_t bytesLeft = buf_.size() - (size_t)(p_ - &buf_[0]);
   3325     if (bytesLeft < needed) {
   3326       Flush();
   3327     }
   3328 
   3329     bytesLeft = buf_.size() - (size_t)(p_ - &buf_[0]);
   3330     if (bytesLeft < needed) {
   3331       LOG(WARNING) << "Chunk is too big to transmit (chunk_len=" << length << ", "
   3332           << needed << " bytes)";
   3333       return;
   3334     }
   3335     EnsureHeader(ptr);
   3336     // Write out the chunk description.
   3337     length /= ALLOCATION_UNIT_SIZE;   // Convert to allocation units.
   3338     totalAllocationUnits_ += length;
   3339     while (length > 256) {
   3340       *p_++ = state | HPSG_PARTIAL;
   3341       *p_++ = 255;     // length - 1
   3342       length -= 256;
   3343     }
   3344     *p_++ = state;
   3345     *p_++ = length - 1;
   3346   }
   3347 
   3348   uint8_t ExamineObject(const mirror::Object* o, bool is_native_heap)
   3349       SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_) {
   3350     if (o == NULL) {
   3351       return HPSG_STATE(SOLIDITY_FREE, 0);
   3352     }
   3353 
   3354     // It's an allocated chunk. Figure out what it is.
   3355 
   3356     // If we're looking at the native heap, we'll just return
   3357     // (SOLIDITY_HARD, KIND_NATIVE) for all allocated chunks.
   3358     if (is_native_heap) {
   3359       return HPSG_STATE(SOLIDITY_HARD, KIND_NATIVE);
   3360     }
   3361 
   3362     if (!Runtime::Current()->GetHeap()->IsLiveObjectLocked(o)) {
   3363       return HPSG_STATE(SOLIDITY_HARD, KIND_NATIVE);
   3364     }
   3365 
   3366     mirror::Class* c = o->GetClass();
   3367     if (c == NULL) {
   3368       // The object was probably just created but hasn't been initialized yet.
   3369       return HPSG_STATE(SOLIDITY_HARD, KIND_OBJECT);
   3370     }
   3371 
   3372     if (!Runtime::Current()->GetHeap()->IsHeapAddress(c)) {
   3373       LOG(ERROR) << "Invalid class for managed heap object: " << o << " " << c;
   3374       return HPSG_STATE(SOLIDITY_HARD, KIND_UNKNOWN);
   3375     }
   3376 
   3377     if (c->IsClassClass()) {
   3378       return HPSG_STATE(SOLIDITY_HARD, KIND_CLASS_OBJECT);
   3379     }
   3380 
   3381     if (c->IsArrayClass()) {
   3382       if (o->IsObjectArray()) {
   3383         return HPSG_STATE(SOLIDITY_HARD, KIND_ARRAY_4);
   3384       }
   3385       switch (c->GetComponentSize()) {
   3386       case 1: return HPSG_STATE(SOLIDITY_HARD, KIND_ARRAY_1);
   3387       case 2: return HPSG_STATE(SOLIDITY_HARD, KIND_ARRAY_2);
   3388       case 4: return HPSG_STATE(SOLIDITY_HARD, KIND_ARRAY_4);
   3389       case 8: return HPSG_STATE(SOLIDITY_HARD, KIND_ARRAY_8);
   3390       }
   3391     }
   3392 
   3393     return HPSG_STATE(SOLIDITY_HARD, KIND_OBJECT);
   3394   }
   3395 
   3396   std::vector<uint8_t> buf_;
   3397   uint8_t* p_;
   3398   uint8_t* pieceLenField_;
   3399   void* startOfNextMemoryChunk_;
   3400   size_t totalAllocationUnits_;
   3401   uint32_t type_;
   3402   bool merge_;
   3403   bool needHeader_;
   3404 
   3405   DISALLOW_COPY_AND_ASSIGN(HeapChunkContext);
   3406 };
   3407 
   3408 void Dbg::DdmSendHeapSegments(bool native) {
   3409   Dbg::HpsgWhen when;
   3410   Dbg::HpsgWhat what;
   3411   if (!native) {
   3412     when = gDdmHpsgWhen;
   3413     what = gDdmHpsgWhat;
   3414   } else {
   3415     when = gDdmNhsgWhen;
   3416     what = gDdmNhsgWhat;
   3417   }
   3418   if (when == HPSG_WHEN_NEVER) {
   3419     return;
   3420   }
   3421 
   3422   // Figure out what kind of chunks we'll be sending.
   3423   CHECK(what == HPSG_WHAT_MERGED_OBJECTS || what == HPSG_WHAT_DISTINCT_OBJECTS) << static_cast<int>(what);
   3424 
   3425   // First, send a heap start chunk.
   3426   uint8_t heap_id[4];
   3427   JDWP::Set4BE(&heap_id[0], 1);  // Heap id (bogus; we only have one heap).
   3428   Dbg::DdmSendChunk(native ? CHUNK_TYPE("NHST") : CHUNK_TYPE("HPST"), sizeof(heap_id), heap_id);
   3429 
   3430   // Send a series of heap segment chunks.
   3431   HeapChunkContext context((what == HPSG_WHAT_MERGED_OBJECTS), native);
   3432   if (native) {
   3433     dlmalloc_inspect_all(HeapChunkContext::HeapChunkCallback, &context);
   3434   } else {
   3435     gc::Heap* heap = Runtime::Current()->GetHeap();
   3436     const std::vector<gc::space::ContinuousSpace*>& spaces = heap->GetContinuousSpaces();
   3437     Thread* self = Thread::Current();
   3438     ReaderMutexLock mu(self, *Locks::heap_bitmap_lock_);
   3439     typedef std::vector<gc::space::ContinuousSpace*>::const_iterator It;
   3440     for (It cur = spaces.begin(), end = spaces.end(); cur != end; ++cur) {
   3441       if ((*cur)->IsDlMallocSpace()) {
   3442         (*cur)->AsDlMallocSpace()->Walk(HeapChunkContext::HeapChunkCallback, &context);
   3443       }
   3444     }
   3445     // Walk the large objects, these are not in the AllocSpace.
   3446     heap->GetLargeObjectsSpace()->Walk(HeapChunkContext::HeapChunkCallback, &context);
   3447   }
   3448 
   3449   // Finally, send a heap end chunk.
   3450   Dbg::DdmSendChunk(native ? CHUNK_TYPE("NHEN") : CHUNK_TYPE("HPEN"), sizeof(heap_id), heap_id);
   3451 }
   3452 
   3453 static size_t GetAllocTrackerMax() {
   3454 #ifdef HAVE_ANDROID_OS
   3455   // Check whether there's a system property overriding the number of records.
   3456   const char* propertyName = "dalvik.vm.allocTrackerMax";
   3457   char allocRecordMaxString[PROPERTY_VALUE_MAX];
   3458   if (property_get(propertyName, allocRecordMaxString, "") > 0) {
   3459     char* end;
   3460     size_t value = strtoul(allocRecordMaxString, &end, 10);
   3461     if (*end != '\0') {
   3462       LOG(ERROR) << "Ignoring  " << propertyName << " '" << allocRecordMaxString
   3463                  << "' --- invalid";
   3464       return kDefaultNumAllocRecords;
   3465     }
   3466     if (!IsPowerOfTwo(value)) {
   3467       LOG(ERROR) << "Ignoring  " << propertyName << " '" << allocRecordMaxString
   3468                  << "' --- not power of two";
   3469       return kDefaultNumAllocRecords;
   3470     }
   3471     return value;
   3472   }
   3473 #endif
   3474   return kDefaultNumAllocRecords;
   3475 }
   3476 
   3477 void Dbg::SetAllocTrackingEnabled(bool enabled) {
   3478   MutexLock mu(Thread::Current(), gAllocTrackerLock);
   3479   if (enabled) {
   3480     if (recent_allocation_records_ == NULL) {
   3481       gAllocRecordMax = GetAllocTrackerMax();
   3482       LOG(INFO) << "Enabling alloc tracker (" << gAllocRecordMax << " entries of "
   3483                 << kMaxAllocRecordStackDepth << " frames, taking "
   3484                 << PrettySize(sizeof(AllocRecord) * gAllocRecordMax) << ")";
   3485       gAllocRecordHead = gAllocRecordCount = 0;
   3486       recent_allocation_records_ = new AllocRecord[gAllocRecordMax];
   3487       CHECK(recent_allocation_records_ != NULL);
   3488     }
   3489   } else {
   3490     delete[] recent_allocation_records_;
   3491     recent_allocation_records_ = NULL;
   3492   }
   3493 }
   3494 
   3495 struct AllocRecordStackVisitor : public StackVisitor {
   3496   AllocRecordStackVisitor(Thread* thread, AllocRecord* record)
   3497       SHARED_LOCKS_REQUIRED(Locks::mutator_lock_)
   3498       : StackVisitor(thread, NULL), record(record), depth(0) {}
   3499 
   3500   // TODO: Enable annotalysis. We know lock is held in constructor, but abstraction confuses
   3501   // annotalysis.
   3502   bool VisitFrame() NO_THREAD_SAFETY_ANALYSIS {
   3503     if (depth >= kMaxAllocRecordStackDepth) {
   3504       return false;
   3505     }
   3506     mirror::ArtMethod* m = GetMethod();
   3507     if (!m->IsRuntimeMethod()) {
   3508       record->stack[depth].method = m;
   3509       record->stack[depth].dex_pc = GetDexPc();
   3510       ++depth;
   3511     }
   3512     return true;
   3513   }
   3514 
   3515   ~AllocRecordStackVisitor() {
   3516     // Clear out any unused stack trace elements.
   3517     for (; depth < kMaxAllocRecordStackDepth; ++depth) {
   3518       record->stack[depth].method = NULL;
   3519       record->stack[depth].dex_pc = 0;
   3520     }
   3521   }
   3522 
   3523   AllocRecord* record;
   3524   size_t depth;
   3525 };
   3526 
   3527 void Dbg::RecordAllocation(mirror::Class* type, size_t byte_count) {
   3528   Thread* self = Thread::Current();
   3529   CHECK(self != NULL);
   3530 
   3531   MutexLock mu(self, gAllocTrackerLock);
   3532   if (recent_allocation_records_ == NULL) {
   3533     return;
   3534   }
   3535 
   3536   // Advance and clip.
   3537   if (++gAllocRecordHead == gAllocRecordMax) {
   3538     gAllocRecordHead = 0;
   3539   }
   3540 
   3541   // Fill in the basics.
   3542   AllocRecord* record = &recent_allocation_records_[gAllocRecordHead];
   3543   record->type = type;
   3544   record->byte_count = byte_count;
   3545   record->thin_lock_id = self->GetThinLockId();
   3546 
   3547   // Fill in the stack trace.
   3548   AllocRecordStackVisitor visitor(self, record);
   3549   visitor.WalkStack();
   3550 
   3551   if (gAllocRecordCount < gAllocRecordMax) {
   3552     ++gAllocRecordCount;
   3553   }
   3554 }
   3555 
   3556 // Returns the index of the head element.
   3557 //
   3558 // We point at the most-recently-written record, so if gAllocRecordCount is 1
   3559 // we want to use the current element.  Take "head+1" and subtract count
   3560 // from it.
   3561 //
   3562 // We need to handle underflow in our circular buffer, so we add
   3563 // gAllocRecordMax and then mask it back down.
   3564 static inline int HeadIndex() EXCLUSIVE_LOCKS_REQUIRED(gAllocTrackerLock) {
   3565   return (gAllocRecordHead+1 + gAllocRecordMax - gAllocRecordCount) & (gAllocRecordMax-1);
   3566 }
   3567 
   3568 void Dbg::DumpRecentAllocations() {
   3569   ScopedObjectAccess soa(Thread::Current());
   3570   MutexLock mu(soa.Self(), gAllocTrackerLock);
   3571   if (recent_allocation_records_ == NULL) {
   3572     LOG(INFO) << "Not recording tracked allocations";
   3573     return;
   3574   }
   3575 
   3576   // "i" is the head of the list.  We want to start at the end of the
   3577   // list and move forward to the tail.
   3578   size_t i = HeadIndex();
   3579   size_t count = gAllocRecordCount;
   3580 
   3581   LOG(INFO) << "Tracked allocations, (head=" << gAllocRecordHead << " count=" << count << ")";
   3582   while (count--) {
   3583     AllocRecord* record = &recent_allocation_records_[i];
   3584 
   3585     LOG(INFO) << StringPrintf(" Thread %-2d %6zd bytes ", record->thin_lock_id, record->byte_count)
   3586               << PrettyClass(record->type);
   3587 
   3588     for (size_t stack_frame = 0; stack_frame < kMaxAllocRecordStackDepth; ++stack_frame) {
   3589       const mirror::ArtMethod* m = record->stack[stack_frame].method;
   3590       if (m == NULL) {
   3591         break;
   3592       }
   3593       LOG(INFO) << "    " << PrettyMethod(m) << " line " << record->stack[stack_frame].LineNumber();
   3594     }
   3595 
   3596     // pause periodically to help logcat catch up
   3597     if ((count % 5) == 0) {
   3598       usleep(40000);
   3599     }
   3600 
   3601     i = (i + 1) & (gAllocRecordMax-1);
   3602   }
   3603 }
   3604 
   3605 class StringTable {
   3606  public:
   3607   StringTable() {
   3608   }
   3609 
   3610   void Add(const char* s) {
   3611     table_.insert(s);
   3612   }
   3613 
   3614   size_t IndexOf(const char* s) const {
   3615     auto it = table_.find(s);
   3616     if (it == table_.end()) {
   3617       LOG(FATAL) << "IndexOf(\"" << s << "\") failed";
   3618     }
   3619     return std::distance(table_.begin(), it);
   3620   }
   3621 
   3622   size_t Size() const {
   3623     return table_.size();
   3624   }
   3625 
   3626   void WriteTo(std::vector<uint8_t>& bytes) const {
   3627     for (const std::string& str : table_) {
   3628       const char* s = str.c_str();
   3629       size_t s_len = CountModifiedUtf8Chars(s);
   3630       UniquePtr<uint16_t> s_utf16(new uint16_t[s_len]);
   3631       ConvertModifiedUtf8ToUtf16(s_utf16.get(), s);
   3632       JDWP::AppendUtf16BE(bytes, s_utf16.get(), s_len);
   3633     }
   3634   }
   3635 
   3636  private:
   3637   std::set<std::string> table_;
   3638   DISALLOW_COPY_AND_ASSIGN(StringTable);
   3639 };
   3640 
   3641 /*
   3642  * The data we send to DDMS contains everything we have recorded.
   3643  *
   3644  * Message header (all values big-endian):
   3645  * (1b) message header len (to allow future expansion); includes itself
   3646  * (1b) entry header len
   3647  * (1b) stack frame len
   3648  * (2b) number of entries
   3649  * (4b) offset to string table from start of message
   3650  * (2b) number of class name strings
   3651  * (2b) number of method name strings
   3652  * (2b) number of source file name strings
   3653  * For each entry:
   3654  *   (4b) total allocation size
   3655  *   (2b) thread id
   3656  *   (2b) allocated object's class name index
   3657  *   (1b) stack depth
   3658  *   For each stack frame:
   3659  *     (2b) method's class name
   3660  *     (2b) method name
   3661  *     (2b) method source file
   3662  *     (2b) line number, clipped to 32767; -2 if native; -1 if no source
   3663  * (xb) class name strings
   3664  * (xb) method name strings
   3665  * (xb) source file strings
   3666  *
   3667  * As with other DDM traffic, strings are sent as a 4-byte length
   3668  * followed by UTF-16 data.
   3669  *
   3670  * We send up 16-bit unsigned indexes into string tables.  In theory there
   3671  * can be (kMaxAllocRecordStackDepth * gAllocRecordMax) unique strings in
   3672  * each table, but in practice there should be far fewer.
   3673  *
   3674  * The chief reason for using a string table here is to keep the size of
   3675  * the DDMS message to a minimum.  This is partly to make the protocol
   3676  * efficient, but also because we have to form the whole thing up all at
   3677  * once in a memory buffer.
   3678  *
   3679  * We use separate string tables for class names, method names, and source
   3680  * files to keep the indexes small.  There will generally be no overlap
   3681  * between the contents of these tables.
   3682  */
   3683 jbyteArray Dbg::GetRecentAllocations() {
   3684   if (false) {
   3685     DumpRecentAllocations();
   3686   }
   3687 
   3688   Thread* self = Thread::Current();
   3689   std::vector<uint8_t> bytes;
   3690   {
   3691     MutexLock mu(self, gAllocTrackerLock);
   3692     //
   3693     // Part 1: generate string tables.
   3694     //
   3695     StringTable class_names;
   3696     StringTable method_names;
   3697     StringTable filenames;
   3698 
   3699     int count = gAllocRecordCount;
   3700     int idx = HeadIndex();
   3701     while (count--) {
   3702       AllocRecord* record = &recent_allocation_records_[idx];
   3703 
   3704       class_names.Add(ClassHelper(record->type).GetDescriptor());
   3705 
   3706       MethodHelper mh;
   3707       for (size_t i = 0; i < kMaxAllocRecordStackDepth; i++) {
   3708         mirror::ArtMethod* m = record->stack[i].method;
   3709         if (m != NULL) {
   3710           mh.ChangeMethod(m);
   3711           class_names.Add(mh.GetDeclaringClassDescriptor());
   3712           method_names.Add(mh.GetName());
   3713           filenames.Add(mh.GetDeclaringClassSourceFile());
   3714         }
   3715       }
   3716 
   3717       idx = (idx + 1) & (gAllocRecordMax-1);
   3718     }
   3719 
   3720     LOG(INFO) << "allocation records: " << gAllocRecordCount;
   3721 
   3722     //
   3723     // Part 2: Generate the output and store it in the buffer.
   3724     //
   3725 
   3726     // (1b) message header len (to allow future expansion); includes itself
   3727     // (1b) entry header len
   3728     // (1b) stack frame len
   3729     const int kMessageHeaderLen = 15;
   3730     const int kEntryHeaderLen = 9;
   3731     const int kStackFrameLen = 8;
   3732     JDWP::Append1BE(bytes, kMessageHeaderLen);
   3733     JDWP::Append1BE(bytes, kEntryHeaderLen);
   3734     JDWP::Append1BE(bytes, kStackFrameLen);
   3735 
   3736     // (2b) number of entries
   3737     // (4b) offset to string table from start of message
   3738     // (2b) number of class name strings
   3739     // (2b) number of method name strings
   3740     // (2b) number of source file name strings
   3741     JDWP::Append2BE(bytes, gAllocRecordCount);
   3742     size_t string_table_offset = bytes.size();
   3743     JDWP::Append4BE(bytes, 0);  // We'll patch this later...
   3744     JDWP::Append2BE(bytes, class_names.Size());
   3745     JDWP::Append2BE(bytes, method_names.Size());
   3746     JDWP::Append2BE(bytes, filenames.Size());
   3747 
   3748     count = gAllocRecordCount;
   3749     idx = HeadIndex();
   3750     ClassHelper kh;
   3751     while (count--) {
   3752       // For each entry:
   3753       // (4b) total allocation size
   3754       // (2b) thread id
   3755       // (2b) allocated object's class name index
   3756       // (1b) stack depth
   3757       AllocRecord* record = &recent_allocation_records_[idx];
   3758       size_t stack_depth = record->GetDepth();
   3759       kh.ChangeClass(record->type);
   3760       size_t allocated_object_class_name_index = class_names.IndexOf(kh.GetDescriptor());
   3761       JDWP::Append4BE(bytes, record->byte_count);
   3762       JDWP::Append2BE(bytes, record->thin_lock_id);
   3763       JDWP::Append2BE(bytes, allocated_object_class_name_index);
   3764       JDWP::Append1BE(bytes, stack_depth);
   3765 
   3766       MethodHelper mh;
   3767       for (size_t stack_frame = 0; stack_frame < stack_depth; ++stack_frame) {
   3768         // For each stack frame:
   3769         // (2b) method's class name
   3770         // (2b) method name
   3771         // (2b) method source file
   3772         // (2b) line number, clipped to 32767; -2 if native; -1 if no source
   3773         mh.ChangeMethod(record->stack[stack_frame].method);
   3774         size_t class_name_index = class_names.IndexOf(mh.GetDeclaringClassDescriptor());
   3775         size_t method_name_index = method_names.IndexOf(mh.GetName());
   3776         size_t file_name_index = filenames.IndexOf(mh.GetDeclaringClassSourceFile());
   3777         JDWP::Append2BE(bytes, class_name_index);
   3778         JDWP::Append2BE(bytes, method_name_index);
   3779         JDWP::Append2BE(bytes, file_name_index);
   3780         JDWP::Append2BE(bytes, record->stack[stack_frame].LineNumber());
   3781       }
   3782 
   3783       idx = (idx + 1) & (gAllocRecordMax-1);
   3784     }
   3785 
   3786     // (xb) class name strings
   3787     // (xb) method name strings
   3788     // (xb) source file strings
   3789     JDWP::Set4BE(&bytes[string_table_offset], bytes.size());
   3790     class_names.WriteTo(bytes);
   3791     method_names.WriteTo(bytes);
   3792     filenames.WriteTo(bytes);
   3793   }
   3794   JNIEnv* env = self->GetJniEnv();
   3795   jbyteArray result = env->NewByteArray(bytes.size());
   3796   if (result != NULL) {
   3797     env->SetByteArrayRegion(result, 0, bytes.size(), reinterpret_cast<const jbyte*>(&bytes[0]));
   3798   }
   3799   return result;
   3800 }
   3801 
   3802 }  // namespace art
   3803