Home | History | Annotate | Download | only in runtime

Lines Matching refs:tlsPtr_

84   tlsPtr_.card_table = Runtime::Current()->GetHeap()->GetCardTable()->GetBiasedBegin();
96 uintptr_t* begin = reinterpret_cast<uintptr_t*>(&tlsPtr_.interpreter_entrypoints);
98 sizeof(tlsPtr_.quick_entrypoints));
102 InitEntryPoints(&tlsPtr_.interpreter_entrypoints, &tlsPtr_.jni_entrypoints,
103 &tlsPtr_.portable_entrypoints, &tlsPtr_.quick_entrypoints);
107 ResetQuickAllocEntryPoints(&tlsPtr_.quick_entrypoints);
111 tlsPtr_.deoptimization_shadow_frame = sf;
119 ShadowFrame* sf = tlsPtr_.deoptimization_shadow_frame;
120 tlsPtr_.deoptimization_shadow_frame = nullptr;
126 sf->SetLink(tlsPtr_.shadow_frame_under_construction);
127 tlsPtr_.shadow_frame_under_construction = sf;
131 CHECK_NE(static_cast<ShadowFrame*>(nullptr), tlsPtr_.shadow_frame_under_construction);
132 tlsPtr_.shadow_frame_under_construction = tlsPtr_.shadow_frame_under_construction->GetLink();
166 CHECK(self->tlsPtr_.jpeer != nullptr);
167 self->tlsPtr_.opeer = soa.Decode<mirror::Object*>(self->tlsPtr_.jpeer);
168 self->GetJniEnv()->DeleteGlobalRef(self->tlsPtr_.jpeer);
169 self->tlsPtr_.jpeer = nullptr;
173 self->SetNativePriority(priorityField->GetInt(self->tlsPtr_.opeer));
177 mirror::Object* receiver = self->tlsPtr_.opeer;
255 byte* pregion = tlsPtr_.stack_begin - kStackOverflowProtectedSize;
256 byte* stack_himem = tlsPtr_.stack_end;
311 child_thread->tlsPtr_.jpeer = env->NewGlobalRef(java_peer);
334 env->DeleteGlobalRef(child_thread->tlsPtr_.jpeer);
335 child_thread->tlsPtr_.jpeer = nullptr;
363 tlsPtr_.pthread_self = pthread_self();
371 tlsPtr_.jni_env = new JNIEnvExt(this, java_vm);
408 self->tlsPtr_.name->assign(thread_name);
426 JNIEnv* env = tlsPtr_.jni_env;
446 tlsPtr_.opeer = soa.Decode<mirror::Object*>(peer.get());
484 SetBoolean<kTransactionActive>(tlsPtr_.opeer, thread_is_daemon);
486 SetObject<kTransactionActive>(tlsPtr_.opeer, soa.Decode<mirror::Object*>(thread_group));
488 SetObject<kTransactionActive>(tlsPtr_.opeer, soa.Decode<mirror::Object*>(thread_name));
490 SetInt<kTransactionActive>(tlsPtr_.opeer, thread_priority);
494 tlsPtr_.name->assign(name);
503 GetThreadStack(tlsPtr_.pthread_self, &read_stack_base, &read_stack_size, &read_guard_size);
511 tlsPtr_.stack_begin = reinterpret_cast<byte*>(read_stack_base);
512 tlsPtr_.stack_size = read_stack_size;
547 tlsPtr_.stack_size = default_stack_size;
548 tlsPtr_.stack_begin += (old_stack_size - default_stack_size);
551 << " with base " << reinterpret_cast<void*>(tlsPtr_.stack_begin);
568 tlsPtr_.stack_begin += read_guard_size + kStackOverflowProtectedSize;
569 tlsPtr_.stack_end += read_guard_size + kStackOverflowProtectedSize;
570 tlsPtr_.stack_size -= read_guard_size;
577 CHECK_GT(&stack_variable, reinterpret_cast<void*>(tlsPtr_.stack_end));
589 << ",peer=" << tlsPtr_.opeer
590 << ",\"" << *tlsPtr_.name << "\""
601 return (tlsPtr_.opeer != nullptr) ? reinterpret_cast<mirror::String*>(f->GetObject(tlsPtr_.opeer)) : nullptr;
605 name.assign(*tlsPtr_.name);
611 pthread_getcpuclockid(tlsPtr_.pthread_self, &cpu_clock_id);
680 checkpoints[i] = tlsPtr_.checkpoint_functions[i];
681 tlsPtr_.checkpoint_functions[i] = nullptr;
709 if (tlsPtr_.checkpoint_functions[i] == nullptr) {
718 tlsPtr_.checkpoint_functions[available_checkpoint] = function;
731 CHECK_EQ(tlsPtr_.checkpoint_functions[available_checkpoint], function);
732 tlsPtr_.checkpoint_functions[available_checkpoint] = nullptr;
759 if (gAborting == 0 && self != nullptr && thread != nullptr && thread->tlsPtr_.opeer != nullptr) {
762 ->GetInt(thread->tlsPtr_.opeer);
764 ->GetBoolean(thread->tlsPtr_.opeer);
767 soa.DecodeField(WellKnownClasses::java_lang_Thread_group)->GetObject(thread->tlsPtr_.opeer);
786 os << '"' << *thread->tlsPtr_.name << '"';
808 << " obj=" << reinterpret_cast<void*>(thread->tlsPtr_.opeer)
818 CHECK_PTHREAD_CALL(pthread_getschedparam, (thread->tlsPtr_.pthread_self, &policy, &sp),
821 << " handle=" << reinterpret_cast<void*>(thread->tlsPtr_.pthread_self);
846 os << " | stack=" << reinterpret_cast<void*>(thread->tlsPtr_.stack_begin) << "-"
847 << reinterpret_cast<void*>(thread->tlsPtr_
848 << PrettySize(thread->tlsPtr_.stack_size) << "\n";
1110 tlsPtr_.debug_invoke_req = new DebugInvokeReq;
1111 tlsPtr_.single_step_control = new SingleStepControl;
1112 tlsPtr_.instrumentation_stack = new std::deque<instrumentation::InstrumentationStackFrame>;
1113 tlsPtr_.name = new std::string(kThreadNameDuringStartup);
1114 tlsPtr_.nested_signal_state = static_cast<jmp_buf*>(malloc(sizeof(jmp_buf)));
1119 memset(&tlsPtr_.held_mutexes[0], 0, sizeof(tlsPtr_.held_mutexes));
1120 std::fill(tlsPtr_.rosalloc_runs,
1121 tlsPtr_.rosalloc_runs + kNumRosAllocThreadLocalSizeBrackets,
1124 tlsPtr_.checkpoint_functions[i] = nullptr;
1135 return (tlsPtr_.jpeer == nullptr && tlsPtr_.opeer == nullptr) ||
1136 (*tlsPtr_.name == kThreadNameDuringStartup);
1173 if (tlsPtr_.opeer != nullptr) {
1182 ->SetLong<true>(tlsPtr_.opeer, 0);
1185 ->SetLong<false>(tlsPtr_.opeer, 0);
1192 soa.DecodeField(WellKnownClasses::java_lang_Thread_lock)->GetObject(tlsPtr_.opeer);
1203 if (tlsPtr_.jni_env != nullptr) {
1204 tlsPtr_.jni_env->monitors.VisitRoots(MonitorExitVisitor, self, RootInfo(kRootVMInternal));
1209 if (tlsPtr_.jni_env != nullptr && tlsPtr_.jpeer != nullptr) {
1211 tlsPtr_.jni_env->DeleteGlobalRef(tlsPtr_.jpeer);
1212 tlsPtr_.jpeer = nullptr;
1214 tlsPtr_.opeer = nullptr;
1216 bool initialized = (tlsPtr_.jni_env != nullptr); // Did Thread::Init run?
1218 delete tlsPtr_.jni_env;
1219 tlsPtr_.jni_env = nullptr;
1223 CHECK(tlsPtr_.checkpoint_functions[0] == nullptr);
1224 CHECK(tlsPtr_.checkpoint_functions[1] == nullptr);
1225 CHECK(tlsPtr_.checkpoint_functions[2] == nullptr);
1233 if (tlsPtr_.long_jump_context != nullptr) {
1234 delete tlsPtr_.long_jump_context;
1241 delete tlsPtr_.debug_invoke_req;
1242 delete tlsPtr_.single_step_control;
1243 delete tlsPtr_.instrumentation_stack;
1244 delete tlsPtr_.name;
1245 delete tlsPtr_.stack_trace_sample;
1246 free(tlsPtr_.nested_signal_state);
1257 ScopedLocalRef<jobject> peer(tlsPtr_.jni_env, soa.AddLocalReference<jobject>(tlsPtr_.opeer));
1261 ScopedLocalRef<jthrowable> exception(tlsPtr_.jni_env, tlsPtr_.jni_env->ExceptionOccurred());
1262 tlsPtr_.jni_env->ExceptionClear();
1265 ScopedLocalRef<jobject> handler(tlsPtr_.jni_env,
1266 tlsPtr_.jni_env->GetObjectField(peer.get(),
1270 handler.reset(tlsPtr_.jni_env->GetObjectField(peer.get(),
1275 tlsPtr_
1280 tlsPtr_.jni_env->ExceptionClear();
1287 ->GetObject(tlsPtr_.opeer);
1290 ScopedLocalRef<jobject> peer(soa.Env(), soa.AddLocalReference<jobject>(tlsPtr_.opeer));
1292 tlsPtr_.jni_env->CallVoidMethod(group.get(),
1300 for (HandleScope* cur = tlsPtr_.top_handle_scope; cur; cur = cur->GetLink()) {
1309 for (HandleScope* cur = tlsPtr_.top_handle_scope; cur; cur = cur->GetLink()) {
1315 return tlsPtr_.managed_stack.ShadowFramesContain(hs_entry);
1319 for (HandleScope* cur = tlsPtr_.top_handle_scope; cur; cur = cur->GetLink()) {
1344 IndirectReferenceTable& locals = tlsPtr_.jni_env->locals;
1973 Context* result = tlsPtr_.long_jump_context;
1977 tlsPtr_.long_jump_context = nullptr; // Avoid context being shared.
2186 tlsPtr_.class_loader_override = class_loader_override;
2191 if (tlsPtr_.opeer != nullptr) {
2192 visitor(&tlsPtr_.opeer, arg, RootInfo(kRootThreadObject, thread_id));
2194 if (tlsPtr_.exception != nullptr && tlsPtr_.exception != GetDeoptimizationException()) {
2195 visitor(reinterpret_cast<mirror::Object**>(&tlsPtr_.exception), arg,
2198 tlsPtr_.throw_location.VisitRoots(visitor, arg);
2199 if (tlsPtr_.class_loader_override != nullptr) {
2200 visitor(reinterpret_cast<mirror::Object**>(&tlsPtr_.class_loader_override), arg,
2203 if (tlsPtr_.monitor_enter_object != nullptr) {
2204 visitor(&tlsPtr_.monitor_enter_object, arg, RootInfo(kRootNativeStack, thread_id));
2206 tlsPtr_.jni_env->locals.VisitRoots(visitor, arg, RootInfo(kRootJNILocal, thread_id));
2207 tlsPtr_.jni_env->monitors.VisitRoots(visitor, arg, RootInfo(kRootJNIMonitor, thread_id));
2209 if (tlsPtr_.debug_invoke_req != nullptr) {
2210 tlsPtr_.debug_invoke_req->VisitRoots(visitor, arg, RootInfo(kRootDebugger, thread_id));
2212 if (tlsPtr_.single_step_control != nullptr) {
2213 tlsPtr_.single_step_control->VisitRoots(visitor, arg, RootInfo(kRootDebugger, thread_id));
2215 if (tlsPtr_.deoptimization_shadow_frame != nullptr) {
2218 for (ShadowFrame* shadow_frame = tlsPtr_.deoptimization_shadow_frame; shadow_frame != nullptr;
2223 if (tlsPtr_.shadow_frame_under_construction != nullptr) {
2226 for (ShadowFrame* shadow_frame = tlsPtr_.shadow_frame_under_construction;
2263 if (tlsPtr_.stack_end == tlsPtr_.stack_begin) {
2271 tlsPtr_.stack_end = tlsPtr_.stack_begin;
2284 tlsPtr_.thread_local_start = start;
2285 tlsPtr_.thread_local_pos = tlsPtr_.thread_local_start;
2286 tlsPtr_.thread_local_end = end;
2287 tlsPtr_.thread_local_objects = 0;
2291 bool has_tlab = tlsPtr_.thread_local_pos != nullptr;
2293 DCHECK(tlsPtr_.thread_local_start != nullptr && tlsPtr_.thread_local_end != nullptr);
2295 DCHECK(tlsPtr_.thread_local_start == nullptr && tlsPtr_.thread_local_end == nullptr);
2306 void* pregion = tlsPtr_.stack_begin - kStackOverflowProtectedSize;
2316 void* pregion = tlsPtr_.stack_begin - kStackOverflowProtectedSize;