HomeSort by relevance Sort by last modified time
    Searched refs:GcRoot (Results 1 - 25 of 71) sorted by null

1 2 3

  /art/runtime/
gc_root-inl.h 31 inline MirrorType* GcRoot<MirrorType>::Read(GcRootSource* gc_root_source) const {
37 inline GcRoot<MirrorType>::GcRoot(MirrorType* ref)
41 inline GcRoot<MirrorType>::GcRoot(ObjPtr<MirrorType> ref)
42 : GcRoot(ref.Ptr()) { }
class_table-inl.h 37 for (GcRoot<mirror::Object>& root : strong_roots_) {
41 for (GcRoot<mirror::Object>& root : oat_file->GetBssGcRoots()) {
55 for (GcRoot<mirror::Object>& root : strong_roots_) {
59 for (GcRoot<mirror::Object>& root : oat_file->GetBssGcRoots()) {
96 GcRoot<mirror::Class>(before_ptr).Read<kReadBarrierOption>());
109 GcRoot<mirror::Class> root(before_ptr);
gc_root.h 186 class GcRoot {
216 ALWAYS_INLINE GcRoot() {}
217 explicit ALWAYS_INLINE GcRoot(MirrorType* ref)
219 explicit ALWAYS_INLINE GcRoot(ObjPtr<MirrorType> ref)
246 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root)
262 void VisitRoot(GcRoot<MirrorType>& root) REQUIRES_SHARED(Locks::mutator_lock_) {
293 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root) const
309 void VisitRoot(GcRoot<MirrorType>& root) const REQUIRES_SHARED(Locks::mutator_lock_) {
reference_table.h 58 typedef std::vector<GcRoot<mirror::Object>,
59 TrackingAllocator<GcRoot<mirror::Object>, kAllocatorTagReferenceTable>> Table;
intern_table.h 80 std::size_t operator()(const GcRoot<mirror::String>& root) const NO_THREAD_SAFETY_ANALYSIS;
81 bool operator()(const GcRoot<mirror::String>& a, const GcRoot<mirror::String>& b) const
90 bool operator()(const GcRoot<mirror::String>& a, const Utf8String& b) const
96 void MakeEmpty(GcRoot<mirror::String>& item) const {
97 item = GcRoot<mirror::String>();
99 bool IsEmpty(const GcRoot<mirror::String>& item) const {
104 using UnorderedSet = HashSet<GcRoot<mirror::String>,
108 TrackingAllocator<GcRoot<mirror::String>, kAllocatorTagInternTable>>;
329 std::vector<GcRoot<mirror::String>> new_strong_intern_roots
    [all...]
reference_table.cc 56 entries_.push_back(GcRoot<mirror::Object>(obj));
128 bool operator()(GcRoot<mirror::Object> root1, GcRoot<mirror::Object> root2) const
222 GcRoot<mirror::Object>& stack_for_object = it->first;
263 for (GcRoot<mirror::Object>& root : entries) {
275 GcRoot<mirror::Object> root;
288 void Reset(GcRoot<mirror::Object>& _root) {
298 for (GcRoot<mirror::Object>& root : sorted_entries) {
358 for (GcRoot<mirror::Object>& root : entries_) {
indirect_reference_table-inl.h 114 references_[serial_] = GcRoot<mirror::Object>(obj);
119 references_[serial_] = GcRoot<mirror::Object>(obj);
indirect_reference_table.h 161 GcRoot<mirror::Object>* GetReference() {
166 const GcRoot<mirror::Object>* GetReference() const {
179 GcRoot<mirror::Object> references_[kIRTPrevCount];
198 GcRoot<mirror::Object>* operator*() REQUIRES_SHARED(Locks::mutator_lock_) {
intern_table.cc 133 new_strong_intern_roots_.push_back(GcRoot<mirror::String>(s));
300 std::size_t InternTable::StringHashEquals::operator()(const GcRoot<mirror::String>& root) const {
309 bool InternTable::StringHashEquals::operator()(const GcRoot<mirror::String>& a,
310 const GcRoot<mirror::String>& b) const {
317 bool InternTable::StringHashEquals::operator()(const GcRoot<mirror::String>& a,
355 for (GcRoot<mirror::String>& string : table.set_) {
367 auto it = table.set_.find(GcRoot<mirror::String>(s));
379 auto it = table.set_.find(GcRoot<mirror::String>(s));
406 tables_.back().set_.insert(GcRoot<mirror::String>(s));
433 *it = GcRoot<mirror::String>(new_object->AsString())
    [all...]
art_method-inl.h 70 declaring_class_ = GcRoot<mirror::Class>(new_declaring_class);
75 GcRoot<mirror::Class> expected_root(expected_class);
76 GcRoot<mirror::Class> desired_root(desired_class);
77 auto atomic_root_class = reinterpret_cast<Atomic<GcRoot<mirror::Class>>*>(&declaring_class_);
intern_table-inl.h 64 for (GcRoot<mirror::String>& string : intern_strings) {
class_table_test.cc 39 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root) const
55 void VisitRoot(GcRoot<MirrorType>& root) const REQUIRES_SHARED(Locks::mutator_lock_) {
  /art/runtime/entrypoints/quick/
quick_dexcache_entrypoints.cc 40 static_assert(sizeof(GcRoot<mirror::Class>) == sizeof(GcRoot<mirror::Object>), "Size check.");
41 static_assert(sizeof(GcRoot<mirror::String>) == sizeof(GcRoot<mirror::Object>), "Size check.");
43 DCHECK_ALIGNED(bss_offset, sizeof(GcRoot<mirror::Object>));
50 GcRoot<mirror::Object>* slot = reinterpret_cast<GcRoot<mirror::Object>*>(
56 *slot = GcRoot<mirror::Object>(object);
88 sizeof(GcRoot<mirror::Class>));
106 sizeof(GcRoot<mirror::Class>))
    [all...]
quick_entrypoints.h 40 template<class MirrorType> class GcRoot;
107 extern "C" mirror::Object* artReadBarrierForRootSlow(GcRoot<mirror::Object>* root)
  /art/runtime/jit/
profiling_info.cc 95 GcRoot<mirror::Class> expected_root(existing);
96 GcRoot<mirror::Class> desired_root(cls);
97 auto atomic_root = reinterpret_cast<Atomic<GcRoot<mirror::Class>>*>(&cache->classes_[i]);
profiling_info.h 46 GcRoot<mirror::Class> classes_[kIndividualCacheSize];
107 InlineCache::kIndividualCacheSize * sizeof(GcRoot<mirror::Class>));
  /art/runtime/mirror/
dex_cache-inl.h 51 first_elem.object = GcRoot<T>(nullptr);
89 GcRoot<mirror::String>* preresolved_strings = GetPreResolvedStrings();
121 GetPreResolvedStrings()[string_idx.index_] = GcRoot<mirror::String>(resolved);
133 GcRoot<mirror::String>*>(PreResolvedStringsOffset(), nullptr);
212 GcRoot<mirror::CallSite>& target = GetResolvedCallSites()[call_site_idx];
213 Atomic<GcRoot<mirror::CallSite>>& ref =
214 reinterpret_cast<Atomic<GcRoot<mirror::CallSite>>&>(target);
223 GcRoot<mirror::CallSite> null_call_site(nullptr);
224 GcRoot<mirror::CallSite> candidate(call_site);
225 GcRoot<mirror::CallSite>& target = GetResolvedCallSites()[call_site_idx]
    [all...]
dex_cache.h 50 GcRoot<T> object;
209 void FixupResolvedCallSites(GcRoot<mirror::CallSite>* dest, const Visitor& visitor)
271 return alignof(GcRoot<mirror::String>);
340 GcRoot<mirror::String>* GetPreResolvedStrings() ALWAYS_INLINE
342 return GetFieldPtr64<GcRoot<mirror::String>*, kVerifyFlags>(PreResolvedStringsOffset());
349 void SetPreResolvedStrings(GcRoot<mirror::String>* strings)
398 GcRoot<CallSite>* GetResolvedCallSites()
401 return GetFieldPtr<GcRoot<CallSite>*, kVerifyFlags>(ResolvedCallSitesOffset());
404 void SetResolvedCallSites(GcRoot<CallSite>* resolved_call_sites)
467 return sizeof(GcRoot<mirror::String>) * num_strings
    [all...]
dex_cache.cc 106 GcRoot<mirror::CallSite>* call_sites = (dex_file->NumCallSiteIds() == 0)
108 : reinterpret_cast<GcRoot<CallSite>*>(raw_arrays + layout.CallSitesOffset());
180 GcRoot<mirror::String>* strings =
181 linear_alloc->AllocArray<GcRoot<mirror::String>>(self, num_strings);
208 GcRoot<CallSite>* resolved_call_sites,
  /art/runtime/gc/
system_weak_test.cc 79 weak_ = GcRoot<mirror::Object>(new_object);
84 GcRoot<mirror::Object> Get()
94 void Set(GcRoot<mirror::Object> obj)
107 GcRoot<mirror::Object> weak_ GUARDED_BY(allow_disallow_lock_);
145 cswh.Set(GcRoot<mirror::Object>(s.Get()));
166 cswh.Set(GcRoot<mirror::Object>(mirror::String::AllocFromModifiedUtf8(soa.Self(), "ABC")));
190 cswh.Set(GcRoot<mirror::Object>(s.Get()));
allocation_record.cc 78 GcRoot<mirror::Class>& klass = record->GetClassGcRoot();
87 klass = GcRoot<mirror::Class>(new_object->AsClass());
105 it->first = GcRoot<mirror::Object>(nullptr);
114 it->first = GcRoot<mirror::Object>(new_object);
allocation_record.h 185 GcRoot<mirror::Class>& GetClassGcRoot() REQUIRES_SHARED(Locks::mutator_lock_) {
196 GcRoot<mirror::Class> klass_;
208 // GcRoot<mirror::Object> pointers in the list are weak roots, and the last recent_record_max_
211 // recent allocation tracking, but GcRoot<mirror::Object> pointers in these pairs can become null.
213 using EntryPair = std::pair<GcRoot<mirror::Object>, AllocRecord>;
235 entries_.push_back(EntryPair(GcRoot<mirror::Object>(obj), std::move(record)));
  /art/openjdkjvmti/
jvmti_weak_table.h 147 auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj));
199 size_t operator()(const art::GcRoot<art::mirror::Object>& r) const
206 bool operator()(const art::GcRoot<art::mirror::Object>& r1,
207 const art::GcRoot<art::mirror::Object>& r2) const
213 using TagAllocator = JvmtiAllocator<std::pair<const art::GcRoot<art::mirror::Object>, T>>;
214 std::unordered_map<art::GcRoot<art::mirror::Object>,
jvmti_weak_table-inl.h 70 auto WithReadBarrierUpdater = [&](const art::GcRoot<art::mirror::Object>& original_root,
108 auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj));
153 auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj));
173 auto insert_it = tagged_objects_.emplace(art::GcRoot<art::mirror::Object>(obj), new_tag);
201 auto IsMarkedUpdater = [&](const art::GcRoot<art::mirror::Object>& original_root ATTRIBUTE_UNUSED,
232 tagged_objects_.emplace(art::GcRoot<art::mirror::Object>(target_obj), tag);
  /art/runtime/utils/
dex_cache_arrays_layout-inl.h 74 static_assert(sizeof(GcRoot<T>) == 4U, "Unexpected GcRoot size");
92 return alignof(GcRoot<mirror::Class>);
167 return alignof(GcRoot<mirror::CallSite>);

Completed in 377 milliseconds

1 2 3