Home | History | Annotate | Download | only in jit
      1 /*
      2  * Copyright (C) 2015 The Android Open Source Project
      3  *
      4  * Licensed under the Apache License, Version 2.0 (the "License");
      5  * you may not use this file except in compliance with the License.
      6  * You may obtain a copy of the License at
      7  *
      8  *      http://www.apache.org/licenses/LICENSE-2.0
      9  *
     10  * Unless required by applicable law or agreed to in writing, software
     11  * distributed under the License is distributed on an "AS IS" BASIS,
     12  * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
     13  * See the License for the specific language governing permissions and
     14  * limitations under the License.
     15  */
     16 
     17 #ifndef ART_RUNTIME_JIT_PROFILING_INFO_H_
     18 #define ART_RUNTIME_JIT_PROFILING_INFO_H_
     19 
     20 #include <vector>
     21 
     22 #include "base/macros.h"
     23 #include "gc_root.h"
     24 
     25 namespace art {
     26 
     27 class ArtMethod;
     28 class ProfilingInfo;
     29 
     30 namespace jit {
     31 class JitCodeCache;
     32 }
     33 
     34 namespace mirror {
     35 class Class;
     36 }
     37 
     38 // Structure to store the classes seen at runtime for a specific instruction.
     39 // Once the classes_ array is full, we consider the INVOKE to be megamorphic.
     40 class InlineCache {
     41  public:
     42   bool IsMonomorphic() const {
     43     DCHECK_GE(kIndividualCacheSize, 2);
     44     return !classes_[0].IsNull() && classes_[1].IsNull();
     45   }
     46 
     47   bool IsMegamorphic() const {
     48     for (size_t i = 0; i < kIndividualCacheSize; ++i) {
     49       if (classes_[i].IsNull()) {
     50         return false;
     51       }
     52     }
     53     return true;
     54   }
     55 
     56   mirror::Class* GetMonomorphicType() const SHARED_REQUIRES(Locks::mutator_lock_) {
     57     // Note that we cannot ensure the inline cache is actually monomorphic
     58     // at this point, as other threads may have updated it.
     59     DCHECK(!classes_[0].IsNull());
     60     return classes_[0].Read();
     61   }
     62 
     63   bool IsUninitialized() const {
     64     return classes_[0].IsNull();
     65   }
     66 
     67   bool IsPolymorphic() const {
     68     DCHECK_GE(kIndividualCacheSize, 3);
     69     return !classes_[1].IsNull() && classes_[kIndividualCacheSize - 1].IsNull();
     70   }
     71 
     72   mirror::Class* GetTypeAt(size_t i) const SHARED_REQUIRES(Locks::mutator_lock_) {
     73     return classes_[i].Read();
     74   }
     75 
     76   static constexpr uint16_t kIndividualCacheSize = 5;
     77 
     78  private:
     79   uint32_t dex_pc_;
     80   GcRoot<mirror::Class> classes_[kIndividualCacheSize];
     81 
     82   friend class ProfilingInfo;
     83 
     84   DISALLOW_COPY_AND_ASSIGN(InlineCache);
     85 };
     86 
     87 /**
     88  * Profiling info for a method, created and filled by the interpreter once the
     89  * method is warm, and used by the compiler to drive optimizations.
     90  */
     91 class ProfilingInfo {
     92  public:
     93   // Create a ProfilingInfo for 'method'. Return whether it succeeded, or if it is
     94   // not needed in case the method does not have virtual/interface invocations.
     95   static bool Create(Thread* self, ArtMethod* method, bool retry_allocation)
     96       SHARED_REQUIRES(Locks::mutator_lock_);
     97 
     98   // Add information from an executed INVOKE instruction to the profile.
     99   void AddInvokeInfo(uint32_t dex_pc, mirror::Class* cls)
    100       // Method should not be interruptible, as it manipulates the ProfilingInfo
    101       // which can be concurrently collected.
    102       REQUIRES(Roles::uninterruptible_)
    103       SHARED_REQUIRES(Locks::mutator_lock_);
    104 
    105   // NO_THREAD_SAFETY_ANALYSIS since we don't know what the callback requires.
    106   template<typename RootVisitorType>
    107   void VisitRoots(RootVisitorType& visitor) NO_THREAD_SAFETY_ANALYSIS {
    108     for (size_t i = 0; i < number_of_inline_caches_; ++i) {
    109       InlineCache* cache = &cache_[i];
    110       for (size_t j = 0; j < InlineCache::kIndividualCacheSize; ++j) {
    111         visitor.VisitRootIfNonNull(cache->classes_[j].AddressWithoutBarrier());
    112       }
    113     }
    114   }
    115 
    116   ArtMethod* GetMethod() const {
    117     return method_;
    118   }
    119 
    120   InlineCache* GetInlineCache(uint32_t dex_pc);
    121 
    122   bool IsMethodBeingCompiled(bool osr) const {
    123     return osr
    124         ? is_osr_method_being_compiled_
    125         : is_method_being_compiled_;
    126   }
    127 
    128   void SetIsMethodBeingCompiled(bool value, bool osr) {
    129     if (osr) {
    130       is_osr_method_being_compiled_ = value;
    131     } else {
    132       is_method_being_compiled_ = value;
    133     }
    134   }
    135 
    136   void SetSavedEntryPoint(const void* entry_point) {
    137     saved_entry_point_ = entry_point;
    138   }
    139 
    140   const void* GetSavedEntryPoint() const {
    141     return saved_entry_point_;
    142   }
    143 
    144   void ClearGcRootsInInlineCaches() {
    145     for (size_t i = 0; i < number_of_inline_caches_; ++i) {
    146       InlineCache* cache = &cache_[i];
    147       memset(&cache->classes_[0],
    148              0,
    149              InlineCache::kIndividualCacheSize * sizeof(GcRoot<mirror::Class>));
    150     }
    151   }
    152 
    153   void IncrementInlineUse() {
    154     DCHECK_NE(current_inline_uses_, std::numeric_limits<uint16_t>::max());
    155     current_inline_uses_++;
    156   }
    157 
    158   void DecrementInlineUse() {
    159     DCHECK_GT(current_inline_uses_, 0);
    160     current_inline_uses_--;
    161   }
    162 
    163   bool IsInUseByCompiler() const {
    164     return IsMethodBeingCompiled(/*osr*/ true) || IsMethodBeingCompiled(/*osr*/ false) ||
    165         (current_inline_uses_ > 0);
    166   }
    167 
    168  private:
    169   ProfilingInfo(ArtMethod* method, const std::vector<uint32_t>& entries)
    170       : number_of_inline_caches_(entries.size()),
    171         method_(method),
    172         is_method_being_compiled_(false),
    173         is_osr_method_being_compiled_(false),
    174         current_inline_uses_(0),
    175         saved_entry_point_(nullptr) {
    176     memset(&cache_, 0, number_of_inline_caches_ * sizeof(InlineCache));
    177     for (size_t i = 0; i < number_of_inline_caches_; ++i) {
    178       cache_[i].dex_pc_ = entries[i];
    179     }
    180   }
    181 
    182   // Number of instructions we are profiling in the ArtMethod.
    183   const uint32_t number_of_inline_caches_;
    184 
    185   // Method this profiling info is for.
    186   ArtMethod* const method_;
    187 
    188   // Whether the ArtMethod is currently being compiled. This flag
    189   // is implicitly guarded by the JIT code cache lock.
    190   // TODO: Make the JIT code cache lock global.
    191   bool is_method_being_compiled_;
    192   bool is_osr_method_being_compiled_;
    193 
    194   // When the compiler inlines the method associated to this ProfilingInfo,
    195   // it updates this counter so that the GC does not try to clear the inline caches.
    196   uint16_t current_inline_uses_;
    197 
    198   // Entry point of the corresponding ArtMethod, while the JIT code cache
    199   // is poking for the liveness of compiled code.
    200   const void* saved_entry_point_;
    201 
    202   // Dynamically allocated array of size `number_of_inline_caches_`.
    203   InlineCache cache_[0];
    204 
    205   friend class jit::JitCodeCache;
    206 
    207   DISALLOW_COPY_AND_ASSIGN(ProfilingInfo);
    208 };
    209 
    210 }  // namespace art
    211 
    212 #endif  // ART_RUNTIME_JIT_PROFILING_INFO_H_
    213