HomeSort by relevance Sort by last modified time
    Searched refs:ALWAYS_INLINE (Results 51 - 75 of 95) sorted by null

1 23 4

  /art/runtime/verifier/
register_line.h 102 ALWAYS_INLINE bool SetRegisterType(MethodVerifier* verifier,
123 ALWAYS_INLINE bool VerifyRegisterType(MethodVerifier* verifier,
206 ALWAYS_INLINE static size_t ComputeSize(size_t num_regs);
method_verifier.h 247 ALWAYS_INLINE const InstructionFlags& GetInstructionFlags(size_t index) const;
248 ALWAYS_INLINE InstructionFlags& GetInstructionFlags(size_t index);
    [all...]
  /art/runtime/base/
mutex.h 327 void SharedLock(Thread* self) ACQUIRE_SHARED() ALWAYS_INLINE;
334 void SharedUnlock(Thread* self) RELEASE_SHARED() ALWAYS_INLINE;
436 void TransitionFromRunnableToSuspended(Thread* self) UNLOCK_FUNCTION() ALWAYS_INLINE;
437 void TransitionFromSuspendedToRunnable(Thread* self) SHARED_LOCK_FUNCTION() ALWAYS_INLINE;
arena_allocator.h 303 void* Alloc(size_t bytes, ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE {
319 ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE {
scoped_arena_containers.h 207 ALWAYS_INLINE void ProtectMemory(T* ptr, size_t size) const {
  /art/runtime/
type_lookup_table.h 44 ALWAYS_INLINE uint32_t Lookup(const char* str, uint32_t hash) const {
indirect_reference_table.h 290 ALWAYS_INLINE;
monitor.h 263 ALWAYS_INLINE static void AtraceMonitorLock(Thread* self,
271 ALWAYS_INLINE static void AtraceMonitorUnlock();
art_field.h 202 ALWAYS_INLINE void UpdateObjects(const Visitor& visitor)
  /external/compiler-rt/lib/sanitizer_common/
sanitizer_stacktrace.h 73 ALWAYS_INLINE
sanitizer_internal_defs.h 127 # define ALWAYS_INLINE __forceinline
139 # define ALWAYS_INLINE inline __attribute__((always_inline))
  /art/runtime/gc/accounting/
space_bitmap.h 76 bool Set(const mirror::Object* obj) ALWAYS_INLINE {
80 bool Clear(const mirror::Object* obj) ALWAYS_INLINE {
atomic_stack.h 222 bool AtomicPushBackInternal(T* value, size_t limit) ALWAYS_INLINE
  /art/runtime/mirror/
object.cc 50 ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) {
59 ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) {
  /art/runtime/native/
java_lang_reflect_Field.cc 33 ALWAYS_INLINE inline static bool VerifyFieldAccess(Thread* self, mirror::Field* field,
61 ALWAYS_INLINE inline static bool GetFieldValue(mirror::Object* o, mirror::Field* f,
105 ALWAYS_INLINE inline static bool CheckReceiver(const ScopedFastNativeObjectAccess& soa,
158 ALWAYS_INLINE inline static JValue GetPrimitiveField(JNIEnv* env, jobject javaField,
232 ALWAYS_INLINE inline static void SetFieldValue(mirror::Object* o, mirror::Field* f,
java_lang_Class.cc 44 ALWAYS_INLINE static inline mirror::Class* DecodeClass(
188 ALWAYS_INLINE static inline ArtField* FindFieldByName(
220 ALWAYS_INLINE static inline mirror::Field* GetDeclaredField(
333 static ALWAYS_INLINE inline bool MethodMatchesConstructor(ArtMethod* m, bool public_only)
  /art/runtime/gc/
heap.h 216 ALWAYS_INLINE mirror::Object* AllocObjectWithAllocator(Thread* self,
438 ALWAYS_INLINE void WriteBarrierField(const mirror::Object* dst,
445 ALWAYS_INLINE void WriteBarrierArray(const mirror::Object* dst,
452 ALWAYS_INLINE void WriteBarrierEveryFieldOf(const mirror::Object* obj) {
812 static ALWAYS_INLINE bool AllocatorHasAllocationStack(AllocatorType allocator_type) {
819 static ALWAYS_INLINE bool AllocatorMayHaveConcurrentGC(AllocatorType allocator_type) {
    [all...]
  /art/runtime/gc/space/
image_space.cc 667 ALWAYS_INLINE T* ForwardObject(T* src) const {
682 ALWAYS_INLINE const void* ForwardCode(const void* src) const {
695 ALWAYS_INLINE bool IsInAppImage(mirror::Object* object) const {
724 ALWAYS_INLINE void VisitRootIfNonNull(mirror::CompressedReference<mirror::Object>* root) const
731 ALWAYS_INLINE void VisitRoot(mirror::CompressedReference<mirror::Object>* root) const
752 ALWAYS_INLINE void VisitRootIfNonNull(
755 ALWAYS_INLINE void VisitRoot(mirror::CompressedReference<mirror::Object>* root ATTRIBUTE_UNUSED)
758 ALWAYS_INLINE void operator()(mirror::Object* obj,
    [all...]
region_space.h 53 ALWAYS_INLINE mirror::Object* AllocNonvirtual(size_t num_bytes, size_t* bytes_allocated,
275 ALWAYS_INLINE mirror::Object* Alloc(size_t num_bytes, size_t* bytes_allocated,
375 ALWAYS_INLINE bool ShouldBeEvacuated();
  /art/runtime/gc/collector/
mark_compact.cc 326 REQUIRES(Locks::mutator_lock_) ALWAYS_INLINE {
417 ALWAYS_INLINE REQUIRES(Locks::mutator_lock_, Locks::heap_bitmap_lock_) {
513 REQUIRES(Locks::mutator_lock_) ALWAYS_INLINE {
554 void operator()(mirror::Object* obj, MemberOffset offset, bool /*is_static*/) const ALWAYS_INLINE
mark_sweep.cc 271 explicit ScanObjectVisitor(MarkSweep* const mark_sweep) ALWAYS_INLINE
275 ALWAYS_INLINE
674 ALWAYS_INLINE MarkObjectParallelVisitor(MarkStackTask<kUseFinger>* chunk_task,
678 ALWAYS_INLINE void operator()(mirror::Object* obj,
702 ALWAYS_INLINE void Mark(mirror::Object* ref) const SHARED_REQUIRES(Locks::mutator_lock_) {
721 ALWAYS_INLINE explicit ScanObjectParallelVisitor(MarkStackTask<kUseFinger>* chunk_task)
753 ALWAYS_INLINE void MarkStackPush(mirror::Object* obj)
    [all...]
concurrent_copying.cc 665 SHARED_REQUIRES(Locks::mutator_lock_) ALWAYS_INLINE {
705 SHARED_REQUIRES(Locks::mutator_lock_) ALWAYS_INLINE {
712 SHARED_REQUIRES(Locks::mutator_lock_) ALWAYS_INLINE {
    [all...]
  /external/compiler-rt/lib/asan/
asan_mac.cc 110 ALWAYS_INLINE
asan_globals.cc 59 ALWAYS_INLINE void PoisonShadowForGlobal(const Global *g, u8 value) {
63 ALWAYS_INLINE void PoisonRedZones(const Global &g) {
  /art/runtime/gc/allocator/
rosalloc.h 421 ALWAYS_INLINE void* AllocSlot();
437 ALWAYS_INLINE bool IsFull();
630 ALWAYS_INLINE bool ShouldCheckZeroMemory();
    [all...]

Completed in 395 milliseconds

1 23 4