HomeSort by relevance Sort by last modified time
    Searched refs:ALWAYS_INLINE (Results 1 - 25 of 95) sorted by null

1 2 3 4

  /art/runtime/gc/accounting/
bitmap.h 49 static ALWAYS_INLINE constexpr size_t BitIndexToWordIndex(uintptr_t offset) {
54 static ALWAYS_INLINE constexpr T WordIndexToBitIndex(T word_index) {
58 static ALWAYS_INLINE constexpr uintptr_t BitIndexToMask(uintptr_t bit_index) {
62 ALWAYS_INLINE bool SetBit(size_t bit_index) {
66 ALWAYS_INLINE bool ClearBit(size_t bit_index) {
70 ALWAYS_INLINE bool TestBit(size_t bit_index) const;
73 ALWAYS_INLINE bool AtomicTestAndSetBit(size_t bit_index);
96 ALWAYS_INLINE void CheckValidBitIndex(size_t bit_index) const {
112 ALWAYS_INLINE bool ModifyBit(uintptr_t bit_index);
137 ALWAYS_INLINE uintptr_t CoverBegin() const
    [all...]
mod_union_table-inl.h 36 bool ShouldAddReference(const mirror::Object* ref) const OVERRIDE ALWAYS_INLINE {
card_table.h 59 ALWAYS_INLINE void MarkCard(const void *addr) {
126 void* AddrFromCard(const uint8_t *card_addr) const ALWAYS_INLINE;
129 uint8_t* CardFromAddr(const void *addr) const ALWAYS_INLINE;
137 bool IsValidCard(const uint8_t* card_addr) const ALWAYS_INLINE;
139 void CheckCardValid(uint8_t* card) const ALWAYS_INLINE;
  /art/runtime/native/
scoped_fast_native_object_access.h 31 SHARED_LOCK_FUNCTION(Locks::mutator_lock_) ALWAYS_INLINE
39 ~ScopedFastNativeObjectAccess() UNLOCK_FUNCTION(Locks::mutator_lock_) ALWAYS_INLINE {
  /art/runtime/
handle.h 42 ALWAYS_INLINE Handle(const Handle<T>& handle) : reference_(handle.reference_) {
45 ALWAYS_INLINE Handle<T>& operator=(const Handle<T>& handle) {
50 ALWAYS_INLINE explicit Handle(StackReference<T>* reference) : reference_(reference) {
53 ALWAYS_INLINE T& operator*() const SHARED_REQUIRES(Locks::mutator_lock_) {
57 ALWAYS_INLINE T* operator->() const SHARED_REQUIRES(Locks::mutator_lock_) {
61 ALWAYS_INLINE T* Get() const SHARED_REQUIRES(Locks::mutator_lock_) {
65 ALWAYS_INLINE jobject ToJObject() const SHARED_REQUIRES(Locks::mutator_lock_) {
73 ALWAYS_INLINE StackReference<mirror::Object>* GetReference() {
77 ALWAYS_INLINE const StackReference<mirror::Object>* GetReference() const {
108 ALWAYS_INLINE MutableHandle(const MutableHandle<T>& handle
    [all...]
handle_scope.h 62 ALWAYS_INLINE mirror::Object* GetReference(size_t i) const
65 ALWAYS_INLINE Handle<mirror::Object> GetHandle(size_t i);
67 ALWAYS_INLINE MutableHandle<mirror::Object> GetMutableHandle(size_t i)
70 ALWAYS_INLINE void SetReference(size_t i, mirror::Object* object)
73 ALWAYS_INLINE bool Contains(StackReference<mirror::Object>* handle_scope_entry) const;
98 ALWAYS_INLINE StackReference<mirror::Object>* GetReferences() const {
148 explicit ALWAYS_INLINE StackHandleScope(Thread* self, mirror::Object* fill_value = nullptr);
149 ALWAYS_INLINE ~StackHandleScope();
152 ALWAYS_INLINE MutableHandle<T> NewHandle(T* object) SHARED_REQUIRES(Locks::mutator_lock_);
155 ALWAYS_INLINE HandleWrapper<T> NewHandleWrapper(T** object
    [all...]
verify_object.h 55 ALWAYS_INLINE void VerifyObject(mirror::Object* obj) NO_THREAD_SAFETY_ANALYSIS;
58 ALWAYS_INLINE bool VerifyClassClass(mirror::Class* c) NO_THREAD_SAFETY_ANALYSIS;
memory_region.h 55 ALWAYS_INLINE T Load(uintptr_t offset) const {
65 ALWAYS_INLINE void Store(uintptr_t offset, T value) const {
74 ALWAYS_INLINE T LoadUnaligned(uintptr_t offset) const {
89 ALWAYS_INLINE void StoreUnaligned(uintptr_t offset, T value) const {
101 ALWAYS_INLINE T* PointerTo(uintptr_t offset) const {
107 ALWAYS_INLINE bool LoadBit(uintptr_t bit_offset) const {
113 ALWAYS_INLINE void StoreBit(uintptr_t bit_offset, bool value) const {
165 ALWAYS_INLINE T* ComputeInternalPointer(size_t offset) const {
173 ALWAYS_INLINE uint8_t* ComputeBitPointer(uintptr_t bit_offset, uint8_t* bit_mask) const {
scoped_thread_state_change.h 37 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE
62 ~ScopedThreadStateChange() REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE {
179 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE
184 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE
195 ~ScopedObjectAccessAlreadyRunnable() ALWAYS_INLINE {
223 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE
230 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE
238 explicit ScopedObjectAccessUnchecked(JavaVM* vm) ALWAYS_INLINE
254 SHARED_LOCK_FUNCTION(Locks::mutator_lock_) ALWAYS_INLINE
260 SHARED_LOCK_FUNCTION(Locks::mutator_lock_) ALWAYS_INLINE
    [all...]
read_barrier.h 50 ALWAYS_INLINE static MirrorType* Barrier(
57 ALWAYS_INLINE static MirrorType* BarrierForRoot(MirrorType** root,
64 ALWAYS_INLINE static MirrorType* BarrierForRoot(mirror::CompressedReference<MirrorType>* root,
83 // ALWAYS_INLINE on this caused a performance regression b/26744236.
96 ALWAYS_INLINE static bool HasGrayReadBarrierPointer(mirror::Object* obj,
utf.h 61 ALWAYS_INLINE int CompareModifiedUtf8ToModifiedUtf8AsUtf16CodePointValues(const char* utf8_1,
116 ALWAYS_INLINE uint16_t GetLeadingUtf16Char(uint32_t maybe_pair);
122 ALWAYS_INLINE uint16_t GetTrailingUtf16Char(uint32_t maybe_pair);
gc_root.h 93 ALWAYS_INLINE void VisitRoot(mirror::Object** root, const RootInfo& info)
99 ALWAYS_INLINE void VisitRootIfNonNull(mirror::Object** root, const RootInfo& info)
171 ALWAYS_INLINE MirrorType* Read(GcRootSource* gc_root_source = nullptr) const
189 ALWAYS_INLINE mirror::CompressedReference<mirror::Object>* AddressWithoutBarrier() {
193 ALWAYS_INLINE bool IsNull() const {
198 ALWAYS_INLINE GcRoot() {}
199 explicit ALWAYS_INLINE GcRoot(MirrorType* ref) SHARED_REQUIRES(Locks::mutator_lock_);
225 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root)
233 ALWAYS_INLINE void VisitRootIfNonNull(mirror::CompressedReference<MirrorType>* root)
  /art/compiler/debug/dwarf/
debug_frame_opcode_writer.h 47 void ALWAYS_INLINE AdvancePC(int absolute_pc) {
73 void ALWAYS_INLINE RelOffset(Reg reg, int offset) {
78 void ALWAYS_INLINE AdjustCFAOffset(int delta) {
83 void ALWAYS_INLINE RelOffsetForMany(Reg reg_base, int offset,
99 void ALWAYS_INLINE RestoreMany(Reg reg_base, uint32_t reg_mask) {
111 void ALWAYS_INLINE Nop() {
117 void ALWAYS_INLINE Offset(Reg reg, int offset) {
139 void ALWAYS_INLINE Restore(Reg reg) {
151 void ALWAYS_INLINE Undefined(Reg reg) {
159 void ALWAYS_INLINE SameValue(Reg reg)
    [all...]
  /external/compiler-rt/lib/builtins/ppc/
DD.h 22 static __inline ALWAYS_INLINE double local_fabs(double x) {
28 static __inline ALWAYS_INLINE double high26bits(double x) {
34 static __inline ALWAYS_INLINE int different_sign(double x, double y) {
  /external/libunwind/include/
compiler.h 41 # define ALWAYS_INLINE inline __attribute__((always_inline))
45 # define ALWAYS_INLINE
59 # define ALWAYS_INLINE
  /art/runtime/base/
arena_object.h 45 ALWAYS_INLINE void* operator new(size_t, void* ptr) noexcept { return ptr; }
46 ALWAYS_INLINE void operator delete(void*, void*) noexcept { }
  /external/compiler-rt/lib/dfsan/
dfsan_platform.h 85 ALWAYS_INLINE
90 ALWAYS_INLINE
95 ALWAYS_INLINE
100 ALWAYS_INLINE
  /art/runtime/mirror/
dex_cache.h 112 String* GetResolvedString(uint32_t string_idx) ALWAYS_INLINE
115 void SetResolvedString(uint32_t string_idx, String* resolved) ALWAYS_INLINE
122 ALWAYS_INLINE ArtMethod* GetResolvedMethod(uint32_t method_idx, size_t ptr_size)
125 ALWAYS_INLINE void SetResolvedMethod(uint32_t method_idx, ArtMethod* resolved, size_t ptr_size)
129 ALWAYS_INLINE ArtField* GetResolvedField(uint32_t idx, size_t ptr_size)
133 ALWAYS_INLINE void SetResolvedField(uint32_t idx, ArtField* field, size_t ptr_size)
136 GcRoot<String>* GetStrings() ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) {
140 void SetStrings(GcRoot<String>* strings) ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) {
144 GcRoot<Class>* GetResolvedTypes() ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) {
149 ALWAYS_INLINE
    [all...]
object_array.h 42 ALWAYS_INLINE T* Get(int32_t i) SHARED_REQUIRES(Locks::mutator_lock_);
50 ALWAYS_INLINE void Set(int32_t i, T* object) SHARED_REQUIRES(Locks::mutator_lock_);
54 ALWAYS_INLINE void Set(int32_t i, T* object) NO_THREAD_SAFETY_ANALYSIS;
62 ALWAYS_INLINE void SetWithoutChecks(int32_t i, T* object) NO_THREAD_SAFETY_ANALYSIS;
67 ALWAYS_INLINE void SetWithoutChecksAndWriteBarrier(int32_t i, T* object)
70 ALWAYS_INLINE T* GetWithoutChecks(int32_t i) SHARED_REQUIRES(Locks::mutator_lock_);
array.h 40 ALWAYS_INLINE static Array* Alloc(Thread* self, Class* array_class, int32_t component_count,
52 ALWAYS_INLINE int32_t GetLength() SHARED_REQUIRES(Locks::mutator_lock_) {
85 ALWAYS_INLINE bool CheckIsValidIndex(int32_t index) SHARED_REQUIRES(Locks::mutator_lock_);
114 const T* GetData() const ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) {
118 T* GetData() ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) {
122 T Get(int32_t i) ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_);
124 T GetWithoutChecks(int32_t i) ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) {
129 void Set(int32_t i, T value) ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_);
134 void Set(int32_t i, T value) ALWAYS_INLINE NO_THREAD_SAFETY_ANALYSIS;
141 void SetWithoutChecks(int32_t i, T value) ALWAYS_INLINE NO_THREAD_SAFETY_ANALYSIS
    [all...]
object.h 90 ALWAYS_INLINE Class* GetClass() SHARED_REQUIRES(Locks::mutator_lock_);
103 ALWAYS_INLINE bool AtomicSetReadBarrierPointer(Object* expected_rb_ptr, Object* rb_ptr)
112 ALWAYS_INLINE bool InstanceOf(Class* klass) SHARED_REQUIRES(Locks::mutator_lock_);
266 ALWAYS_INLINE T* GetFieldObject(MemberOffset field_offset)
271 ALWAYS_INLINE T* GetFieldObjectVolatile(MemberOffset field_offset)
276 ALWAYS_INLINE void SetFieldObjectWithoutWriteBarrier(MemberOffset field_offset, Object* new_value)
281 ALWAYS_INLINE void SetFieldObject(MemberOffset field_offset, Object* new_value)
286 ALWAYS_INLINE void SetFieldObjectVolatile(MemberOffset field_offset, Object* new_value)
328 ALWAYS_INLINE uint8_t GetFieldBoolean(MemberOffset field_offset)
332 ALWAYS_INLINE int8_t GetFieldByte(MemberOffset field_offset
    [all...]
class.h 195 ALWAYS_INLINE uint32_t GetAccessFlags() SHARED_REQUIRES(Locks::mutator_lock_);
201 ALWAYS_INLINE uint32_t GetClassFlags() SHARED_REQUIRES(Locks::mutator_lock_) {
209 ALWAYS_INLINE bool IsInterface() SHARED_REQUIRES(Locks::mutator_lock_) {
214 ALWAYS_INLINE bool IsPublic() SHARED_REQUIRES(Locks::mutator_lock_) {
219 ALWAYS_INLINE bool IsFinal() SHARED_REQUIRES(Locks::mutator_lock_) {
223 ALWAYS_INLINE bool IsFinalizable() SHARED_REQUIRES(Locks::mutator_lock_) {
227 ALWAYS_INLINE void SetRecursivelyInitialized() SHARED_REQUIRES(Locks::mutator_lock_) {
233 ALWAYS_INLINE void SetHasDefaultMethods() SHARED_REQUIRES(Locks::mutator_lock_) {
239 ALWAYS_INLINE void SetFinalizable() SHARED_REQUIRES(Locks::mutator_lock_) {
244 ALWAYS_INLINE bool IsStringClass() SHARED_REQUIRES(Locks::mutator_lock_)
    [all...]
  /art/runtime/gc/collector/
immune_spaces.h 52 ALWAYS_INLINE bool IsInImmuneRegion(const mirror::Object* obj) const {
70 ALWAYS_INLINE bool ContainsObject(const mirror::Object* obj) const {
  /toolchain/binutils/binutils-2.25/gold/testsuite/
gdb_index_test.cc 127 #define ALWAYS_INLINE __attribute__((always_inline))
129 #define ALWAYS_INLINE
132 static inline ALWAYS_INLINE int
  /external/compiler-rt/lib/tsan/rtl/
tsan_platform.h 382 ALWAYS_INLINE
386 ALWAYS_INLINE
392 ALWAYS_INLINE
396 ALWAYS_INLINE
402 ALWAYS_INLINE
406 ALWAYS_INLINE
411 ALWAYS_INLINE
415 ALWAYS_INLINE
420 ALWAYS_INLINE
427 ALWAYS_INLINE
    [all...]

Completed in 1264 milliseconds

1 2 3 4