OpenGrok
Home
Sort by relevance
Sort by last modified time
Full Search
Definition
Symbol
File Path
History
|
|
Help
Searched
refs:ALWAYS_INLINE
(Results
1 - 25
of
95
) sorted by null
1
2
3
4
/art/runtime/gc/accounting/
bitmap.h
49
static
ALWAYS_INLINE
constexpr size_t BitIndexToWordIndex(uintptr_t offset) {
54
static
ALWAYS_INLINE
constexpr T WordIndexToBitIndex(T word_index) {
58
static
ALWAYS_INLINE
constexpr uintptr_t BitIndexToMask(uintptr_t bit_index) {
62
ALWAYS_INLINE
bool SetBit(size_t bit_index) {
66
ALWAYS_INLINE
bool ClearBit(size_t bit_index) {
70
ALWAYS_INLINE
bool TestBit(size_t bit_index) const;
73
ALWAYS_INLINE
bool AtomicTestAndSetBit(size_t bit_index);
96
ALWAYS_INLINE
void CheckValidBitIndex(size_t bit_index) const {
112
ALWAYS_INLINE
bool ModifyBit(uintptr_t bit_index);
137
ALWAYS_INLINE
uintptr_t CoverBegin() const
[
all
...]
mod_union_table-inl.h
36
bool ShouldAddReference(const mirror::Object* ref) const OVERRIDE
ALWAYS_INLINE
{
card_table.h
59
ALWAYS_INLINE
void MarkCard(const void *addr) {
126
void* AddrFromCard(const uint8_t *card_addr) const
ALWAYS_INLINE
;
129
uint8_t* CardFromAddr(const void *addr) const
ALWAYS_INLINE
;
137
bool IsValidCard(const uint8_t* card_addr) const
ALWAYS_INLINE
;
139
void CheckCardValid(uint8_t* card) const
ALWAYS_INLINE
;
/art/runtime/native/
scoped_fast_native_object_access.h
31
SHARED_LOCK_FUNCTION(Locks::mutator_lock_)
ALWAYS_INLINE
39
~ScopedFastNativeObjectAccess() UNLOCK_FUNCTION(Locks::mutator_lock_)
ALWAYS_INLINE
{
/art/runtime/
handle.h
42
ALWAYS_INLINE
Handle(const Handle<T>& handle) : reference_(handle.reference_) {
45
ALWAYS_INLINE
Handle<T>& operator=(const Handle<T>& handle) {
50
ALWAYS_INLINE
explicit Handle(StackReference<T>* reference) : reference_(reference) {
53
ALWAYS_INLINE
T& operator*() const SHARED_REQUIRES(Locks::mutator_lock_) {
57
ALWAYS_INLINE
T* operator->() const SHARED_REQUIRES(Locks::mutator_lock_) {
61
ALWAYS_INLINE
T* Get() const SHARED_REQUIRES(Locks::mutator_lock_) {
65
ALWAYS_INLINE
jobject ToJObject() const SHARED_REQUIRES(Locks::mutator_lock_) {
73
ALWAYS_INLINE
StackReference<mirror::Object>* GetReference() {
77
ALWAYS_INLINE
const StackReference<mirror::Object>* GetReference() const {
108
ALWAYS_INLINE
MutableHandle(const MutableHandle<T>& handle
[
all
...]
handle_scope.h
62
ALWAYS_INLINE
mirror::Object* GetReference(size_t i) const
65
ALWAYS_INLINE
Handle<mirror::Object> GetHandle(size_t i);
67
ALWAYS_INLINE
MutableHandle<mirror::Object> GetMutableHandle(size_t i)
70
ALWAYS_INLINE
void SetReference(size_t i, mirror::Object* object)
73
ALWAYS_INLINE
bool Contains(StackReference<mirror::Object>* handle_scope_entry) const;
98
ALWAYS_INLINE
StackReference<mirror::Object>* GetReferences() const {
148
explicit
ALWAYS_INLINE
StackHandleScope(Thread* self, mirror::Object* fill_value = nullptr);
149
ALWAYS_INLINE
~StackHandleScope();
152
ALWAYS_INLINE
MutableHandle<T> NewHandle(T* object) SHARED_REQUIRES(Locks::mutator_lock_);
155
ALWAYS_INLINE
HandleWrapper<T> NewHandleWrapper(T** object
[
all
...]
verify_object.h
55
ALWAYS_INLINE
void VerifyObject(mirror::Object* obj) NO_THREAD_SAFETY_ANALYSIS;
58
ALWAYS_INLINE
bool VerifyClassClass(mirror::Class* c) NO_THREAD_SAFETY_ANALYSIS;
memory_region.h
55
ALWAYS_INLINE
T Load(uintptr_t offset) const {
65
ALWAYS_INLINE
void Store(uintptr_t offset, T value) const {
74
ALWAYS_INLINE
T LoadUnaligned(uintptr_t offset) const {
89
ALWAYS_INLINE
void StoreUnaligned(uintptr_t offset, T value) const {
101
ALWAYS_INLINE
T* PointerTo(uintptr_t offset) const {
107
ALWAYS_INLINE
bool LoadBit(uintptr_t bit_offset) const {
113
ALWAYS_INLINE
void StoreBit(uintptr_t bit_offset, bool value) const {
165
ALWAYS_INLINE
T* ComputeInternalPointer(size_t offset) const {
173
ALWAYS_INLINE
uint8_t* ComputeBitPointer(uintptr_t bit_offset, uint8_t* bit_mask) const {
scoped_thread_state_change.h
37
REQUIRES(!Locks::thread_suspend_count_lock_)
ALWAYS_INLINE
62
~ScopedThreadStateChange() REQUIRES(!Locks::thread_suspend_count_lock_)
ALWAYS_INLINE
{
179
REQUIRES(!Locks::thread_suspend_count_lock_)
ALWAYS_INLINE
184
REQUIRES(!Locks::thread_suspend_count_lock_)
ALWAYS_INLINE
195
~ScopedObjectAccessAlreadyRunnable()
ALWAYS_INLINE
{
223
REQUIRES(!Locks::thread_suspend_count_lock_)
ALWAYS_INLINE
230
REQUIRES(!Locks::thread_suspend_count_lock_)
ALWAYS_INLINE
238
explicit ScopedObjectAccessUnchecked(JavaVM* vm)
ALWAYS_INLINE
254
SHARED_LOCK_FUNCTION(Locks::mutator_lock_)
ALWAYS_INLINE
260
SHARED_LOCK_FUNCTION(Locks::mutator_lock_)
ALWAYS_INLINE
[
all
...]
read_barrier.h
50
ALWAYS_INLINE
static MirrorType* Barrier(
57
ALWAYS_INLINE
static MirrorType* BarrierForRoot(MirrorType** root,
64
ALWAYS_INLINE
static MirrorType* BarrierForRoot(mirror::CompressedReference<MirrorType>* root,
83
//
ALWAYS_INLINE
on this caused a performance regression b/26744236.
96
ALWAYS_INLINE
static bool HasGrayReadBarrierPointer(mirror::Object* obj,
utf.h
61
ALWAYS_INLINE
int CompareModifiedUtf8ToModifiedUtf8AsUtf16CodePointValues(const char* utf8_1,
116
ALWAYS_INLINE
uint16_t GetLeadingUtf16Char(uint32_t maybe_pair);
122
ALWAYS_INLINE
uint16_t GetTrailingUtf16Char(uint32_t maybe_pair);
gc_root.h
93
ALWAYS_INLINE
void VisitRoot(mirror::Object** root, const RootInfo& info)
99
ALWAYS_INLINE
void VisitRootIfNonNull(mirror::Object** root, const RootInfo& info)
171
ALWAYS_INLINE
MirrorType* Read(GcRootSource* gc_root_source = nullptr) const
189
ALWAYS_INLINE
mirror::CompressedReference<mirror::Object>* AddressWithoutBarrier() {
193
ALWAYS_INLINE
bool IsNull() const {
198
ALWAYS_INLINE
GcRoot() {}
199
explicit
ALWAYS_INLINE
GcRoot(MirrorType* ref) SHARED_REQUIRES(Locks::mutator_lock_);
225
ALWAYS_INLINE
void VisitRootIfNonNull(GcRoot<MirrorType>& root)
233
ALWAYS_INLINE
void VisitRootIfNonNull(mirror::CompressedReference<MirrorType>* root)
/art/compiler/debug/dwarf/
debug_frame_opcode_writer.h
47
void
ALWAYS_INLINE
AdvancePC(int absolute_pc) {
73
void
ALWAYS_INLINE
RelOffset(Reg reg, int offset) {
78
void
ALWAYS_INLINE
AdjustCFAOffset(int delta) {
83
void
ALWAYS_INLINE
RelOffsetForMany(Reg reg_base, int offset,
99
void
ALWAYS_INLINE
RestoreMany(Reg reg_base, uint32_t reg_mask) {
111
void
ALWAYS_INLINE
Nop() {
117
void
ALWAYS_INLINE
Offset(Reg reg, int offset) {
139
void
ALWAYS_INLINE
Restore(Reg reg) {
151
void
ALWAYS_INLINE
Undefined(Reg reg) {
159
void
ALWAYS_INLINE
SameValue(Reg reg)
[
all
...]
/external/compiler-rt/lib/builtins/ppc/
DD.h
22
static __inline
ALWAYS_INLINE
double local_fabs(double x) {
28
static __inline
ALWAYS_INLINE
double high26bits(double x) {
34
static __inline
ALWAYS_INLINE
int different_sign(double x, double y) {
/external/libunwind/include/
compiler.h
41
# define
ALWAYS_INLINE
inline __attribute__((
always_inline
))
45
# define
ALWAYS_INLINE
59
# define
ALWAYS_INLINE
/art/runtime/base/
arena_object.h
45
ALWAYS_INLINE
void* operator new(size_t, void* ptr) noexcept { return ptr; }
46
ALWAYS_INLINE
void operator delete(void*, void*) noexcept { }
/external/compiler-rt/lib/dfsan/
dfsan_platform.h
85
ALWAYS_INLINE
90
ALWAYS_INLINE
95
ALWAYS_INLINE
100
ALWAYS_INLINE
/art/runtime/mirror/
dex_cache.h
112
String* GetResolvedString(uint32_t string_idx)
ALWAYS_INLINE
115
void SetResolvedString(uint32_t string_idx, String* resolved)
ALWAYS_INLINE
122
ALWAYS_INLINE
ArtMethod* GetResolvedMethod(uint32_t method_idx, size_t ptr_size)
125
ALWAYS_INLINE
void SetResolvedMethod(uint32_t method_idx, ArtMethod* resolved, size_t ptr_size)
129
ALWAYS_INLINE
ArtField* GetResolvedField(uint32_t idx, size_t ptr_size)
133
ALWAYS_INLINE
void SetResolvedField(uint32_t idx, ArtField* field, size_t ptr_size)
136
GcRoot<String>* GetStrings()
ALWAYS_INLINE
SHARED_REQUIRES(Locks::mutator_lock_) {
140
void SetStrings(GcRoot<String>* strings)
ALWAYS_INLINE
SHARED_REQUIRES(Locks::mutator_lock_) {
144
GcRoot<Class>* GetResolvedTypes()
ALWAYS_INLINE
SHARED_REQUIRES(Locks::mutator_lock_) {
149
ALWAYS_INLINE
[
all
...]
object_array.h
42
ALWAYS_INLINE
T* Get(int32_t i) SHARED_REQUIRES(Locks::mutator_lock_);
50
ALWAYS_INLINE
void Set(int32_t i, T* object) SHARED_REQUIRES(Locks::mutator_lock_);
54
ALWAYS_INLINE
void Set(int32_t i, T* object) NO_THREAD_SAFETY_ANALYSIS;
62
ALWAYS_INLINE
void SetWithoutChecks(int32_t i, T* object) NO_THREAD_SAFETY_ANALYSIS;
67
ALWAYS_INLINE
void SetWithoutChecksAndWriteBarrier(int32_t i, T* object)
70
ALWAYS_INLINE
T* GetWithoutChecks(int32_t i) SHARED_REQUIRES(Locks::mutator_lock_);
array.h
40
ALWAYS_INLINE
static Array* Alloc(Thread* self, Class* array_class, int32_t component_count,
52
ALWAYS_INLINE
int32_t GetLength() SHARED_REQUIRES(Locks::mutator_lock_) {
85
ALWAYS_INLINE
bool CheckIsValidIndex(int32_t index) SHARED_REQUIRES(Locks::mutator_lock_);
114
const T* GetData() const
ALWAYS_INLINE
SHARED_REQUIRES(Locks::mutator_lock_) {
118
T* GetData()
ALWAYS_INLINE
SHARED_REQUIRES(Locks::mutator_lock_) {
122
T Get(int32_t i)
ALWAYS_INLINE
SHARED_REQUIRES(Locks::mutator_lock_);
124
T GetWithoutChecks(int32_t i)
ALWAYS_INLINE
SHARED_REQUIRES(Locks::mutator_lock_) {
129
void Set(int32_t i, T value)
ALWAYS_INLINE
SHARED_REQUIRES(Locks::mutator_lock_);
134
void Set(int32_t i, T value)
ALWAYS_INLINE
NO_THREAD_SAFETY_ANALYSIS;
141
void SetWithoutChecks(int32_t i, T value)
ALWAYS_INLINE
NO_THREAD_SAFETY_ANALYSIS
[
all
...]
object.h
90
ALWAYS_INLINE
Class* GetClass() SHARED_REQUIRES(Locks::mutator_lock_);
103
ALWAYS_INLINE
bool AtomicSetReadBarrierPointer(Object* expected_rb_ptr, Object* rb_ptr)
112
ALWAYS_INLINE
bool InstanceOf(Class* klass) SHARED_REQUIRES(Locks::mutator_lock_);
266
ALWAYS_INLINE
T* GetFieldObject(MemberOffset field_offset)
271
ALWAYS_INLINE
T* GetFieldObjectVolatile(MemberOffset field_offset)
276
ALWAYS_INLINE
void SetFieldObjectWithoutWriteBarrier(MemberOffset field_offset, Object* new_value)
281
ALWAYS_INLINE
void SetFieldObject(MemberOffset field_offset, Object* new_value)
286
ALWAYS_INLINE
void SetFieldObjectVolatile(MemberOffset field_offset, Object* new_value)
328
ALWAYS_INLINE
uint8_t GetFieldBoolean(MemberOffset field_offset)
332
ALWAYS_INLINE
int8_t GetFieldByte(MemberOffset field_offset
[
all
...]
class.h
195
ALWAYS_INLINE
uint32_t GetAccessFlags() SHARED_REQUIRES(Locks::mutator_lock_);
201
ALWAYS_INLINE
uint32_t GetClassFlags() SHARED_REQUIRES(Locks::mutator_lock_) {
209
ALWAYS_INLINE
bool IsInterface() SHARED_REQUIRES(Locks::mutator_lock_) {
214
ALWAYS_INLINE
bool IsPublic() SHARED_REQUIRES(Locks::mutator_lock_) {
219
ALWAYS_INLINE
bool IsFinal() SHARED_REQUIRES(Locks::mutator_lock_) {
223
ALWAYS_INLINE
bool IsFinalizable() SHARED_REQUIRES(Locks::mutator_lock_) {
227
ALWAYS_INLINE
void SetRecursivelyInitialized() SHARED_REQUIRES(Locks::mutator_lock_) {
233
ALWAYS_INLINE
void SetHasDefaultMethods() SHARED_REQUIRES(Locks::mutator_lock_) {
239
ALWAYS_INLINE
void SetFinalizable() SHARED_REQUIRES(Locks::mutator_lock_) {
244
ALWAYS_INLINE
bool IsStringClass() SHARED_REQUIRES(Locks::mutator_lock_)
[
all
...]
/art/runtime/gc/collector/
immune_spaces.h
52
ALWAYS_INLINE
bool IsInImmuneRegion(const mirror::Object* obj) const {
70
ALWAYS_INLINE
bool ContainsObject(const mirror::Object* obj) const {
/toolchain/binutils/binutils-2.25/gold/testsuite/
gdb_index_test.cc
127
#define
ALWAYS_INLINE
__attribute__((
always_inline
))
129
#define
ALWAYS_INLINE
132
static inline
ALWAYS_INLINE
int
/external/compiler-rt/lib/tsan/rtl/
tsan_platform.h
382
ALWAYS_INLINE
386
ALWAYS_INLINE
392
ALWAYS_INLINE
396
ALWAYS_INLINE
402
ALWAYS_INLINE
406
ALWAYS_INLINE
411
ALWAYS_INLINE
415
ALWAYS_INLINE
420
ALWAYS_INLINE
427
ALWAYS_INLINE
[
all
...]
Completed in 1264 milliseconds
1
2
3
4