/prebuilts/gcc/linux-x86/host/x86_64-linux-glibc2.11-4.8/x86_64-linux/include/c++/4.8/ext/ |
pool_allocator.h | 100 _M_round_up(size_t __bytes) 101 { return ((__bytes + (size_t)_S_align - 1) & ~((size_t)_S_align - 1)); } 104 _M_get_free_list(size_t __bytes) throw (); 233 const size_t __bytes = __n * sizeof(_Tp); local 234 if (__bytes > size_t(_S_max_bytes) || _S_force_new > 0) 235 __ret = static_cast<_Tp*>(::operator new(__bytes)); 238 _Obj* volatile* __free_list = _M_get_free_list(__bytes); 243 __ret = static_cast<_Tp*>(_M_refill(_M_round_up(__bytes))); 262 const size_t __bytes = __n * sizeof(_Tp); local 263 if (__bytes > static_cast<size_t>(_S_max_bytes) || _S_force_new > 0 [all...] |
mt_allocator.h | 148 _M_check_threshold(size_t __bytes) 149 { return __bytes > _M_options._M_max_bytes || _M_options._M_force_new; } 152 _M_get_binmap(size_t __bytes) 153 { return _M_binmap[__bytes]; } 225 _M_reserve_block(size_t __bytes, const size_t __thread_id); 228 _M_reclaim_block(char* __p, size_t __bytes) throw (); 336 _M_reserve_block(size_t __bytes, const size_t __thread_id); 339 _M_reclaim_block(char* __p, size_t __bytes) throw (); 699 const size_t __bytes = __n * sizeof(_Tp); local 700 if (__pool._M_check_threshold(__bytes)) 743 const size_t __bytes = __n * sizeof(_Tp); local 747 __pool._M_reclaim_block(reinterpret_cast<char*>(__p), __bytes); local [all...] |
/prebuilts/gcc/linux-x86/host/x86_64-linux-glibc2.15-4.8/x86_64-linux/include/c++/4.8/ext/ |
pool_allocator.h | 100 _M_round_up(size_t __bytes) 101 { return ((__bytes + (size_t)_S_align - 1) & ~((size_t)_S_align - 1)); } 104 _M_get_free_list(size_t __bytes) throw (); 233 const size_t __bytes = __n * sizeof(_Tp); local 234 if (__bytes > size_t(_S_max_bytes) || _S_force_new > 0) 235 __ret = static_cast<_Tp*>(::operator new(__bytes)); 238 _Obj* volatile* __free_list = _M_get_free_list(__bytes); 243 __ret = static_cast<_Tp*>(_M_refill(_M_round_up(__bytes))); 262 const size_t __bytes = __n * sizeof(_Tp); local 263 if (__bytes > static_cast<size_t>(_S_max_bytes) || _S_force_new > 0 [all...] |
mt_allocator.h | 148 _M_check_threshold(size_t __bytes) 149 { return __bytes > _M_options._M_max_bytes || _M_options._M_force_new; } 152 _M_get_binmap(size_t __bytes) 153 { return _M_binmap[__bytes]; } 225 _M_reserve_block(size_t __bytes, const size_t __thread_id); 228 _M_reclaim_block(char* __p, size_t __bytes) throw (); 336 _M_reserve_block(size_t __bytes, const size_t __thread_id); 339 _M_reclaim_block(char* __p, size_t __bytes) throw (); 699 const size_t __bytes = __n * sizeof(_Tp); local 700 if (__pool._M_check_threshold(__bytes)) 743 const size_t __bytes = __n * sizeof(_Tp); local 747 __pool._M_reclaim_block(reinterpret_cast<char*>(__p), __bytes); local [all...] |
/prebuilts/gcc/linux-x86/host/x86_64-w64-mingw32-4.8/x86_64-w64-mingw32/include/c++/4.8.3/ext/ |
pool_allocator.h | 100 _M_round_up(size_t __bytes) 101 { return ((__bytes + (size_t)_S_align - 1) & ~((size_t)_S_align - 1)); } 104 _M_get_free_list(size_t __bytes) throw (); 233 const size_t __bytes = __n * sizeof(_Tp); local 234 if (__bytes > size_t(_S_max_bytes) || _S_force_new > 0) 235 __ret = static_cast<_Tp*>(::operator new(__bytes)); 238 _Obj* volatile* __free_list = _M_get_free_list(__bytes); 243 __ret = static_cast<_Tp*>(_M_refill(_M_round_up(__bytes))); 262 const size_t __bytes = __n * sizeof(_Tp); local 263 if (__bytes > static_cast<size_t>(_S_max_bytes) || _S_force_new > 0 [all...] |
mt_allocator.h | 148 _M_check_threshold(size_t __bytes) 149 { return __bytes > _M_options._M_max_bytes || _M_options._M_force_new; } 152 _M_get_binmap(size_t __bytes) 153 { return _M_binmap[__bytes]; } 225 _M_reserve_block(size_t __bytes, const size_t __thread_id); 228 _M_reclaim_block(char* __p, size_t __bytes) throw (); 336 _M_reserve_block(size_t __bytes, const size_t __thread_id); 339 _M_reclaim_block(char* __p, size_t __bytes) throw (); 699 const size_t __bytes = __n * sizeof(_Tp); local 700 if (__pool._M_check_threshold(__bytes)) 743 const size_t __bytes = __n * sizeof(_Tp); local 747 __pool._M_reclaim_block(reinterpret_cast<char*>(__p), __bytes); local [all...] |
/prebuilts/ndk/r10/sources/cxx-stl/gnu-libstdc++/4.9/include/ext/ |
pool_allocator.h | 100 _M_round_up(size_t __bytes) 101 { return ((__bytes + (size_t)_S_align - 1) & ~((size_t)_S_align - 1)); } 104 _M_get_free_list(size_t __bytes) throw (); 233 const size_t __bytes = __n * sizeof(_Tp); local 234 if (__bytes > size_t(_S_max_bytes) || _S_force_new > 0) 235 __ret = static_cast<_Tp*>(::operator new(__bytes)); 238 _Obj* volatile* __free_list = _M_get_free_list(__bytes); 243 __ret = static_cast<_Tp*>(_M_refill(_M_round_up(__bytes))); 262 const size_t __bytes = __n * sizeof(_Tp); local 263 if (__bytes > static_cast<size_t>(_S_max_bytes) || _S_force_new > 0 [all...] |
mt_allocator.h | 148 _M_check_threshold(size_t __bytes) 149 { return __bytes > _M_options._M_max_bytes || _M_options._M_force_new; } 152 _M_get_binmap(size_t __bytes) 153 { return _M_binmap[__bytes]; } 225 _M_reserve_block(size_t __bytes, const size_t __thread_id); 228 _M_reclaim_block(char* __p, size_t __bytes) throw (); 336 _M_reserve_block(size_t __bytes, const size_t __thread_id); 339 _M_reclaim_block(char* __p, size_t __bytes) throw (); 699 const size_t __bytes = __n * sizeof(_Tp); local 700 if (__pool._M_check_threshold(__bytes)) 743 const size_t __bytes = __n * sizeof(_Tp); local 747 __pool._M_reclaim_block(reinterpret_cast<char*>(__p), __bytes); local [all...] |
/prebuilts/ndk/r11/sources/cxx-stl/gnu-libstdc++/4.9/include/ext/ |
pool_allocator.h | 100 _M_round_up(size_t __bytes) 101 { return ((__bytes + (size_t)_S_align - 1) & ~((size_t)_S_align - 1)); } 104 _M_get_free_list(size_t __bytes) throw (); 233 const size_t __bytes = __n * sizeof(_Tp); local 234 if (__bytes > size_t(_S_max_bytes) || _S_force_new > 0) 235 __ret = static_cast<_Tp*>(::operator new(__bytes)); 238 _Obj* volatile* __free_list = _M_get_free_list(__bytes); 243 __ret = static_cast<_Tp*>(_M_refill(_M_round_up(__bytes))); 262 const size_t __bytes = __n * sizeof(_Tp); local 263 if (__bytes > static_cast<size_t>(_S_max_bytes) || _S_force_new > 0 [all...] |
mt_allocator.h | 148 _M_check_threshold(size_t __bytes) 149 { return __bytes > _M_options._M_max_bytes || _M_options._M_force_new; } 152 _M_get_binmap(size_t __bytes) 153 { return _M_binmap[__bytes]; } 225 _M_reserve_block(size_t __bytes, const size_t __thread_id); 228 _M_reclaim_block(char* __p, size_t __bytes) throw (); 336 _M_reserve_block(size_t __bytes, const size_t __thread_id); 339 _M_reclaim_block(char* __p, size_t __bytes) throw (); 699 const size_t __bytes = __n * sizeof(_Tp); local 700 if (__pool._M_check_threshold(__bytes)) 743 const size_t __bytes = __n * sizeof(_Tp); local 747 __pool._M_reclaim_block(reinterpret_cast<char*>(__p), __bytes); local [all...] |
/prebuilts/ndk/r13/sources/cxx-stl/gnu-libstdc++/4.9/include/ext/ |
pool_allocator.h | 100 _M_round_up(size_t __bytes) 101 { return ((__bytes + (size_t)_S_align - 1) & ~((size_t)_S_align - 1)); } 104 _M_get_free_list(size_t __bytes) throw (); 233 const size_t __bytes = __n * sizeof(_Tp); local 234 if (__bytes > size_t(_S_max_bytes) || _S_force_new > 0) 235 __ret = static_cast<_Tp*>(::operator new(__bytes)); 238 _Obj* volatile* __free_list = _M_get_free_list(__bytes); 243 __ret = static_cast<_Tp*>(_M_refill(_M_round_up(__bytes))); 262 const size_t __bytes = __n * sizeof(_Tp); local 263 if (__bytes > static_cast<size_t>(_S_max_bytes) || _S_force_new > 0 [all...] |
mt_allocator.h | 148 _M_check_threshold(size_t __bytes) 149 { return __bytes > _M_options._M_max_bytes || _M_options._M_force_new; } 152 _M_get_binmap(size_t __bytes) 153 { return _M_binmap[__bytes]; } 225 _M_reserve_block(size_t __bytes, const size_t __thread_id); 228 _M_reclaim_block(char* __p, size_t __bytes) throw (); 336 _M_reserve_block(size_t __bytes, const size_t __thread_id); 339 _M_reclaim_block(char* __p, size_t __bytes) throw (); 699 const size_t __bytes = __n * sizeof(_Tp); local 700 if (__pool._M_check_threshold(__bytes)) 743 const size_t __bytes = __n * sizeof(_Tp); local 747 __pool._M_reclaim_block(reinterpret_cast<char*>(__p), __bytes); local [all...] |