/frameworks/base/core/java/com/android/internal/util/ |
TokenBucket.java | 30 * For every filling period, the bucket gains one token, up to its maximum capacity from 33 * The available amount of tokens is computed lazily when the bucket state is inspected. 43 private long mLastFill; // Last time in ms the bucket generated tokens.
|
/packages/providers/ContactsProvider/tests/src/com/android/providers/contacts/ |
ContactsDatabaseHelperTest.java | 347 // Zero primary phone book bucket and null primary sort key 354 // Assert that the primary phone book bucket/label has been set to the number bucket/label 364 // Zero alt phone book bucket and null alt sort key 371 // Assert that the alt phone book bucket/label has been set to the number bucket/label 381 // Non-zero primary phone book bucket 389 // Assert that the primary phone book bucket/label is unchanged 402 // Assert that the primary phone book bucket/label is unchanged 408 // Non-zero alt phone book bucket [all...] |
/prebuilts/gcc/linux-x86/host/x86_64-linux-glibc2.11-4.8/x86_64-linux/include/c++/4.8/profile/ |
unordered_base.h | 43 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 52 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 54 { return __uc.bucket(__node->_M_v); } 64 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 66 { return __uc.bucket(__node->_M_v.first); } 76 return __bucket_index_helper::bucket(__uc, __node);
|
/prebuilts/gcc/linux-x86/host/x86_64-linux-glibc2.15-4.8/x86_64-linux/include/c++/4.8/profile/ |
unordered_base.h | 43 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 52 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 54 { return __uc.bucket(__node->_M_v); } 64 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 66 { return __uc.bucket(__node->_M_v.first); } 76 return __bucket_index_helper::bucket(__uc, __node);
|
/prebuilts/gcc/linux-x86/host/x86_64-w64-mingw32-4.8/x86_64-w64-mingw32/include/c++/4.8.3/profile/ |
unordered_base.h | 43 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 52 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 54 { return __uc.bucket(__node->_M_v); } 64 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 66 { return __uc.bucket(__node->_M_v.first); } 76 return __bucket_index_helper::bucket(__uc, __node);
|
/prebuilts/ndk/r10/sources/cxx-stl/gnu-libstdc++/4.9/include/profile/ |
unordered_base.h | 43 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 52 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 54 { return __uc.bucket(__node->_M_v()); } 64 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 66 { return __uc.bucket(__node->_M_v().first); } 76 return __bucket_index_helper::bucket(__uc, __node);
|
/prebuilts/ndk/r11/sources/cxx-stl/gnu-libstdc++/4.9/include/profile/ |
unordered_base.h | 43 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 52 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 54 { return __uc.bucket(__node->_M_v()); } 64 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 66 { return __uc.bucket(__node->_M_v().first); } 76 return __bucket_index_helper::bucket(__uc, __node);
|
/prebuilts/ndk/r13/sources/cxx-stl/gnu-libstdc++/4.9/include/profile/ |
unordered_base.h | 43 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 52 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 54 { return __uc.bucket(__node->_M_v()); } 64 bucket(const _UnorderedCont& __uc, function in struct:__profile::_Bucket_index_helper 66 { return __uc.bucket(__node->_M_v().first); } 76 return __bucket_index_helper::bucket(__uc, __node);
|
/prebuilts/gcc/linux-x86/host/x86_64-linux-glibc2.11-4.8/x86_64-linux/include/c++/4.8/bits/ |
hashtable.h | 90 * which govern the bucket count. _M_next_bkt(n) returns a bucket 92 * bucket count appropriate for an element count of n. 94 * current bucket count is n_bkt and the current element count is 95 * n_elt, we need to increase the bucket count. If so, returns 96 * make_pair(true, n), where n is the new bucket count. If not, 122 * bucket. This design makes it possible to implement something like a 129 * node pointer in a bucket shall never be dereferenced, only its 132 * Walking through a bucket's nodes requires a check on the hash code to 133 * see if each node is still in the bucket. Such a design assumes 514 bucket(const key_type& __k) const function in class:_Hashtable [all...] |
/prebuilts/gcc/linux-x86/host/x86_64-linux-glibc2.15-4.8/x86_64-linux/include/c++/4.8/bits/ |
hashtable.h | 90 * which govern the bucket count. _M_next_bkt(n) returns a bucket 92 * bucket count appropriate for an element count of n. 94 * current bucket count is n_bkt and the current element count is 95 * n_elt, we need to increase the bucket count. If so, returns 96 * make_pair(true, n), where n is the new bucket count. If not, 122 * bucket. This design makes it possible to implement something like a 129 * node pointer in a bucket shall never be dereferenced, only its 132 * Walking through a bucket's nodes requires a check on the hash code to 133 * see if each node is still in the bucket. Such a design assumes 514 bucket(const key_type& __k) const function in class:_Hashtable [all...] |
/prebuilts/gcc/linux-x86/host/x86_64-w64-mingw32-4.8/x86_64-w64-mingw32/include/c++/4.8.3/bits/ |
hashtable.h | 90 * which govern the bucket count. _M_next_bkt(n) returns a bucket 92 * bucket count appropriate for an element count of n. 94 * current bucket count is n_bkt and the current element count is 95 * n_elt, we need to increase the bucket count. If so, returns 96 * make_pair(true, n), where n is the new bucket count. If not, 122 * bucket. This design makes it possible to implement something like a 129 * node pointer in a bucket shall never be dereferenced, only its 132 * Walking through a bucket's nodes requires a check on the hash code to 133 * see if each node is still in the bucket. Such a design assumes 514 bucket(const key_type& __k) const function in class:_Hashtable [all...] |
/external/antlr/antlr-3.4/runtime/C/src/ |
antlr3collections.c | 133 ANTLR3_UINT32 bucket; // Used to traverse the buckets local 153 // Modulo of the table, (bucket count). 161 for (bucket = 0; bucket < sizeHint; bucket++) 163 table->buckets[bucket].entries = NULL; 197 ANTLR3_UINT32 bucket; /* Used to traverse the buckets */ local 208 for (bucket = 0; bucket < table->modulo; bucket++ 277 pANTLR3_HASH_BUCKET bucket; local 334 pANTLR3_HASH_BUCKET bucket; local 447 pANTLR3_HASH_BUCKET bucket; local 486 pANTLR3_HASH_BUCKET bucket; local 526 pANTLR3_HASH_BUCKET bucket; local 597 pANTLR3_HASH_BUCKET bucket; local 767 pANTLR3_HASH_BUCKET bucket; local [all...] |
/docs/source.android.com/src/source/ |
life-of-a-bug.jd | 58 <p>Each bucket includes a number of states that provide more detail on the 62 <p>Bugs in the "Resolved" bucket will eventually be included in a future 66 <h1 id="bucket-details">Bucket Details</h1> 67 <p>Here is some additional information on each bucket, what it means, and how 87 <p>This bucket contains bugs that need action, but which are still 113 <p>This bucket contains bugs that have for one reason or another been 173 <p>This bucket contains bugs that have had action taken, and are now
|
/external/autotest/server/site_tests/video_VDAStress/ |
generate_controlfiles.py | 14 # TODO(ihf): Populate public bucket with test videos. 41 # Download the test videos from the gs bucket to the server.
|
/external/compiler-rt/lib/sanitizer_common/ |
sanitizer_addrhashmap.h | 59 struct Bucket { 82 Bucket *bucket_; 93 Bucket *table_; 151 table_ = (Bucket*)MmapOrDie(kSize * sizeof(table_[0]), "AddrHashMap"); 158 Bucket *b = &table_[hash]; 165 // If we want to remove the element, we need exclusive access to the bucket, 287 Bucket *b = h->bucket_; 301 // See if we need to compact the bucket.
|
/external/emma/core/java12/com/vladium/util/ |
IntSet.java | 84 // index into the corresponding hash bucket: 88 // traverse the singly-linked list of entries in the bucket: 141 // index into the corresponding hash bucket: 144 // traverse the singly-linked list of entries in the bucket: 185 out.append ("size = " + m_size + ", bucket table size = " + m_buckets.length + ", load factor = " + m_loadFactor + EOL); 217 // and then with that value reduced to actually shrink capacity. As it is right now, the bucket table can 225 // rehash all entry chains in every bucket: 233 // index into the corresponding new hash bucket:
|
SoftValueMap.java | 161 // index into the corresponding hash bucket: 168 // traverse the singly-linked list of entries in the bucket: 232 // index into the corresponding hash bucket: 237 // traverse the singly-linked list of entries in the bucket: 280 // recompute the hash bucket index: 299 // index into the corresponding hash bucket: 306 // traverse the singly-linked list of entries in the bucket: 415 out.append ("size = " + m_size + ", bucket table size = " + m_buckets.length + ", load factor = " + m_loadFactor + EOL); 425 * An extension of WeakReference that can store an index of the bucket it 473 // and then with that value reduced to actually shrink capacity. As it is right now, the bucket table ca [all...] |
/external/libchrome/base/metrics/ |
sample_vector.h | 45 // Get count of a specific bucket. 90 // SampleVector uses predefined buckets, so iterator can return bucket index.
|
/external/libchrome/base/trace_event/ |
heap_profiler_allocation_register.h | 59 buckets_(static_cast<Bucket*>( 60 AllocateGuardedVirtualMemory(NumBuckets * sizeof(Bucket)))), 66 FreeGuardedVirtualMemory(buckets_, NumBuckets * sizeof(Bucket)); 137 bits::Align(sizeof(Bucket) * NumBuckets, page_size); 148 // also participate in the bucket's cell list - they point to the list's 151 // that points to this Cell" kind of thing. So |p_prev| points to a bucket 159 using Bucket = Cell*; 228 mutable Bucket* buckets_;
|
/external/ltp/utils/ffsb-6.0-rc2/ |
ffsb_stats.h | 39 /* max = 0 indicates uninitialized bucket */ 85 uint32_t *buckets[FFSB_NUM_SYSCALLS]; /* bucket counters */
|
/packages/apps/Settings/src/com/android/settings/core/instrumentation/ |
EventLogWriter.java | 82 public void histogram(Context context, String name, int bucket) { 83 MetricsLogger.histogram(context, name, bucket);
|
/prebuilts/ndk/r10/sources/cxx-stl/gnu-libstdc++/4.9/include/debug/ |
safe_local_iterator.tcc | 41 if (bucket() != __rhs.bucket())
|
/prebuilts/ndk/r11/sources/cxx-stl/gnu-libstdc++/4.9/include/debug/ |
safe_local_iterator.tcc | 41 if (bucket() != __rhs.bucket())
|
/prebuilts/ndk/r13/sources/cxx-stl/gnu-libstdc++/4.9/include/debug/ |
safe_local_iterator.tcc | 41 if (bucket() != __rhs.bucket())
|
/system/core/logd/ |
LogAudit.cpp | 80 bucket.emplace(android_log_clockid()); 81 oldest = bucket.back() - oldest; 82 while (bucket.front() < oldest) bucket.pop(); 89 if (bucket.size() >= upperThreshold) { 100 bucket.pop(); 101 } while (bucket.size() >= upperThreshold); 110 if (bucket.size() >= lowerThreshold) return;
|