/external/webrtc/webrtc/call/ |
bitrate_allocator_unittest.cc | 39 BitrateAllocatorTest() : allocator_(new BitrateAllocator()) { 40 allocator_->OnNetworkChanged(300000u, 0, 0); 44 rtc::scoped_ptr<BitrateAllocator> allocator_; member in class:webrtc::BitrateAllocatorTest 50 allocator_->AddBitrateObserver(&bitrate_observer, 100000, 1500000); 52 allocator_->OnNetworkChanged(200000, 0, 0); 57 allocator_->OnNetworkChanged(4000000, 0, 0); 60 allocator_->AddBitrateObserver(&bitrate_observer, 100000, 4000000); 64 allocator_->AddBitrateObserver(&bitrate_observer, 100000, 1500000); 67 allocator_->OnNetworkChanged(1500000, 0, 0); 75 allocator_->AddBitrateObserver(&bitrate_observer_1, 100000, 300000) 111 rtc::scoped_ptr<BitrateAllocator> allocator_; member in class:webrtc::BitrateAllocatorTestNoEnforceMin [all...] |
/art/runtime/ |
linear_alloc.cc | 23 LinearAlloc::LinearAlloc(ArenaPool* pool) : lock_("linear alloc"), allocator_(pool) { 28 return allocator_.Realloc(ptr, old_size, new_size); 33 return allocator_.Alloc(size); 38 return allocator_.AllocAlign16(size); 43 return allocator_.BytesUsed(); 48 return allocator_.GetArenaPool(); 53 return allocator_.Contains(ptr); 57 return allocator_.Contains(ptr);
|
/external/tensorflow/tensorflow/core/common_runtime/ |
visitable_allocator.h | 54 : TrackingAllocator(allocator, track_ids), allocator_(allocator) {} 68 allocator_->AddAllocVisitor(visitor); 72 allocator_->AddFreeVisitor(visitor); 76 VisitableAllocator* allocator_; member in class:tensorflow::TrackingVisitableAllocator
|
mkl_cpu_allocator.h | 60 ~MklCPUAllocator() override { delete allocator_; } 94 allocator_ = new BFCAllocator(new MklSubAllocator, max_mem_bytes, 110 return allocator_->AllocateRaw(alignment, num_bytes); 114 allocator_->DeallocateRaw(ptr); 117 void GetStats(AllocatorStats* stats) override { allocator_->GetStats(stats); } 119 void ClearStats() override { allocator_->ClearStats(); } 122 allocator_->AddAllocVisitor(visitor); 126 allocator_->AddFreeVisitor(visitor); 163 VisitableAllocator* allocator_; // owned by this class member in class:tensorflow::MklCPUAllocator
|
threadpool_device.h | 41 Allocator* allocator_; // Not owned member in class:tensorflow::ThreadPoolDevice
|
threadpool_device.cc | 43 allocator_(allocator) {} 65 return allocator_;
|
/packages/apps/Test/connectivity/sl4n/rapidjson/include/rapidjson/internal/ |
stack.h | 35 Stack(Allocator* allocator, size_t stackCapacity) : allocator_(allocator), ownAllocator_(0), stack_(0), stackTop_(0), stackEnd_(0), initialCapacity_(stackCapacity) {
41 : allocator_(rhs.allocator_),
48 rhs.allocator_ = 0;
67 allocator_ = rhs.allocator_;
74 rhs.allocator_ = 0;
86 internal::Swap(allocator_, rhs.allocator_);
138 return allocator_ != 0; 185 Allocator* allocator_; member in class:internal::Stack [all...] |
/external/tensorflow/tensorflow/core/framework/ |
tracking_allocator.cc | 24 : allocator_(allocator), 29 track_sizes_locally_(track_sizes && !allocator_->TracksAllocationSizes()), 35 void* ptr = allocator_->AllocateRaw(alignment, num_bytes, allocation_attr); 41 if (allocator_->TracksAllocationSizes()) { 42 size_t allocated_bytes = allocator_->AllocatedSize(ptr); 55 size_t allocated_bytes = allocator_->AllocatedSizeSlow(ptr); 83 bool tracks_allocation_sizes = allocator_->TracksAllocationSizes(); 86 allocated_bytes = allocator_->AllocatedSize(ptr); 96 Allocator* allocator = allocator_; 113 return track_sizes_locally_ || allocator_->TracksAllocationSizes() [all...] |
/external/google-breakpad/src/common/ |
memory_unittest.cc | 70 PageAllocator allocator_; local 71 wasteful_vector<int> v(&allocator_); 77 PageAllocator allocator_; local 78 wasteful_vector<unsigned> v(&allocator_); 92 PageAllocator allocator_; local 93 wasteful_vector<unsigned> v(&allocator_); 96 ASSERT_TRUE(allocator_.OwnsPointer(&v[0]));
|
/external/libchrome/base/metrics/ |
persistent_memory_allocator_unittest.cc | 65 allocator_.reset(); 67 allocator_.reset(new PersistentMemoryAllocator( 73 allocator_.reset(); 77 PersistentMemoryAllocator::Iterator iter(allocator_.get()); 92 std::unique_ptr<PersistentMemoryAllocator> allocator_; member in class:base::PersistentMemoryAllocatorTest 96 allocator_->CreateTrackingHistograms(allocator_->Name()); 99 EXPECT_EQ(TEST_ID, allocator_->Id()); 100 EXPECT_TRUE(allocator_->used_histogram_); 102 allocator_->used_histogram_->histogram_name()) 317 PersistentMemoryAllocator allocator_; member in class:base::AllocatorThread [all...] |
/art/compiler/utils/ |
assembler.cc | 29 : allocator_(allocator) { 31 contents_ = allocator_->AllocArray<uint8_t>(kInitialBufferCapacity, kArenaAllocAssembler); 48 if (allocator_->IsRunningOnMemoryTool()) { 49 allocator_->MakeInaccessible(contents_, Capacity()); 84 allocator_->Realloc(contents_, old_capacity, new_capacity, kArenaAllocAssembler));
|
jni_macro_assembler_test.h | 61 allocator_.reset(new ArenaAllocator(&pool_)); 62 assembler_.reset(CreateAssembler(allocator_.get())); 79 allocator_.reset(); 143 std::unique_ptr<ArenaAllocator> allocator_; member in class:art::JNIMacroAssemblerTest
|
/external/tensorflow/tensorflow/compiler/xla/service/gpu/ |
cudnn_convolution_algorithm_picker.h | 39 : stream_exec_(stream_exec), allocator_(allocator) {} 56 DeviceMemoryAllocator* allocator_; // may be null member in class:xla::gpu::CudnnConvolutionAlgorithmPicker
|
/external/tensorflow/tensorflow/compiler/xla/tests/ |
local_client_allocation_test.cc | 50 int64 allocation_count_before = allocator_->allocation_count(); 65 EXPECT_GT(allocator_->allocation_count(), allocation_count_before); 68 int64 deallocation_count_before = allocator_->deallocation_count(); 70 EXPECT_EQ(deallocation_count_before + 1, allocator_->deallocation_count());
|
/art/runtime/base/ |
arena_bit_vector.cc | 66 return allocator_->Alloc(size, this->Kind()); 73 : ArenaBitVectorAllocatorKind(kind), allocator_(allocator) { } 75 ArenaAlloc* const allocator_; member in class:art::FINAL
|
arena_containers.h | 143 allocator_(allocator) { 148 allocator_(other.allocator_) { 155 ArenaAllocator* allocator_; member in class:art::ArenaAllocatorAdapter 179 allocator_(allocator) { 184 allocator_(other.allocator_) { 200 return allocator_->AllocArray<T>(n, ArenaAllocatorAdapterKind::Kind()); 203 allocator_->MakeInaccessible(p, sizeof(T) * n); 216 ArenaAllocator* allocator_; member in class:art::ArenaAllocatorAdapter [all...] |
/art/runtime/mirror/ |
class_loader.h | 59 GetField64(OFFSET_OF_OBJECT_MEMBER(ClassLoader, allocator_))); 63 SetField64<false>(OFFSET_OF_OBJECT_MEMBER(ClassLoader, allocator_), 84 uint64_t allocator_; member in class:art::mirror::ClassLoader
|
/system/core/libmemunreachable/ |
ScopedSignalHandler.h | 35 explicit ScopedSignalHandler(Allocator<Fn> allocator) : allocator_(allocator), signal_(-1) {} 42 handler_ = SignalFn(std::allocator_arg, allocator_, 71 Allocator<Fn> allocator_; member in class:android::ScopedSignalHandler
|
/art/compiler/optimizing/ |
gvn.cc | 41 : allocator_(allocator), 55 : allocator_(allocator), 95 buckets_[index] = new (allocator_) Node(instruction, hash_code, buckets_[index]); 180 buckets_[new_index] = node->Dup(allocator_, buckets_[new_index]); 221 clone_current = node->Dup(allocator_, nullptr); 320 ScopedArenaAllocator* const allocator_; member in class:art::ValueSet 347 allocator_(graph->GetArenaStack()), 349 sets_(graph->GetBlocks().size(), nullptr, allocator_.Adapter(kArenaAllocGvn)), 351 &allocator_, graph->GetBlocks().size(), /* expandable */ false, kArenaAllocGvn) { 363 ScopedArenaAllocator allocator_; member in class:art::GlobalValueNumberer [all...] |
bounds_check_elimination.cc | 292 : allocator_(allocator), lower_(lower), upper_(upper) {} 301 ScopedArenaAllocator* GetAllocator() const { return allocator_; } 329 return new (allocator_) ValueRange( 330 allocator_, 350 return new (allocator_) ValueRange(allocator_, lower, upper); 354 ScopedArenaAllocator* const allocator_; member in class:art::ValueRange 514 allocator_(graph->GetArenaStack()), 518 allocator_.Adapter(kArenaAllocBoundsCheckElimination)), 519 allocator_.Adapter(kArenaAllocBoundsCheckElimination)) 1905 ScopedArenaAllocator allocator_; member in class:art::BCEVisitor [all...] |
intrinsics_arm64.h | 43 : allocator_(allocator), codegen_(codegen) {} 60 ArenaAllocator* const allocator_; member in class:art::arm64::FINAL
|
/external/ImageMagick/Magick++/lib/Magick++/ |
Blob.h | 68 // Specify allocator_ as "MallocAllocator" if memory is allocated 72 const Allocator allocator_=NewAllocator);
|
/external/webrtc/talk/app/webrtc/ |
datachannel_unittest.cc | 513 SctpSidAllocator allocator_; member in class:SctpSidAllocatorTest 520 EXPECT_TRUE(allocator_.AllocateSid(rtc::SSL_SERVER, &id)); 522 EXPECT_TRUE(allocator_.AllocateSid(rtc::SSL_CLIENT, &id)); 524 EXPECT_TRUE(allocator_.AllocateSid(rtc::SSL_SERVER, &id)); 526 EXPECT_TRUE(allocator_.AllocateSid(rtc::SSL_CLIENT, &id)); 533 EXPECT_TRUE(allocator_.ReserveSid(old_id)); 536 EXPECT_TRUE(allocator_.AllocateSid(rtc::SSL_SERVER, &new_id)); 540 EXPECT_TRUE(allocator_.ReserveSid(old_id)); 541 EXPECT_TRUE(allocator_.AllocateSid(rtc::SSL_CLIENT, &new_id)); 549 EXPECT_TRUE(allocator_.ReserveSid(odd_id)) [all...] |
/external/tensorflow/tensorflow/compiler/xla/ |
executable_run_options.cc | 30 allocator_ = allocator; 35 return allocator_;
|
/external/tensorflow/tensorflow/core/kernels/ |
ops_testutil.cc | 33 allocator_ = managed_allocator_.get(); 36 allocator_ = device_->GetAllocator(AllocatorAttributes());
|