/external/tensorflow/tensorflow/core/framework/ |
tracking_allocator.h | 59 string Name() override { return allocator_->Name(); } 98 Allocator* allocator_; // not owned. member in class:tensorflow::TrackingAllocator
|
/art/compiler/optimizing/ |
instruction_builder.cc | 56 : allocator_(graph->GetAllocator()), 109 HPhi* phi = new (allocator_) HPhi( 110 allocator_, 159 HPhi* phi = new (allocator_) HPhi( 160 allocator_, 198 HPhi* phi = new (allocator_) HPhi( 199 allocator_, 260 HEnvironment* environment = new (allocator_) HEnvironment( 261 allocator_, 277 HNullCheck* null_check = new (allocator_) HNullCheck(ref, dex_pc) [all...] |
intrinsics_mips64.cc | 38 : codegen_(codegen), allocator_(codegen->GetGraph()->GetAllocator()) { 164 CreateFPToIntLocations(allocator_, invoke); 173 CreateFPToIntLocations(allocator_, invoke); 200 CreateIntToFPLocations(allocator_, invoke); 209 CreateIntToFPLocations(allocator_, invoke); 250 CreateIntToIntLocations(allocator_, invoke); 259 CreateIntToIntLocations(allocator_, invoke); 268 CreateIntToIntLocations(allocator_, invoke); 290 CreateIntToIntLocations(allocator_, invoke); 299 CreateIntToIntLocations(allocator_, invoke) [all...] |
intrinsics_arm64.cc | 268 CreateFPToIntLocations(allocator_, invoke); 271 CreateIntToFPLocations(allocator_, invoke); 282 CreateFPToIntLocations(allocator_, invoke); 285 CreateIntToFPLocations(allocator_, invoke); 324 CreateIntToIntLocations(allocator_, invoke); 332 CreateIntToIntLocations(allocator_, invoke); 340 CreateIntToIntLocations(allocator_, invoke); 367 CreateIntToIntLocations(allocator_, invoke); 375 CreateIntToIntLocations(allocator_, invoke); 395 CreateIntToIntLocations(allocator_, invoke) [all...] |
intrinsics_x86_64.cc | 44 : allocator_(codegen->GetGraph()->GetAllocator()), codegen_(codegen) { 158 CreateFPToIntLocations(allocator_, invoke); 161 CreateIntToFPLocations(allocator_, invoke); 172 CreateFPToIntLocations(allocator_, invoke); 175 CreateIntToFPLocations(allocator_, invoke); 216 CreateIntToIntLocations(allocator_, invoke); 224 CreateIntToIntLocations(allocator_, invoke); 232 CreateIntToIntLocations(allocator_, invoke); 274 CreateFloatToFloatPlusTemps(allocator_, invoke); 282 CreateFloatToFloatPlusTemps(allocator_, invoke) [all...] |
intrinsics_arm_vixl.cc | 249 : allocator_(codegen->GetGraph()->GetAllocator()), 298 CreateFPToIntLocations(allocator_, invoke); 301 CreateIntToFPLocations(allocator_, invoke); 312 CreateFPToIntLocations(allocator_, invoke); 315 CreateIntToFPLocations(allocator_, invoke); 374 CreateIntToIntLocations(allocator_, invoke); 382 CreateLongToLongLocationsWithOverlap(allocator_, invoke); 420 CreateIntToIntLocations(allocator_, invoke); 428 CreateLongToLongLocationsWithOverlap(allocator_, invoke); 440 CreateFPToFPLocations(allocator_, invoke) [all...] |
intrinsics_x86.cc | 49 : allocator_(codegen->GetGraph()->GetAllocator()), 231 CreateFPToIntLocations(allocator_, invoke, /* is64bit */ true); 234 CreateIntToFPLocations(allocator_, invoke, /* is64bit */ true); 245 CreateFPToIntLocations(allocator_, invoke, /* is64bit */ false); 248 CreateIntToFPLocations(allocator_, invoke, /* is64bit */ false); 300 CreateIntToIntLocations(allocator_, invoke); 308 CreateLongToLongLocations(allocator_, invoke); 329 CreateIntToIntLocations(allocator_, invoke); 398 CreateFloatToFloat(allocator_, invoke); 406 CreateFloatToFloat(allocator_, invoke) [all...] |
intrinsics_mips.cc | 38 : codegen_(codegen), allocator_(codegen->GetGraph()->GetAllocator()) { 180 CreateFPToIntLocations(allocator_, invoke); 189 CreateFPToIntLocations(allocator_, invoke); 221 CreateIntToFPLocations(allocator_, invoke); 230 CreateIntToFPLocations(allocator_, invoke); 402 CreateIntToIntLocations(allocator_, invoke); 416 CreateIntToIntLocations(allocator_, invoke); 430 CreateIntToIntLocations(allocator_, invoke); 474 CreateIntToIntLocations(allocator_, invoke); 483 CreateIntToIntLocations(allocator_, invoke) [all...] |
parallel_move_resolver.h | 129 allocator_(allocator) { 197 ArenaAllocator* const allocator_; member in class:art::ParallelMoveResolverNoSwap
|
/art/runtime/verifier/ |
reg_type_cache.cc | 173 char* ptr = allocator_.AllocArray<char>(string_piece.length()); 207 new (&allocator_) PreciseReferenceType(klass, AddString(sp_descriptor), entries_.size()); 209 entry = new (&allocator_) ReferenceType(klass, AddString(sp_descriptor), entries_.size()); 223 new (&allocator_) UnresolvedReferenceType(AddString(sp_descriptor), entries_.size())); 234 return AddEntry(new (&allocator_) UnresolvedReferenceType(AddString("a"), entries_.size())); 263 new (&allocator_) PreciseReferenceType(klass, descriptor, entries_.size())) 264 : new (&allocator_) ReferenceType(klass, descriptor, entries_.size()); 281 allocator_(allocator) { 383 ArenaBitVector types(&allocator_, 463 return AddEntry(new (&allocator_) UnresolvedMergedType(resolved_parts_merged [all...] |
/frameworks/native/vulkan/libvulkan/ |
api.cpp | 54 allocator_(allocator), 63 allocator_.pfnFree(allocator_.pUserData, names_); 64 allocator_.pfnFree(allocator_.pUserData, implicit_layers_.elements); 65 allocator_.pfnFree(allocator_.pUserData, implicit_layers_.name_pool); 248 new_mem = allocator_.pfnReallocation( 249 allocator_.pUserData, arr.elements, 272 new_mem = allocator_.pfnReallocation 315 const VkAllocationCallbacks& allocator_; member in class:vulkan::api::__anon46623::OverrideLayerNames 381 const VkAllocationCallbacks& allocator_; member in class:vulkan::api::__anon46623::OverrideExtensionNames 484 const VkAllocationCallbacks& allocator_; member in class:vulkan::api::__anon46623::LayerChain [all...] |
/external/tensorflow/tensorflow/core/common_runtime/gpu/ |
pool_allocator.cc | 41 allocator_(allocator), 130 void* ptr = allocator_->Alloc(kPoolAlignment, num_bytes); 146 allocator_->Free(cp, cp->num_bytes); 169 allocator_->Free(pr->ptr, pr->num_bytes); 226 allocator_->Free(prec->ptr, prec->num_bytes);
|
/external/webrtc/webrtc/p2p/client/ |
portallocator_unittest.cc | 105 allocator_.reset(new cricket::BasicPortAllocator( 109 allocator_->set_step_delay(cricket::kMinimumStepDelay); 139 return allocator_->SetPortRange(min_port, max_port); 143 allocator_.reset(new cricket::BasicPortAllocator(&network_manager_)); 144 allocator_->set_step_delay(cricket::kMinimumStepDelay); 175 allocator_->AddTurnServer(turn_server); 206 allocator_->CreateSession( 324 return *allocator_; 341 for (size_t i = 0; i < allocator_->turn_servers().size(); ++i) { 342 cricket::RelayServerConfig server_config = allocator_->turn_servers()[i] 384 rtc::scoped_ptr<cricket::BasicPortAllocator> allocator_; member in class:PortAllocatorTest [all...] |
/art/libartbase/base/ |
bit_vector.h | 110 allocator_(other.allocator_), 289 Allocator* const allocator_; // Allocator if expandable. member in class:art::BitVector
|
/external/libchrome/base/containers/ |
stack_container.h | 147 StackContainer() : allocator_(&stack_data_), container_(allocator_) { 178 Allocator allocator_; member in class:base::StackContainer
|
/external/tensorflow/tensorflow/compiler/xla/service/ |
shaped_buffer.cc | 125 allocator_(allocator) {} 129 : ShapedBuffer(std::move(shaped_buffer)), allocator_(allocator) {} 142 this->allocator_->Deallocate(this->device_ordinal(), &memory_base));
|
/external/v8/src/ |
splay-tree.h | 39 : root_(NULL), allocator_(allocator) {} 54 AllocationPolicy allocator() { return allocator_; } 195 AllocationPolicy allocator_; member in class:v8::internal::SplayTree
|
string-stream.h | 108 : allocator_(allocator), 112 buffer_(allocator_->allocate(kInitialCapacity)) { 170 StringAllocator* allocator_; member in class:v8::internal::final
|
splay-tree-inl.h | 26 root_ = new(allocator_) Node(key, Config::NoValue()); 38 Node* node = new(allocator_) Node(key, Config::NoValue()); 281 List<Node*, Allocator> nodes_to_visit(10, allocator_); 282 nodes_to_visit.Add(root_, allocator_); 286 if (node->left() != NULL) nodes_to_visit.Add(node->left(), allocator_); 287 if (node->right() != NULL) nodes_to_visit.Add(node->right(), allocator_);
|
/external/libchrome/base/metrics/ |
sparse_histogram_unittest.cc | 42 if (allocator_) { 43 ASSERT_FALSE(allocator_->IsFull()); 44 ASSERT_FALSE(allocator_->IsCorrupt()); 67 allocator_ = GlobalHistogramAllocator::Get()->memory_allocator(); 71 allocator_ = nullptr; 82 PersistentMemoryAllocator* allocator_ = nullptr; member in class:base::SparseHistogramTest
|
/external/gemmlowp/internal/ |
pack.h | 53 : allocator_(allocator), pos_(0) { 56 allocator_->Reserve<std::uint8_t>(params_.l2_width * params_.l2_depth); 58 allocator_->Reserve<std::int32_t>(params_.l2_width); 76 return allocator_->GetPointer<std::uint8_t>(data_handle_) + pos_; 80 return allocator_->GetPointer<std::uint8_t>(data_handle_) + pos_; 84 return allocator_->GetPointer<std::int32_t>(sums_of_each_slice_handle_); 88 return allocator_->GetPointer<const std::int32_t>( 103 Allocator* const allocator_; member in class:gemmlowp::PackedSideBlock
|
/external/tensorflow/tensorflow/compiler/tf2xla/ |
xla_compilation_device.cc | 84 allocator_(new XlaCompilationAllocator()) {} 89 return allocator_.get();
|
/external/v8/src/compiler/ |
zone-stats.cc | 64 : max_allocated_bytes_(0), total_deleted_bytes_(0), allocator_(allocator) {} 88 Zone* zone = new Zone(allocator_, zone_name);
|
/hardware/qcom/display/msm8909/gralloc/ |
gr_buf_mgr.cpp | 53 allocator_ = new Allocator(); 54 allocator_->Init(); 79 if (allocator_) { 80 delete allocator_; 108 descriptor->SetColorFormat(allocator_->GetImplDefinedFormat(descriptor->GetProducerUsage(), 117 shared = allocator_->CheckForBufferSharing(num_descriptors, descriptors, &max_buf_index); 202 if (allocator_->FreeBuffer(reinterpret_cast<void *>(hnd->base), hnd->size, hnd->offset, 208 if (allocator_->FreeBuffer(reinterpret_cast<void *>(hnd->base_metadata), meta_size, 231 int ion_handle = allocator_->ImportBuffer(hnd->fd); 236 int ion_handle_meta = allocator_->ImportBuffer(hnd->fd_metadata) [all...] |
/hardware/qcom/display/msm8909w_3100/libgralloc1/ |
gr_buf_mgr.cpp | 53 allocator_ = new Allocator(); 54 allocator_->Init(); 79 if (allocator_) { 80 delete allocator_; 108 descriptor->SetColorFormat(allocator_->GetImplDefinedFormat(descriptor->GetProducerUsage(), 117 shared = allocator_->CheckForBufferSharing(num_descriptors, descriptors, &max_buf_index); 196 if (allocator_->FreeBuffer(reinterpret_cast<void *>(hnd->base), hnd->size, hnd->offset, 202 if (allocator_->FreeBuffer(reinterpret_cast<void *>(hnd->base_metadata), meta_size, 225 int ion_handle = allocator_->ImportBuffer(hnd->fd); 230 int ion_handle_meta = allocator_->ImportBuffer(hnd->fd_metadata) [all...] |