HomeSort by relevance Sort by last modified time
    Searched refs:AddScalar (Results 1 - 24 of 24) sorted by null

  /external/libmojo/base/trace_event/
java_heap_dump_provider_android.cc 35 outer_dump->AddScalar(MemoryAllocatorDump::kNameSize,
40 inner_dump->AddScalar(MemoryAllocatorDump::kNameSize,
  /external/libchrome/base/trace_event/
memory_allocator_dump_unittest.cc 32 root_heap->AddScalar(MemoryAllocatorDump::kNameSize,
34 root_heap->AddScalar(MemoryAllocatorDump::kNameObjectCount,
36 root_heap->AddScalar("attr1", "units1", 1234);
42 sub_heap->AddScalar(MemoryAllocatorDump::kNameSize,
44 sub_heap->AddScalar(MemoryAllocatorDump::kNameObjectCount,
179 dump->AddScalar(MemoryAllocatorDump::kNameSize,
181 dump->AddScalar("foo", MemoryAllocatorDump::kUnitsBytes, 2);
trace_event_memory_overhead.cc 145 mad->AddScalar(MemoryAllocatorDump::kNameSize,
148 mad->AddScalar("resident_size", MemoryAllocatorDump::kUnitsBytes,
150 mad->AddScalar(MemoryAllocatorDump::kNameObjectCount,
memory_allocator_dump.h 44 // Standard attribute |name|s for the AddScalar and AddString() methods.
48 // Standard attribute |unit|s for the AddScalar and AddString() methods.
58 // AddScalar(kNameSize, kUnitsBytes, 1234);
60 // AddScalar("number_of_freelist_entires", kUnitsObjects, 42)
63 void AddScalar(const char* name, const char* units, uint64_t value);
94 // The size is the value set with AddScalar(kNameSize, kUnitsBytes, size);
106 // temporary strings on each AddScalar() call.
malloc_dump_provider.cc 258 outer_dump->AddScalar("virtual_size", MemoryAllocatorDump::kUnitsBytes,
260 outer_dump->AddScalar(MemoryAllocatorDump::kNameSize,
264 inner_dump->AddScalar(MemoryAllocatorDump::kNameSize,
268 inner_dump->AddScalar(MemoryAllocatorDump::kNameObjectCount,
279 other_dump->AddScalar(MemoryAllocatorDump::kNameSize,
memory_allocator_dump.cc 60 void MemoryAllocatorDump::AddScalar(const char* name,
  /external/libchrome/base/memory/
shared_memory_tracker.cc 68 local_dump->AddScalar(trace_event::MemoryAllocatorDump::kNameSize,
73 global_dump->AddScalar(trace_event::MemoryAllocatorDump::kNameSize,
  /external/tensorflow/tensorflow/core/kernels/data/
range_dataset_op.cc 79 TF_RETURN_IF_ERROR(b->AddScalar(start_, &start));
80 TF_RETURN_IF_ERROR(b->AddScalar(stop_, &stop));
81 TF_RETURN_IF_ERROR(b->AddScalar(step_, &step));
random_dataset_op.cc 84 TF_RETURN_IF_ERROR(b->AddScalar(seed_, &seed));
85 TF_RETURN_IF_ERROR(b->AddScalar(seed2_, &seed2));
shuffle_dataset_op.cc 418 TF_RETURN_IF_ERROR(b->AddScalar(buffer_size_, &buffer_size));
419 TF_RETURN_IF_ERROR(b->AddScalar(seed_, &seed));
420 TF_RETURN_IF_ERROR(b->AddScalar(seed2_, &seed2));
502 TF_RETURN_IF_ERROR(b->AddScalar(buffer_size_, &buffer_size));
503 TF_RETURN_IF_ERROR(b->AddScalar(seed_, &seed));
504 TF_RETURN_IF_ERROR(b->AddScalar(seed2_, &seed2));
505 TF_RETURN_IF_ERROR(b->AddScalar(count_, &count));
reader_dataset_ops.cc 118 TF_RETURN_IF_ERROR(b->AddScalar(compression_type_, &compression_type));
120 b->AddScalar(options_.input_buffer_size, &buffer_size));
356 TF_RETURN_IF_ERROR(b->AddScalar(header_bytes_, &header_bytes));
357 TF_RETURN_IF_ERROR(b->AddScalar(record_bytes_, &record_bytes));
358 TF_RETURN_IF_ERROR(b->AddScalar(footer_bytes_, &footer_bytes));
359 TF_RETURN_IF_ERROR(b->AddScalar(buffer_size_, &buffer_size));
571 TF_RETURN_IF_ERROR(b->AddScalar(compression_type_, &compression_type));
573 TF_RETURN_IF_ERROR(b->AddScalar(options_.buffer_size, &buffer_size));
stats_dataset_ops.cc 80 TF_RETURN_IF_ERROR(b->AddScalar(tag_, &tag_node));
176 TF_RETURN_IF_ERROR(b->AddScalar(tag_, &tag_node));
skip_dataset_op.cc 79 TF_RETURN_IF_ERROR(b->AddScalar(count_, &count));
take_dataset_op.cc 80 TF_RETURN_IF_ERROR(b->AddScalar(count_, &count));
interleave_dataset_op.cc 121 TF_RETURN_IF_ERROR(b->AddScalar(cycle_length_, &cycle_length_node));
123 TF_RETURN_IF_ERROR(b->AddScalar(block_length_, &block_length_node));
batch_dataset_op.cc 88 TF_RETURN_IF_ERROR(b->AddScalar(batch_size_, &batch_size));
repeat_dataset_op.cc 80 TF_RETURN_IF_ERROR(b->AddScalar(count_, &count));
dense_to_sparse_batch_dataset_op.cc 123 TF_RETURN_IF_ERROR(b->AddScalar(batch_size_, &batch_size_node));
prefetch_dataset_op.cc 77 TF_RETURN_IF_ERROR(b->AddScalar(buffer_size_, &buffer_size));
padded_batch_dataset_op.cc 147 TF_RETURN_IF_ERROR(b->AddScalar(batch_size_, &batch_size));
parallel_map_dataset_op.cc 126 b->AddScalar(num_parallel_calls_, &num_parallel_calls));
tensor_queue_dataset_op.cc 107 TF_RETURN_IF_ERROR(b->AddScalar(batch_size_, &batch_size));
  /external/tensorflow/tensorflow/contrib/kafka/kernels/
kafka_dataset_ops.cc 94 TF_RETURN_IF_ERROR(b->AddScalar(servers_, &servers));
96 TF_RETURN_IF_ERROR(b->AddScalar(group_, &group));
98 TF_RETURN_IF_ERROR(b->AddScalar(eof_, &eof));
100 TF_RETURN_IF_ERROR(b->AddScalar(timeout_, &timeout));
  /external/tensorflow/tensorflow/core/framework/
dataset.h 81 Status AddScalar(const T& val, Node** output) {
86 return errors::Internal("AddScalar: Failed to build Const op.");

Completed in 230 milliseconds