/external/tensorflow/tensorflow/python/data/ops/ |
dataset_ops.py | 728 def batch(self, batch_size): member in class:Dataset 732 multiple of `batch_size`, the final batch contain smaller tensors with 733 shape `N % batch_size` in the batch dimension. If your program depends on 739 consecutive elements of this dataset to combine in a single batch. 757 consecutive elements of this dataset to combine in a single batch. 764 dimension in each batch. [all...] |
/frameworks/rs/cpu_ref/ |
rsCpuScriptGroup2.cpp | 91 Batch::Batch(CpuScriptGroup2Impl* group, const char* name) : 96 Batch::~Batch() { 103 bool Batch::conflict(CPUClosure* cpuClosure) const { 111 // An invoke should be in a batch by itself, so it conflicts with any other 164 Batch* batch = new Batch(this, "Batch0"); local 179 if (batch->conflict(cc)) [all...] |
/hardware/interfaces/camera/device/3.4/default/ |
CameraDeviceSession.cpp | 585 std::shared_ptr<InflightBatch> batch = pair.second; local 587 Mutex::Autolock _l(batch->mLock); 588 // Check if the batch is removed (mostly by notify error) before lock was acquired 589 if (batch->mRemoved) { 590 // Fall back to non-batch path 598 batch->mResultMds[result.v3_2.partialResult].mMds.push_back( 606 auto it = batch->mBatchBufs.find(buffer.streamId); 607 if (it != batch->mBatchBufs.end()) { 631 if (result.v3_2.frameNumber == batch->mLastFrame) { 634 sendBatchMetadataLocked(batch, result.v3_2.partialResult) [all...] |
/hardware/libhardware/modules/sensors/dynamic_sensor/ |
HidRawSensor.cpp | 895 int HidRawSensor::batch(int64_t samplingPeriod, int64_t batchingPeriod) { function in class:android::SensorHalExt::HidRawSensor [all...] |
/libcore/xml/src/main/java/org/kxml2/io/ |
KXmlSerializer.java | 63 int batch = BUFFER_LEN - mPos; local 64 if (batch > length) { 65 batch = length; 67 str.getChars(i, i + batch, mText, mPos); 68 i += batch; 69 length -= batch; 70 mPos += batch;
|
/external/mesa3d/src/intel/vulkan/ |
anv_batch_chain.c | 40 * the actual batch buffers as well as link them together and handle 196 anv_batch_emit_dwords(struct anv_batch *batch, int num_dwords) 198 if (batch->next + num_dwords * 4 > batch->end) 199 batch->extend_cb(batch, batch->user_data); 201 void *p = batch->next; 203 batch->next += num_dwords * 4; 204 assert(batch->next <= batch->end) 472 struct anv_batch *batch = &cmd_buffer->batch; local 1230 struct anv_batch *batch = &cmd_buffer->batch; local [all...] |
anv_blorp.c | 213 struct blorp_batch batch; local 214 blorp_batch_init(&cmd_buffer->device->blorp, &batch, cmd_buffer, 0); 257 blorp_copy(&batch, &src_surf, pRegions[r].srcSubresource.mipLevel, 268 blorp_batch_finish(&batch); 279 struct blorp_batch batch; local 280 blorp_batch_init(&cmd_buffer->device->blorp, &batch, cmd_buffer, 0); 346 blorp_copy(&batch, &src->surf, src->level, src->offset.z, 356 blorp_batch_finish(&batch); 441 struct blorp_batch batch; local 442 blorp_batch_init(&cmd_buffer->device->blorp, &batch, cmd_buffer, 0) 614 struct blorp_batch batch; local 675 struct blorp_batch batch; local 724 struct blorp_batch batch; local 806 struct blorp_batch batch; local 858 struct blorp_batch batch; local 1043 struct blorp_batch batch; local 1178 struct blorp_batch batch; local 1379 struct blorp_batch batch; local 1550 struct blorp_batch batch; local 1622 struct blorp_batch batch; local [all...] |
anv_device.c | 811 struct anv_batch *batch) 821 /* Kernel driver requires 8 byte aligned batch length */ 822 size = align_u32(batch->next - batch->start, 8); 827 memcpy(bo.map, batch->start, size); 1228 struct anv_batch batch; local 1557 struct anv_batch batch; local [all...] |
/external/tensorflow/tensorflow/contrib/labeled_tensor/ |
__init__.py | 133 batch = _ops.batch variable
|
/external/tensorflow/tensorflow/core/kernels/ |
conv_ops.cc | 136 const Tensor& filter, int batch, int input_rows, 152 const Tensor& filter, int batch, int input_rows, 166 args.batch = batch; 193 const Tensor& filter, int batch, int input_rows, 207 const Tensor& filter, int batch, int input_rows, 217 desc.N = batch; 282 "strides in the batch and depth dimensions.")); 294 "dilations in the batch and depth dimensions.")); 303 // [ batch, in_rows, in_cols, in_depth 362 const int batch = static_cast<int>(batch_raw); variable [all...] |
mkl_lrn_op.cc | 286 const int batch = static_cast<int>(input.dim_size(0)); local 292 auto in_shaped = input.shaped<T, 2>({nodes * batch, depth}); 311 auto out_shaped = output->shaped<T, 2>({nodes * batch, depth}); 658 const int64 batch = static_cast<int64>(in_grads.dim_size(0)); local 664 auto grads_shaped = in_grads.shaped<T, 2>({nodes * batch, depth}); 666 auto in_shaped = in_image.shaped<T, 2>({nodes * batch, depth}); 667 auto activations = out_image.shaped<T, 2>({nodes * batch, depth}); 676 auto out_shaped = output->shaped<T, 2>({nodes * batch, depth}); 706 Shard(worker_threads.num_threads, worker_threads.workers, nodes * batch, [all...] |
pooling_ops_3d_sycl.h | 33 SYCL3DPoolParams(const int depth, const int batch, const int in_planes, 40 batch_(batch), 57 SYCL3DPoolParams(const int depth, const int batch, const int in_planes, 63 : SYCL3DPoolParams(depth, batch, in_planes, in_rows, in_cols, 124 MaxPool3DSYCL(const int depth, const int batch, const int in_planes, 132 : p_(depth, batch, in_planes, in_rows, in_cols, out_planes, out_rows, 189 const int batch = GetTensorDim(tensor_in, data_format, 'N'); local 207 MaxPool3DSYCL<T> max_pool(depth, batch, in_planes, in_rows, in_cols, 235 MaxPool3DGradSYCL(const int depth, const int batch, const int in_planes, 245 : p_(depth, batch, in_planes, in_rows, in_cols, output_shape, window 355 const int batch = GetTensorDim(tensor_in, data_format, 'N'); local 595 const int batch = GetTensorDim(tensor_in, data_format, 'N'); local 727 const int batch = GetTensorDim(tensor_in_shape, data_format, 'N'); local [all...] |
quantized_conv_ops.cc | 102 for (int batch = 0; batch < input_batches; ++batch) { 144 input_data[(batch * input_height * input_width * 176 output_data[(batch * output_height * output_width * filter_count) + 302 const int64 batch = patch_index / (output_height * output_width); local 306 input_data + (batch * input_height * input_width * input_depth); 465 "strides in the batch and depth dimensions.")); 478 "dilations in the batch and depth dimensions.")); 484 // [ batch, in_rows, in_cols, in_depth 533 const int64 batch = input.dim_size(0); variable [all...] |
/frameworks/native/libs/input/ |
InputTransport.cpp | 428 ALOGD("channel '%s' consumer ~ consumed batch event, seq=%u", 456 Batch& batch = mBatches.editItemAt(batchIndex); local 457 if (canAddSample(batch, &mMsg)) { 458 batch.samples.push(mMsg); 460 ALOGD("channel '%s' consumer ~ appended to batch event", 465 // We cannot append to the batch in progress, so we need to consume 466 // the previous batch right now and defer the new message until later. 469 batch, batch.samples.size(), outSeq, outEvent, displayId) 487 Batch& batch = mBatches.editTop(); local 525 Batch& batch = mBatches.editItemAt(i); local 908 const Batch& batch = mBatches.itemAt(i); local [all...] |
/hardware/libhardware/modules/sensors/ |
multihal.cpp | 198 int batch(int handle, int flags, int64_t period_ns, int64_t timeout); 403 int sensors_poll_context_t::batch(int handle, int flags, int64_t period_ns, int64_t timeout) { function in class:sensors_poll_context_t 404 ALOGV("batch"); 409 retval = v1->batch(v1, local_handle, flags, period_ns, timeout); 411 ALOGE("IGNORING batch() call to non-API-compliant sensor handle=%d !", handle); 558 return ctx->batch(handle, flags, period_ns, timeout); 803 dev->proxy_device.batch = device__batch;
|
/libcore/ojluni/src/main/java/java/util/ |
LinkedList.java | 1181 int batch; \/\/ batch size for splits field in class:LinkedList.LLSpliterator [all...] |
Spliterators.java | 1261 private int batch; \/\/ batch size for splits field in class:Spliterators.AbstractSpliterator 1384 private int batch; \/\/ batch size for splits field in class:Spliterators.AbstractIntSpliterator 1494 private int batch; \/\/ batch size for splits field in class:Spliterators.AbstractLongSpliterator 1604 private int batch; \/\/ batch size for splits field in class:Spliterators.AbstractDoubleSpliterator 1698 private int batch; \/\/ batch size for splits field in class:Spliterators.IteratorSpliterator 1849 private int batch; \/\/ batch size for splits field in class:Spliterators.IntIteratorSpliterator 1943 private int batch; \/\/ batch size for splits field in class:Spliterators.LongIteratorSpliterator 2037 private int batch; \/\/ batch size for splits field in class:Spliterators.DoubleIteratorSpliterator [all...] |
/libcore/ojluni/src/main/java/java/util/concurrent/ |
ConcurrentLinkedQueue.java | 816 static final int MAX_BATCH = 1 << 25; // max batch array size; 819 int batch; // batch size for splits field in class:ConcurrentLinkedQueue.CLQSpliterator 828 int b = batch; 844 batch = i; [all...] |
LinkedBlockingQueue.java | 830 static final int MAX_BATCH = 1 << 25; // max batch array size; 833 int batch; // batch size for splits field in class:LinkedBlockingQueue.LBQSpliterator 846 int b = batch; 872 batch = i; [all...] |
/packages/apps/Bluetooth/src/com/android/bluetooth/opp/ |
BluetoothOppService.java | 749 * Add info into a batch. The logic is 751 * 2) If there is no batch, create a batch and insert this transfer into batch, 752 * then run the batch 753 * 3) If there is existing batch and timestamp match, insert transfer into batch 754 * 4) If there is existing batch and timestamp does not match, create a new batch and 778 "Service create new Batch " + newBatch.mId + " for OUTBOUND info 886 BluetoothOppBatch batch = mBatches.get(i); local 894 BluetoothOppBatch batch = mBatches.get(i); local 941 BluetoothOppBatch batch = mBatches.get(i); local [all...] |
/external/libdrm/intel/ |
intel_bufmgr_fake.c | 1436 struct drm_i915_batchbuffer batch; local [all...] |
/external/mesa3d/src/gallium/drivers/freedreno/ |
freedreno_context.h | 159 /* Current batch.. the rule here is that you can deref ctx->batch 162 * called from GMEM code), since in those code-paths the batch 163 * you care about is not necessarily the same as ctx->batch. 165 struct fd_batch *batch; member in struct:fd_context 191 * (NOTE: this is kind of related to the batch, but moving it there 192 * means we'd always have to recalc tiles ever batch) 255 void (*emit_tile_init)(struct fd_batch *batch); 256 void (*emit_tile_prep)(struct fd_batch *batch, struct fd_tile *tile); 257 void (*emit_tile_mem2gmem)(struct fd_batch *batch, struct fd_tile *tile) [all...] |
/external/mesa3d/src/mesa/drivers/dri/i965/ |
intel_screen.c | 893 /* Once a batch uses more than 75% of the maximum mappable size, we 1287 uint32_t *batch; local [all...] |
/external/swiftshader/src/Renderer/ |
Renderer.cpp | 251 int batch = batchSize / ms; local 253 int (Renderer::*setupPrimitives)(int batch, int count); 264 batch = 1; 268 batch = 1; 323 draw->batchSize = batch; 650 draw->references = (count + batch - 1) / batch; 792 int batch = draw->batchSize; local 796 primitiveProgress[unit].primitiveCount = count - primitive >= batch ? batch : count - primitive 1088 unsigned int batch[128][3]; \/\/ FIXME: Adjust to dynamic batch size local [all...] |
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
ir_emitter.cc | 1008 llvm::Value* batch = index[dnums.output_batch_dimension()]; local [all...] |