HomeSort by relevance Sort by last modified time
    Searched defs:batch (Results 101 - 125 of 256) sorted by null

1 2 3 45 6 7 8 91011

  /external/tensorflow/tensorflow/python/data/ops/
dataset_ops.py 728 def batch(self, batch_size): member in class:Dataset
732 multiple of `batch_size`, the final batch contain smaller tensors with
733 shape `N % batch_size` in the batch dimension. If your program depends on
739 consecutive elements of this dataset to combine in a single batch.
757 consecutive elements of this dataset to combine in a single batch.
764 dimension in each batch.
    [all...]
  /frameworks/rs/cpu_ref/
rsCpuScriptGroup2.cpp 91 Batch::Batch(CpuScriptGroup2Impl* group, const char* name) :
96 Batch::~Batch() {
103 bool Batch::conflict(CPUClosure* cpuClosure) const {
111 // An invoke should be in a batch by itself, so it conflicts with any other
164 Batch* batch = new Batch(this, "Batch0"); local
179 if (batch->conflict(cc))
    [all...]
  /hardware/interfaces/camera/device/3.4/default/
CameraDeviceSession.cpp 585 std::shared_ptr<InflightBatch> batch = pair.second; local
587 Mutex::Autolock _l(batch->mLock);
588 // Check if the batch is removed (mostly by notify error) before lock was acquired
589 if (batch->mRemoved) {
590 // Fall back to non-batch path
598 batch->mResultMds[result.v3_2.partialResult].mMds.push_back(
606 auto it = batch->mBatchBufs.find(buffer.streamId);
607 if (it != batch->mBatchBufs.end()) {
631 if (result.v3_2.frameNumber == batch->mLastFrame) {
634 sendBatchMetadataLocked(batch, result.v3_2.partialResult)
    [all...]
  /hardware/libhardware/modules/sensors/dynamic_sensor/
HidRawSensor.cpp 895 int HidRawSensor::batch(int64_t samplingPeriod, int64_t batchingPeriod) { function in class:android::SensorHalExt::HidRawSensor
    [all...]
  /libcore/xml/src/main/java/org/kxml2/io/
KXmlSerializer.java 63 int batch = BUFFER_LEN - mPos; local
64 if (batch > length) {
65 batch = length;
67 str.getChars(i, i + batch, mText, mPos);
68 i += batch;
69 length -= batch;
70 mPos += batch;
  /external/mesa3d/src/intel/vulkan/
anv_batch_chain.c 40 * the actual batch buffers as well as link them together and handle
196 anv_batch_emit_dwords(struct anv_batch *batch, int num_dwords)
198 if (batch->next + num_dwords * 4 > batch->end)
199 batch->extend_cb(batch, batch->user_data);
201 void *p = batch->next;
203 batch->next += num_dwords * 4;
204 assert(batch->next <= batch->end)
472 struct anv_batch *batch = &cmd_buffer->batch; local
1230 struct anv_batch *batch = &cmd_buffer->batch; local
    [all...]
anv_blorp.c 213 struct blorp_batch batch; local
214 blorp_batch_init(&cmd_buffer->device->blorp, &batch, cmd_buffer, 0);
257 blorp_copy(&batch, &src_surf, pRegions[r].srcSubresource.mipLevel,
268 blorp_batch_finish(&batch);
279 struct blorp_batch batch; local
280 blorp_batch_init(&cmd_buffer->device->blorp, &batch, cmd_buffer, 0);
346 blorp_copy(&batch, &src->surf, src->level, src->offset.z,
356 blorp_batch_finish(&batch);
441 struct blorp_batch batch; local
442 blorp_batch_init(&cmd_buffer->device->blorp, &batch, cmd_buffer, 0)
614 struct blorp_batch batch; local
675 struct blorp_batch batch; local
724 struct blorp_batch batch; local
806 struct blorp_batch batch; local
858 struct blorp_batch batch; local
1043 struct blorp_batch batch; local
1178 struct blorp_batch batch; local
1379 struct blorp_batch batch; local
1550 struct blorp_batch batch; local
1622 struct blorp_batch batch; local
    [all...]
anv_device.c 811 struct anv_batch *batch)
821 /* Kernel driver requires 8 byte aligned batch length */
822 size = align_u32(batch->next - batch->start, 8);
827 memcpy(bo.map, batch->start, size);
1228 struct anv_batch batch; local
1557 struct anv_batch batch; local
    [all...]
  /external/tensorflow/tensorflow/contrib/labeled_tensor/
__init__.py 133 batch = _ops.batch variable
  /external/tensorflow/tensorflow/core/kernels/
conv_ops.cc 136 const Tensor& filter, int batch, int input_rows,
152 const Tensor& filter, int batch, int input_rows,
166 args.batch = batch;
193 const Tensor& filter, int batch, int input_rows,
207 const Tensor& filter, int batch, int input_rows,
217 desc.N = batch;
282 "strides in the batch and depth dimensions."));
294 "dilations in the batch and depth dimensions."));
303 // [ batch, in_rows, in_cols, in_depth
362 const int batch = static_cast<int>(batch_raw); variable
    [all...]
mkl_lrn_op.cc 286 const int batch = static_cast<int>(input.dim_size(0)); local
292 auto in_shaped = input.shaped<T, 2>({nodes * batch, depth});
311 auto out_shaped = output->shaped<T, 2>({nodes * batch, depth});
658 const int64 batch = static_cast<int64>(in_grads.dim_size(0)); local
664 auto grads_shaped = in_grads.shaped<T, 2>({nodes * batch, depth});
666 auto in_shaped = in_image.shaped<T, 2>({nodes * batch, depth});
667 auto activations = out_image.shaped<T, 2>({nodes * batch, depth});
676 auto out_shaped = output->shaped<T, 2>({nodes * batch, depth});
706 Shard(worker_threads.num_threads, worker_threads.workers, nodes * batch,
    [all...]
pooling_ops_3d_sycl.h 33 SYCL3DPoolParams(const int depth, const int batch, const int in_planes,
40 batch_(batch),
57 SYCL3DPoolParams(const int depth, const int batch, const int in_planes,
63 : SYCL3DPoolParams(depth, batch, in_planes, in_rows, in_cols,
124 MaxPool3DSYCL(const int depth, const int batch, const int in_planes,
132 : p_(depth, batch, in_planes, in_rows, in_cols, out_planes, out_rows,
189 const int batch = GetTensorDim(tensor_in, data_format, 'N'); local
207 MaxPool3DSYCL<T> max_pool(depth, batch, in_planes, in_rows, in_cols,
235 MaxPool3DGradSYCL(const int depth, const int batch, const int in_planes,
245 : p_(depth, batch, in_planes, in_rows, in_cols, output_shape, window
355 const int batch = GetTensorDim(tensor_in, data_format, 'N'); local
595 const int batch = GetTensorDim(tensor_in, data_format, 'N'); local
727 const int batch = GetTensorDim(tensor_in_shape, data_format, 'N'); local
    [all...]
quantized_conv_ops.cc 102 for (int batch = 0; batch < input_batches; ++batch) {
144 input_data[(batch * input_height * input_width *
176 output_data[(batch * output_height * output_width * filter_count) +
302 const int64 batch = patch_index / (output_height * output_width); local
306 input_data + (batch * input_height * input_width * input_depth);
465 "strides in the batch and depth dimensions."));
478 "dilations in the batch and depth dimensions."));
484 // [ batch, in_rows, in_cols, in_depth
533 const int64 batch = input.dim_size(0); variable
    [all...]
  /frameworks/native/libs/input/
InputTransport.cpp 428 ALOGD("channel '%s' consumer ~ consumed batch event, seq=%u",
456 Batch& batch = mBatches.editItemAt(batchIndex); local
457 if (canAddSample(batch, &mMsg)) {
458 batch.samples.push(mMsg);
460 ALOGD("channel '%s' consumer ~ appended to batch event",
465 // We cannot append to the batch in progress, so we need to consume
466 // the previous batch right now and defer the new message until later.
469 batch, batch.samples.size(), outSeq, outEvent, displayId)
487 Batch& batch = mBatches.editTop(); local
525 Batch& batch = mBatches.editItemAt(i); local
908 const Batch& batch = mBatches.itemAt(i); local
    [all...]
  /hardware/libhardware/modules/sensors/
multihal.cpp 198 int batch(int handle, int flags, int64_t period_ns, int64_t timeout);
403 int sensors_poll_context_t::batch(int handle, int flags, int64_t period_ns, int64_t timeout) { function in class:sensors_poll_context_t
404 ALOGV("batch");
409 retval = v1->batch(v1, local_handle, flags, period_ns, timeout);
411 ALOGE("IGNORING batch() call to non-API-compliant sensor handle=%d !", handle);
558 return ctx->batch(handle, flags, period_ns, timeout);
803 dev->proxy_device.batch = device__batch;
  /libcore/ojluni/src/main/java/java/util/
LinkedList.java 1181 int batch; \/\/ batch size for splits field in class:LinkedList.LLSpliterator
    [all...]
Spliterators.java 1261 private int batch; \/\/ batch size for splits field in class:Spliterators.AbstractSpliterator
1384 private int batch; \/\/ batch size for splits field in class:Spliterators.AbstractIntSpliterator
1494 private int batch; \/\/ batch size for splits field in class:Spliterators.AbstractLongSpliterator
1604 private int batch; \/\/ batch size for splits field in class:Spliterators.AbstractDoubleSpliterator
1698 private int batch; \/\/ batch size for splits field in class:Spliterators.IteratorSpliterator
1849 private int batch; \/\/ batch size for splits field in class:Spliterators.IntIteratorSpliterator
1943 private int batch; \/\/ batch size for splits field in class:Spliterators.LongIteratorSpliterator
2037 private int batch; \/\/ batch size for splits field in class:Spliterators.DoubleIteratorSpliterator
    [all...]
  /libcore/ojluni/src/main/java/java/util/concurrent/
ConcurrentLinkedQueue.java 816 static final int MAX_BATCH = 1 << 25; // max batch array size;
819 int batch; // batch size for splits field in class:ConcurrentLinkedQueue.CLQSpliterator
828 int b = batch;
844 batch = i;
    [all...]
LinkedBlockingQueue.java 830 static final int MAX_BATCH = 1 << 25; // max batch array size;
833 int batch; // batch size for splits field in class:LinkedBlockingQueue.LBQSpliterator
846 int b = batch;
872 batch = i;
    [all...]
  /packages/apps/Bluetooth/src/com/android/bluetooth/opp/
BluetoothOppService.java 749 * Add info into a batch. The logic is
751 * 2) If there is no batch, create a batch and insert this transfer into batch,
752 * then run the batch
753 * 3) If there is existing batch and timestamp match, insert transfer into batch
754 * 4) If there is existing batch and timestamp does not match, create a new batch and
778 "Service create new Batch " + newBatch.mId + " for OUTBOUND info
886 BluetoothOppBatch batch = mBatches.get(i); local
894 BluetoothOppBatch batch = mBatches.get(i); local
941 BluetoothOppBatch batch = mBatches.get(i); local
    [all...]
  /external/libdrm/intel/
intel_bufmgr_fake.c 1436 struct drm_i915_batchbuffer batch; local
    [all...]
  /external/mesa3d/src/gallium/drivers/freedreno/
freedreno_context.h 159 /* Current batch.. the rule here is that you can deref ctx->batch
162 * called from GMEM code), since in those code-paths the batch
163 * you care about is not necessarily the same as ctx->batch.
165 struct fd_batch *batch; member in struct:fd_context
191 * (NOTE: this is kind of related to the batch, but moving it there
192 * means we'd always have to recalc tiles ever batch)
255 void (*emit_tile_init)(struct fd_batch *batch);
256 void (*emit_tile_prep)(struct fd_batch *batch, struct fd_tile *tile);
257 void (*emit_tile_mem2gmem)(struct fd_batch *batch, struct fd_tile *tile)
    [all...]
  /external/mesa3d/src/mesa/drivers/dri/i965/
intel_screen.c 893 /* Once a batch uses more than 75% of the maximum mappable size, we
1287 uint32_t *batch; local
    [all...]
  /external/swiftshader/src/Renderer/
Renderer.cpp 251 int batch = batchSize / ms; local
253 int (Renderer::*setupPrimitives)(int batch, int count);
264 batch = 1;
268 batch = 1;
323 draw->batchSize = batch;
650 draw->references = (count + batch - 1) / batch;
792 int batch = draw->batchSize; local
796 primitiveProgress[unit].primitiveCount = count - primitive >= batch ? batch : count - primitive
1088 unsigned int batch[128][3]; \/\/ FIXME: Adjust to dynamic batch size local
    [all...]
  /external/tensorflow/tensorflow/compiler/xla/service/cpu/
ir_emitter.cc 1008 llvm::Value* batch = index[dnums.output_batch_dimension()]; local
    [all...]

Completed in 2493 milliseconds

1 2 3 45 6 7 8 91011