/external/tensorflow/tensorflow/core/kernels/ |
bias_op.cc | 45 int32* batch, int32* height, int32* width, 47 *batch = 1; 55 *batch *= static_cast<int32>(value_tensor.dim_size(i)); 65 *batch *= static_cast<int32>(value_tensor.dim_size(i)); 134 int32 batch, height, width, channel; variable 135 GetBiasValueDims(input, data_format_, &batch, &height, &width, &channel); 137 Eigen::DSizes<int32, 4> broad_cast_dims(batch, 1, height, width); 231 int32 batch, height, width, channel; variable 232 GetBiasValueDims(output_backprop, data_format_, &batch, &height, &width, 249 Eigen::DSizes<int, 4> four_dims(batch, channel, height, width) 331 int32 batch, height, width, channel; variable 387 int32 batch, height, width, channel; variable [all...] |
bias_op_gpu.cu.cc | 70 T* output, int32 batch, int32 height, int32 width, 74 const int32 total_count = batch * bias_size * image_size; 140 T* bias_backprop, int32 batch, 156 int32 total_count = batch * image_size; 161 int32 batch = index / image_size; local 163 (batch * bias_size + bias_index) * image_size + image_offset); 189 T* bias_backprop, int32 batch, int32 height, 194 const int32 total_count = batch * bias_size * image_size; 222 output_backprop, bias_backprop, batch, bias_size, image_size,
|
cwise_op_select.cc | 237 const Eigen::DenseIndex batch = cond_vec.size(); local 242 Eigen::Tensor<Eigen::DenseIndex, 2>::Dimensions reshape_dims{{batch, 1}}; 247 reshape_dims.set(0, batch);
|
dilation_ops_gpu.cu.cc | 40 const T* filter_ptr, int batch, int input_rows, 81 const T* out_backprop_ptr, int batch, int input_rows, int input_cols, 130 const T* out_backprop_ptr, int batch, int input_rows, int input_cols, 185 const int batch = input.dimension(0); local 196 const int total_count = batch * output_rows * output_cols * depth; 201 config.virtual_thread_count, input.data(), filter.data(), batch, 216 const int batch = input.dimension(0); local 231 total_count = batch * input_rows * input_cols * depth; 237 total_count = batch * output_rows * output_cols * depth; 242 out_backprop.data(), batch, input_rows, input_cols, depth, filter_rows 256 const int batch = input.dimension(0); local [all...] |
extract_image_patches_op.cc | 63 // [ batch, in_rows, in_cols, channels ] 69 const int batch = input.dim_size(0); variable 95 const std::vector<int64> out_sizes = {batch, out_rows, out_cols,
|
resize_bilinear_op_test.cc | 67 const int batch = images.dimension(0); local 72 ASSERT_EQ(batch, output.dimension(0)); 81 for (int b = 0; b < batch; ++b) {
|
sparse_xent_op.h | 49 // the batch index b, it calculates: 70 const int batch = coords[0]; local 72 const Index label = tensorflow::internal::SubtleMustCopy(labels_(batch)); 77 ? (Eigen::numext::log(sum_exp_logits_(batch)) - logits_(coords)) 90 // indices. For each minibatch entry, ignoring the batch index b, it calculates: 110 const int batch = coords[0]; local 112 const Index label = tensorflow::internal::SubtleMustCopy(labels_(batch)); 117 return exp_logits_(coords) / sum_exp_logits_(batch) - subtract;
|
/external/tensorflow/tensorflow/core/kernels/neon/ |
neon_depthwise_conv_op.cc | 57 "strides in the batch and depth dimensions.")); 78 const int32 batch = input.dim_size(0); variable 97 TensorShape out_shape({batch, out_rows, out_cols, out_depth}); 110 << " Input: [" << batch << ", " << input_rows << ", " << input_cols 114 << ", pad_cols = " << pad_cols << ", output: [" << batch << ", " 153 // Dims in the neon kernels are channel, x, y, batch order. 164 // Dims in the neon kernels are channel, x, y, batch order. 176 // Dims in the neon kernels are channel, x, y, batch order.
|
/external/tensorflow/tensorflow/core/util/ |
tensor_format.h | 31 // N = Batch, H = Image Height, W = Image Width, C = Number of Channels. 129 // Returns the index of the batch dimension. 242 // data 'tensor_format'. 'dimension' is a char that can be 'N' (batch size), 478 const int64 batch = GetTensorDim(src_shape, src_format, 'N'); local 483 return ShapeFromFormat(dst_format, batch, 490 return ShapeFromFormat(dst_format, batch,
|
/frameworks/base/cmds/incidentd/src/ |
Reporter.h | 89 ReportRequestSet batch; member in class:android::os::incidentd::Reporter 95 // Run the report as described in the batch and args parameters.
|
/hardware/interfaces/sensors/1.0/default/ |
Sensors.cpp | 234 Return<Result> Sensors::batch( function in class:android::hardware::sensors::V1_0::implementation::Sensors 239 mSensorDevice->batch(
|
/hardware/invensense/65xx/libsensors_iio/ |
sensors_mpl.cpp | 95 int batch(int handle, int flags, int64_t period_ns, int64_t timeout); 319 int sensors_poll_context_t::batch(int handle, int flags, int64_t period_ns, function in class:sensors_poll_context_t 323 return mSensor->batch(handle, flags, period_ns, timeout); 370 return ctx->batch(handle, flags, period_ns, timeout); 399 /* Batch processing */ 400 dev->device.batch = poll__batch;
|
/developers/build/prebuilts/gradle/BasicSyncAdapter/Application/src/main/java/com/example/android/basicsyncadapter/ |
SyncAdapter.java | 189 * <p>As an additional optimization, we use a batch operation to perform all database writes at 212 ArrayList<ContentProviderOperation> batch = new ArrayList<ContentProviderOperation>(); local 252 batch.add(ContentProviderOperation.newUpdate(existingUri) 266 batch.add(ContentProviderOperation.newDelete(deleteUri).build()); 275 batch.add(ContentProviderOperation.newInsert(FeedContract.Entry.CONTENT_URI) 283 Log.i(TAG, "Merge solution ready. Applying batch update"); 284 mContentResolver.applyBatch(FeedContract.CONTENT_AUTHORITY, batch);
|
/developers/samples/android/connectivity/sync/BasicSyncAdapter/Application/src/main/java/com/example/android/basicsyncadapter/ |
SyncAdapter.java | 189 * <p>As an additional optimization, we use a batch operation to perform all database writes at 212 ArrayList<ContentProviderOperation> batch = new ArrayList<ContentProviderOperation>(); local 252 batch.add(ContentProviderOperation.newUpdate(existingUri) 266 batch.add(ContentProviderOperation.newDelete(deleteUri).build()); 275 batch.add(ContentProviderOperation.newInsert(FeedContract.Entry.CONTENT_URI) 283 Log.i(TAG, "Merge solution ready. Applying batch update"); 284 mContentResolver.applyBatch(FeedContract.CONTENT_AUTHORITY, batch);
|
/development/samples/browseable/BasicSyncAdapter/src/com.example.android.basicsyncadapter/ |
SyncAdapter.java | 189 * <p>As an additional optimization, we use a batch operation to perform all database writes at 212 ArrayList<ContentProviderOperation> batch = new ArrayList<ContentProviderOperation>(); local 252 batch.add(ContentProviderOperation.newUpdate(existingUri) 266 batch.add(ContentProviderOperation.newDelete(deleteUri).build()); 275 batch.add(ContentProviderOperation.newInsert(FeedContract.Entry.CONTENT_URI) 283 Log.i(TAG, "Merge solution ready. Applying batch update"); 284 mContentResolver.applyBatch(FeedContract.CONTENT_AUTHORITY, batch);
|
/device/google/contexthub/sensorhal/ |
sensors.cpp | 58 device.batch = BatchWrapper; 111 int SensorContext::batch( function in class:SensorContext 115 ALOGV("batch"); 119 return h->batch(handle, sampling_period_ns, max_report_latency_ns); 185 return reinterpret_cast<SensorContext *>(dev)->batch( 286 int SensorContext::HubConnectionOperation::batch( function in class:SensorContext::HubConnectionOperation 351 int SensorContext::DynamicSensorManagerOperation::batch(int handle, int64_t sampling_period_ns, function in class:SensorContext::DynamicSensorManagerOperation 353 return mDynamicSensorManager->batch(handle, sampling_period_ns, max_report_latency_ns);
|
/external/drm_gralloc/ |
gralloc_drm_intel.c | 64 uint32_t *batch, *cur; member in struct:intel_info 78 info->cur = info->batch; 92 return info->cur - info->batch; 106 uint32_t offset = (info->cur - info->batch) * sizeof(info->batch[0]); 129 size *= sizeof(info->batch[0]); 130 ret = drm_intel_bo_subdata(info->batch_ibo, 0, size, info->batch); 132 ALOGE("failed to subdata batch"); 138 ALOGE("failed to exec batch"); 145 info->cur = info->batch; [all...] |
/external/linux-kselftest/tools/testing/selftests/networking/timestamping/ |
txtimestamp.c | 184 int batch = 0; local 223 batch++; 227 if (batch > 1) 228 fprintf(stderr, "batched %d timestamps\n", batch);
|
/external/mesa3d/src/gallium/auxiliary/hud/ |
hud_driver_query.c | 76 bq->result[idx] = MALLOC(sizeof(bq->result[idx]->batch[0]) * 129 "gallium_hud: could not begin batch query. You may have " 201 struct hud_batch_query_context *batch; member in struct:query_info 218 struct hud_batch_query_context *bq = info->batch; 224 info->results_cumulative += bq->result[idx]->batch[result_index].u64; 295 assert(!info->batch); 306 if (info->batch) { 343 if (!info->batch && info->last_time) { 390 info->batch = *pbq;
|
/external/mesa3d/src/gallium/drivers/freedreno/ |
freedreno_query_hw.c | 64 get_sample(struct fd_batch *batch, struct fd_ringbuffer *ring, 67 struct fd_context *ctx = batch->ctx; 73 if (!batch->sample_cache[idx]) { 75 ctx->sample_providers[idx]->get_sample(batch, ring); 76 fd_hw_sample_reference(ctx, &batch->sample_cache[idx], new_samp); 77 util_dynarray_append(&batch->samples, struct fd_hw_sample *, new_samp); 78 batch->needs_flush = true; 81 fd_hw_sample_reference(ctx, &samp, batch->sample_cache[idx]); 87 clear_sample_cache(struct fd_batch *batch) 91 for (i = 0; i < ARRAY_SIZE(batch->sample_cache); i++ 160 struct fd_batch *batch = ctx->batch; local 186 struct fd_batch *batch = ctx->batch; local [all...] |
freedreno_state.c | 122 struct fd_batch *batch, *old_batch = NULL; local 124 fd_batch_reference(&old_batch, ctx->batch); 129 batch = fd_batch_from_fb(&ctx->screen->batch_cache, ctx, framebuffer); 130 fd_batch_reference(&ctx->batch, NULL); 131 fd_reset_wfi(batch); 132 ctx->batch = batch; 137 * to the uncommitted batch (ie. you probably don't blit 146 DBG("%d: cbufs[0]=%p, zsbuf=%p", ctx->batch->needs_flush, 148 fd_batch_flush(ctx->batch, false) [all...] |
/external/mesa3d/src/mesa/state_tracker/ |
st_cb_perfmon.c | 45 unsigned *batch = NULL; local 81 batch = CALLOC(max_batch_counters, sizeof(*batch)); 82 if (!batch) 101 batch[num_batch_counters++] = stc->query_type; 111 /* Create the batch query. */ 114 batch); 115 stm->batch_result = CALLOC(num_batch_counters, sizeof(stm->batch_result->batch[0])); 120 FREE(batch); 124 FREE(batch); [all...] |
/external/oj-libjdwp/src/share/back/ |
ReferenceTypeImpl.c | 251 ObjectBatch batch; local 253 error = classInstances(clazz, &batch, maxInstances); 260 (void)outStream_writeInt(out, batch.count); 261 if (batch.count > 0) { 266 typeKey = specificTypeKey(env, batch.objects[0]); 268 for (kk = 0; kk < batch.count; kk++) { 271 inst = batch.objects[kk]; 276 jvmtiDeallocate(batch.objects);
|
/external/skia/src/gpu/ccpr/ |
GrCoverageCountingPathRenderer.cpp | 522 const AtlasBatch& batch = fAtlasBatches[i]; local 523 SkASSERT(batch.fEndInstanceIdx > baseInstance); 525 if (!batch.fAtlas->textureProxy()) { 530 sk_ref_sp(batch.fAtlas->textureProxy()), this->getFillType()); 536 batch.fEndInstanceIdx - baseInstance, baseInstance);
|
/external/skqp/src/gpu/ccpr/ |
GrCoverageCountingPathRenderer.cpp | 524 const AtlasBatch& batch = fAtlasBatches[i]; local 525 SkASSERT(batch.fEndInstanceIdx > baseInstance); 527 if (!batch.fAtlas->textureProxy()) { 532 sk_ref_sp(batch.fAtlas->textureProxy()), this->getFillType()); 538 batch.fEndInstanceIdx - baseInstance, baseInstance);
|