/hardware/intel/img/psb_video/src/mrst/ |
lnc_MPEG4ES.c | 202 ASSERT(obj_buffer->num_elements == 1); 205 if ((obj_buffer->num_elements != 1) || 288 if ((obj_buffer->num_elements != 1) || 351 if ((ctx->slice_param_cache != NULL) && (obj_buffer->num_elements != ctx->slice_param_num)) { 353 ctx->slice_param_num, obj_buffer->num_elements); 361 ctx->slice_param_num = obj_buffer->num_elements; 370 for (i = 0; i < obj_buffer->num_elements; i++) {
|
/hardware/intel/img/psb_video/src/ |
pnw_H263ES.c | 187 ASSERT(obj_buffer->num_elements == 1); 194 if ((obj_buffer->num_elements != 1) || 282 if ((obj_buffer->num_elements != 1) || 360 if ((ctx->slice_param_cache != NULL) && (obj_buffer->num_elements != ctx->slice_param_num)) { 362 ctx->slice_param_num, obj_buffer->num_elements); 369 ctx->slice_param_num = obj_buffer->num_elements; 379 for (i = 0; i < obj_buffer->num_elements; i++) {
|
pnw_MPEG4ES.c | 196 ASSERT(obj_buffer->num_elements == 1); 204 if ((obj_buffer->num_elements != 1) || 326 if ((obj_buffer->num_elements != 1) || 399 if ((ctx->slice_param_cache != NULL) && (obj_buffer->num_elements != ctx->slice_param_num)) { 401 ctx->slice_param_num, obj_buffer->num_elements); 409 ctx->slice_param_num = obj_buffer->num_elements; 419 for (i = 0; i < obj_buffer->num_elements; i++) {
|
/external/tensorflow/tensorflow/core/framework/ |
tensor.cc | 609 shape_.num_elements() == 0; 629 CHECK_EQ(shape.num_elements(), other.NumElements()); 648 CHECK_EQ(shape.num_elements() * out_size, 649 other.shape().num_elements() * in_size); 717 if (shape_.num_elements() > 0 || a->ShouldAllocateEmptyTensors()) { 718 CASES(type, buf_ = new Buffer<T>(a, shape.num_elements())); 731 if (shape_.num_elements() > 0 || a->ShouldAllocateEmptyTensors()) { 732 CASES(type, buf_ = new Buffer<T>(a, shape.num_elements(), allocation_attr)); 814 const int64 N = shape.num_elements(); 845 CASES(dtype(), ToProtoField<T>(*buf_, shape_.num_elements(), proto)) [all...] |
/external/tensorflow/tensorflow/core/kernels/ |
example_parsing_ops.cc | 507 const int64 num_elements = feature_values.NumElements(); variable 508 TensorShape indices_shape({num_elements, 1}); 517 shape_t(0) = num_elements; 519 std::iota(indices_t.data(), indices_t.data() + num_elements, 0); 626 const int64 num_elements = v.shape().num_elements(); variable 627 total_num_features += num_elements; 628 max_num_features = std::max(max_num_features, num_elements); 649 const int64 num_elements = CopyIntoSparseTensor( variable 651 offset += num_elements; [all...] |
barrier_ops.cc | 94 ctx, keys.NumElements() == 0 || element_shape.num_elements() > 0, 201 void TryTakeMany(int num_elements, bool allow_small_batch, int64 timeout, 204 int num_elements_to_deliver = num_elements; 210 // We want to deliver a maximum of num_elements, if there are less 215 num_elements_to_deliver = std::min(num_elements, available_elements); 585 OP_REQUIRES_OK_ASYNC(ctx, ctx->input("num_elements", &Tnum_elements), 588 errors::InvalidArgument("num_elements must be a scalar."), 590 const int32 num_elements = Tnum_elements->scalar<int32>()(); variable 602 num_elements, allow_small_batch_, timeout_, ctx,
|
fifo_queue.h | 46 void TryDequeueMany(int num_elements, OpKernelContext* ctx,
|
parameterized_truncated_normal_op.h | 38 int64 samples_per_batch, int64 num_elements,
|
quantized_reshape_op_test.cc | 54 for (int i = 0; i < input.shape().num_elements(); ++i) {
|
/external/tensorflow/tensorflow/cc/framework/ |
gradient_checker.cc | 147 const int64 dy_size = y_shapes[y_idx].num_elements(); 161 const int64 x_size = x_shapes[x_idx].num_elements(); 223 const int64 x_size = x_shapes[x_idx].num_elements(); 249 const int64 y_size = y_shapes[y_idx].num_elements(); 309 x_shapes[x_idx].num_elements() * JacobianStride<X_T>::value; 315 y_shapes[y_idx].num_elements() * JacobianStride<Y_T>::value;
|
/external/drm_hwcomposer/ |
drmhwctwo.cpp | 276 uint32_t *num_elements, hwc2_layer_t *layers, int32_t *types) { 281 if (layers && num_changes < *num_elements) 283 if (types && num_changes < *num_elements) 289 *num_elements = num_changes; 418 uint32_t *num_elements, 424 unsupported(__func__, display_requests, num_elements, layers, layer_requests); 425 *num_elements = 0; 449 HWC2::Error DrmHwcTwo::HwcDisplay::GetReleaseFences(uint32_t *num_elements, 459 } else if (num_layers > *num_elements) { 460 ALOGW("Overflow num_elements %d/%d", num_layers, *num_elements) [all...] |
drmhwctwo.h | 152 HWC2::Error GetChangedCompositionTypes(uint32_t *num_elements, 164 uint32_t *num_elements, hwc2_layer_t *layers, 172 HWC2::Error GetReleaseFences(uint32_t *num_elements, hwc2_layer_t *layers,
|
/external/mesa3d/src/gallium/auxiliary/vl/ |
vl_vertex_buffers.c | 145 vl_vb_element_helper(struct pipe_vertex_element* elements, unsigned num_elements, 150 assert(elements && num_elements); 152 for ( i = 0; i < num_elements; ++i ) {
|
/external/mesa3d/src/gallium/drivers/freedreno/a2xx/ |
fd2_draw.c | 64 if (!vtx->num_elements) 67 for (i = 0; i < vtx->num_elements; i++) { 79 fd2_emit_vertex_bufs(ctx->batch->draw, 0x78, bufs, vtx->num_elements);
|
/external/mesa3d/src/gallium/state_trackers/va/ |
picture_mpeg4.c | 38 assert(buf->size >= sizeof(VAPictureParameterBufferMPEG4) && buf->num_elements == 1); 101 assert(buf->size >= sizeof(VAIQMatrixBufferMPEG4) && buf->num_elements == 1); 117 assert(buf->size >= sizeof(VASliceParameterBufferMPEG4) && buf->num_elements == 1);
|
va_private.h | 226 unsigned int num_elements; member in struct:__anon30232 306 unsigned int num_elements, void *data, VABufferID *buf_id); 307 VAStatus vlVaBufferSetNumElements(VADriverContextP ctx, VABufferID buf_id, unsigned int num_elements); 354 unsigned int *size, unsigned int *num_elements);
|
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
concat_op.cc | 165 const int64 dims = inp0_shape.num_elements(); 179 OP_REQUIRES(ctx, dims == inp_shape.num_elements(), 182 inp_shape.num_elements()));
|
/external/tensorflow/tensorflow/compiler/xla/ |
array4d_test.cc | 43 EXPECT_EQ(empty.num_elements(), 120); 86 EXPECT_EQ(arr.num_elements(), 24); 116 EXPECT_EQ(arr.num_elements(), 24);
|
/external/tensorflow/tensorflow/core/lib/random/ |
weighted_picker_test.cc | 158 while (picker.num_elements() != size) { 185 const int size = picker->num_elements(); 204 const int size = picker->num_elements();
|
/toolchain/binutils/binutils-2.27/include/ |
gcc-c-fe.def | 147 /* Create a new array type. If NUM_ELEMENTS is -1, then the array 152 int) /* Argument NUM_ELEMENTS. */ 180 int) /* Argument NUM_ELEMENTS. */
|
/cts/tests/ui/src/android/ui/cts/ |
ScrollingActivity.java | 34 private static final String NUM_ELEMENTS_EXTRA = "num_elements";
|
/external/mesa3d/src/gallium/drivers/etnaviv/ |
etnaviv_state.c | 520 unsigned num_elements, const struct pipe_vertex_element *elements) 528 if (num_elements > ctx->specs.vertex_max_elements) { 529 BUG("number of elements (%u) exceeds chip maximum (%u)", num_elements, 544 for (unsigned idx = 0; idx < num_elements; ++idx) { 549 cs->num_elements = num_elements; 550 if (incompatible || num_elements == 0) { 559 for (unsigned idx = 0; idx < num_elements; ++idx) { 571 nonconsecutive = (idx == (num_elements - 1)) ||
|
/external/mesa3d/src/gallium/drivers/freedreno/ |
freedreno_state.c | 309 fd_vertex_state_create(struct pipe_context *pctx, unsigned num_elements, 317 memcpy(so->pipe, elements, sizeof(*elements) * num_elements); 318 so->num_elements = num_elements;
|
/external/mesa3d/src/util/ |
slab.h | 53 unsigned num_elements; member in struct:slab_parent_pool
|
/external/tensorflow/tensorflow/contrib/boosted_trees/lib/learner/common/stats/ |
node-stats.h | 265 const int num_elements) { 267 num_elements); 271 const int num_elements) { 273 num_elements);
|