/external/tensorflow/tensorflow/core/util/sparse/ |
sparse_tensor.h | 89 int dims() const { return shape_.size(); } function in class:tensorflow::sparse::SparseTensor 141 // out->shape().dims() == shape().dims() 203 ShapeArray vec(shape.dims()); 204 for (int i = 0; i < shape.dims(); ++i) vec[i] = shape.dim_size(i); 366 CHECK_EQ(out->shape().dims(), dims_) 376 if (shape_.size() != out_shape.dims()) return false; 426 const int dims = tensors[0].dims_; local 427 CHECK_GE(dims, 1) << "Cannot concat 0-dimensional SparseTensors"; 437 CHECK_EQ(st.dims_, dims) << "All SparseTensors must have the same rank." 574 const int dims = input_tensor.dims(); local [all...] |
/external/tensorflow/tensorflow/compiler/tf2xla/kernels/ |
lrn_ops.cc | 38 OP_REQUIRES(ctx, in_shape.dims() == 4, 90 OP_REQUIRES(ctx, in_grads_shape.dims() == 4 && in_image_shape.dims() == 4,
|
tensor_array_ops.cc | 224 xla::MakeEdgePaddingConfig({{0, elem_shape.dims()}})); 268 xla::MakeEdgePaddingConfig({{0, ta_shape.dims() - 1}})); 307 OP_REQUIRES(ctx, indices_shape.dims() == 1, 328 std::vector<int64> begin(ta_shape.dims(), 0); 329 std::vector<int64> strides(ta_shape.dims(), 1); 330 std::vector<int64> end(ta_shape.dims(), 1); 332 for (auto i = 1; i < ta_shape.dims(); i++) { 375 OP_REQUIRES(ctx, indices_shape.dims() >= 1, 405 std::vector<int64> value_starts(value_shape.dims(), 0); 408 std::vector<int64> value_strides(value_shape.dims(), 1) [all...] |
matrix_band_part_op.cc | 56 const int64 m = input_shape.dim_size(input_shape.dims() - 2); 57 const int64 n = input_shape.dim_size(input_shape.dims() - 1);
|
/external/tensorflow/tensorflow/contrib/coder/kernels/ |
range_coder_ops_test.cc | 71 std::vector<int64> stride(shape.dims()); 73 for (int i = shape.dims() - 1; i >= 0; --i) { 137 Tensor shape{DT_INT32, {data_shape.dims()}}; 138 for (int i = 0; i < data_shape.dims(); ++i) { 164 CHECK_EQ(cdf_tensor->dims(), maxvalue_tensor.dims() + 1); 165 const int64 chip_size = cdf_tensor->dim_size(cdf_tensor->dims() - 1); 170 for (int i = 0; i < cdf_tensor->dims(); ++i) { 406 CHECK_EQ(shape.dims(), 4); 444 CHECK_EQ(shape.dims(), 4) [all...] |
range_coder_ops.cc | 120 if (TF_PREDICT_FALSE(cdf_shape.dims() != data_shape.dims() + 1)) { 126 if (TF_PREDICT_FALSE(cdf_shape.dim_size(cdf_shape.dims() - 1) <= 1)) { 160 #define RANGE_ENCODE_CASE(dims) \ 161 case dims: { \ 162 RangeEncodeImpl<dims>(data.flat<int16>(), data_shape, \
|
/external/tensorflow/tensorflow/contrib/lite/kernels/ |
embedding_lookup_sparse_test.cc | 64 int rows = tensor->dims->data[0]; 65 int columns = tensor->dims->data[1]; 66 int features = tensor->dims->data[2];
|
hashtable_lookup_test.cc | 62 int rows = tensor->dims->data[0]; 70 int rows = tensor->dims->data[0]; 71 int features = tensor->dims->data[1];
|
split.cc | 62 TfLiteIntArray* output_dims = TfLiteIntArrayCopy(input->dims); 124 all_outputs.dims()); \ 129 all_outputs.data(), all_outputs.dims()); \
|
/external/tensorflow/tensorflow/contrib/lite/toco/graph_transformations/ |
resolve_constant_strided_slice.cc | 37 start += input_shape.dims(axis); 48 stop = input_shape.dims(axis); 54 stop += input_shape.dims(axis);
|
/external/tensorflow/tensorflow/core/kernels/data/ |
tensor_slice_dataset_op.cc | 39 OP_REQUIRES(ctx, inputs[0].dims() > 0, 45 OP_REQUIRES(ctx, t.dims() > 0, 66 for (int i = 1; i < t.dims(); ++i) {
|
/external/tensorflow/tensorflow/core/kernels/ |
serialize_sparse_op.cc | 302 const int ndims = serialized_sparse.shape().dims(); 326 if (num_sparse_tensors == 0 && serialized_sparse.shape().dims() == 1) { 388 context, shape.dims() == expanded_tensor_shape.dims(), 392 i, "] was: ", shape.dims() - 1, " but rank of SparseTensor[", i, 393 "] is: ", expanded_tensor_shape.dims() - 1)); 394 for (int j = 1; j < shape.dims(); ++j) { 409 int rank = shape.dims(); 439 Tensor input_shape(DT_INT64, TensorShape({output.dims()})); 440 std::copy_n(output.shape().data(), output.dims(), [all...] |
sparse_tensor_dense_add_op.cc | 50 if (a_shape->NumElements() != b->dims()) { 53 " and ", b->dims()); 56 for (int i = 0; i < b->dims(); ++i) {
|
unravel_index_op.cc | 55 auto dims = dims_tensor.vec<Tidx>(); variable 66 strides = dims.reverse(reverse) 77 strides_shifted = dims.reverse(reverse)
|
eigen_attention.h | 67 DSizes<Index, 4> dims = in.dimensions(); local 69 dims[0] = in.dimension(0); 70 dims[1] = width_; 71 dims[2] = height_; 72 dims[3] = in.dimension(3); 73 return dims;
|
scatter_op.cc | 41 if (updates.dims() != indices.dims() + params.dims() - 1) return false; 42 for (int d = 0; d < indices.dims(); d++) { 47 for (int d = 1; d < params.dims(); d++) { 48 if (params.dim_size(d) != updates.dim_size(d - 1 + indices.dims())) {
|
transpose_functor_gpu.cu.cc | 63 const int32 ndims = in.dims(); 105 int dims = new_dims.size(); local 106 if (dims < 2 || dims > 3) return false; 110 switch (dims) { 175 switch (in.dims()) {
|
/external/tensorflow/tensorflow/python/lib/core/ |
ndarray_tensor_bridge.cc | 188 Status ArrayFromMemory(int dim_size, npy_intp* dims, void* data, DataType dtype, 192 size *= dims[i]; 207 PyArray_SimpleNewFromData(dim_size, dims, type_num, data);
|
/external/eigen/unsupported/test/ |
cxx11_tensor_argmax.cpp | 65 DimensionList<DenseIndex, 4> dims; local 67 dims, internal::ArgMaxTupleReducer<Tuple<DenseIndex, float> >()); 97 DimensionList<DenseIndex, 4> dims; local 99 dims, internal::ArgMinTupleReducer<Tuple<DenseIndex, float> >()); 164 std::vector<int> dims {2, 3, 5, 7}; 222 std::vector<int> dims {2, 3, 5, 7};
|
cxx11_tensor_simple.cpp | 198 array<Eigen::DenseIndex, 3> dims; local 199 dims[0] = 2; 200 dims[1] = 3; 201 dims[2] = 4; 202 Tensor<int, 3> t1(dims); 203 Tensor<int, 3, RowMajor> t2(dims);
|
/external/jacoco/org.jacoco.core/src/org/jacoco/core/internal/instr/ |
DuplicateFrameEliminator.java | 120 public void visitMultiANewArrayInsn(final String desc, final int dims) { 122 mv.visitMultiANewArrayInsn(desc, dims);
|
/external/mesa3d/src/mesa/drivers/dri/i965/ |
intel_tex_copy.c | 103 intelCopyTexSubImage(struct gl_context *ctx, GLuint dims, 127 _mesa_meta_CopyTexSubImage(ctx, dims, texImage,
|
/external/mesa3d/src/mesa/main/ |
pack.h | 99 _mesa_unpack_color_index_to_rgba_float(struct gl_context *ctx, GLuint dims, 106 _mesa_unpack_color_index_to_rgba_ubyte(struct gl_context *ctx, GLuint dims,
|
/external/tensorflow/tensorflow/compiler/xla/service/cpu/ |
runtime_matmul.cc | 57 const Eigen::array<DimPair, 1> dims( 63 C.device(*run_options->intra_op_thread_pool()) = A.contract(B, dims);
|
runtime_single_threaded_matmul.cc | 51 const Eigen::array<DimPair, 1> dims( 57 C = A.contract(B, dims);
|