HomeSort by relevance Sort by last modified time
    Searched refs:shape (Results 1 - 25 of 2472) sorted by null

1 2 3 4 5 6 7 8 91011>>

  /external/guice/extensions/grapher/src/com/google/inject/grapher/graphviz/
NodeShape.java 31 private final String shape; field in class:NodeShape
33 NodeShape(String shape) {
34 this.shape = shape;
39 return shape;
  /external/snakeyaml/src/test/java/org/yaml/snakeyaml/javabeans/
TriangleBean.java 20 private Shape shape; field in class:TriangleBean
30 public Shape getShape() {
31 return shape;
34 public void setShape(Shape shape) {
35 this.shape = shape;
  /external/tensorflow/tensorflow/java/src/main/java/org/tensorflow/
Shape.java 20 /** The possibly partially known shape of a tensor produced by an operation. */
21 public final class Shape {
23 /** Create a Shape representing an unknown number of dimensions. */
24 public static Shape unknown() {
25 return new Shape(null);
28 /** Create a Shape representing a scalar value. */
29 public static Shape scalar() {
30 return new Shape(new long[0]);
34 * Create a Shape representing an N-dimensional value.
36 * <p>Creates a Shape representing an N-dimensional value (N being at least 1), with the provide
55 long[] shape = new long[otherDimensionSizes.length + 1]; local
118 private long[] shape; field in class:Shape
    [all...]
  /external/libtextclassifier/
tensor-view.cc 22 int NumberOfElements(const std::vector<int>& shape) {
24 for (const int dim : shape) {
  /external/tensorflow/tensorflow/contrib/timeseries/examples/
predict_test.py 40 self.assertAllEqual([500], times.shape)
41 self.assertAllEqual([500], observed.shape)
42 self.assertAllEqual([700], all_times.shape)
43 self.assertAllEqual([700], mean.shape)
44 self.assertAllEqual([700], upper_limit.shape)
45 self.assertAllEqual([700], lower_limit.shape)
54 self.assertAllEqual(times.shape, observed.shape)
55 self.assertAllEqual(all_times.shape, mean.shape)
    [all...]
known_anomaly_test.py 35 self.assertAllEqual([200], times.shape)
36 self.assertAllEqual([200], observed.shape)
37 self.assertAllEqual([300], all_times.shape)
38 self.assertAllEqual([300], mean.shape)
39 self.assertAllEqual([300], upper_limit.shape)
40 self.assertAllEqual([300], lower_limit.shape)
  /external/tensorflow/tensorflow/compiler/xla/client/lib/
testing.cc 35 // specified shape. In case of a (nested) tuple shape this is the total byte
37 int64 DataSizeOfShape(const Shape& shape) {
38 if (ShapeUtil::IsArray(shape)) {
39 return ShapeUtil::ByteSizeOf(shape);
43 for (const Shape& s : shape.tuple_shapes()) {
50 // given shape.
51 ComputationDataHandle BuildFakeDataOpOnDevice(const Shape& shape
    [all...]
  /libcore/ojluni/src/test/java/util/stream/bootlib/java/util/stream/
FlagDeclaringOp.java 31 private final StreamShape shape; field in class:FlagDeclaringOp
37 public FlagDeclaringOp(int flags, StreamShape shape) {
39 this.shape = shape;
44 return shape;
49 return shape;
  /libcore/ojluni/src/test/java/util/stream/testlib/org/openjdk/testlib/java/util/stream/
FlagDeclaringOp.java 34 private final StreamShape shape; field in class:FlagDeclaringOp
40 public FlagDeclaringOp(int flags, StreamShape shape) {
42 this.shape = shape;
47 return shape;
52 return shape;
  /external/tensorflow/tensorflow/compiler/tf2xla/
shape_util.h 27 // Convert an XLA Shape into the equivalent TensorFlow shape. May fail since
29 Status XLAShapeToTensorShape(const xla::Shape& shape,
32 // Convert a TensorShape into the equivalent XLA Shape proto. Unlike Tensorflow,
36 xla::Shape* shape);
literal_util.cc 26 xla::Shape literal_shape;
28 host_tensor.dtype(), host_tensor.shape(), &literal_shape));
45 TF_RET_CHECK(xla::ShapeUtil::IsArray(literal.shape()) &&
46 xla::ShapeUtil::ElementsIn(literal.shape()) ==
51 if (literal.shape().element_type() != primitive_type) {
54 xla::PrimitiveType_Name(literal.shape().element_type()),
68 TensorShape shape; local
69 TF_RETURN_IF_ERROR(XLAShapeToTensorShape(literal.shape(), &shape));
70 *host_tensor = Tensor(target_type, shape);
    [all...]
  /toolchain/binutils/binutils-2.27/gas/testsuite/gas/arm/
neon-vmov-bad.l 2 [^:]*:3: Error: invalid instruction shape -- `vmov Q0,d0'
  /external/tensorflow/tensorflow/contrib/solvers/python/ops/
util.py 33 "LinearOperator", ["shape", "dtype", "apply", "apply_adjoint"])
36 shape = matrix.get_shape()
37 if shape.is_fully_defined():
38 shape = shape.as_list()
40 shape = array_ops.shape(matrix)
42 shape=shape,
52 "LinearOperator", ["shape", "dtype", "apply", "apply_adjoint"]
    [all...]
  /external/tensorflow/tensorflow/compiler/xla/
layout_util.cc 89 /* static */ Layout LayoutUtil::GetDefaultLayoutForShape(const Shape& shape) {
91 DCHECK(!ShapeUtil::IsTuple(shape));
92 return CreateDefaultLayoutForRank(shape.dimensions_size());
111 /* static */ void LayoutUtil::SetToDefaultLayout(Shape* shape) {
112 if (ShapeUtil::IsTuple(*shape)) {
113 // Tuple shape.
114 for (auto& element_shape : *shape->mutable_tuple_shapes()) {
117 shape->clear_layout()
    [all...]
shape_util.cc 72 bool CompareShapes(const Shape& lhs, const Shape& rhs, bool compare_layouts) {
76 [=](const Shape& l, const Shape& r) {
117 // Constructs and returns the new shape with the given minor_to_major order in
119 StatusOr<Shape> MakeShapeWithLayoutInternal(
130 Shape shape = ShapeUtil::MakeShape(element_type, dimensions); local
131 auto min2maj = shape.mutable_layout()->mutable_minor_to_major();
136 if (!shape.has_layout())
210 Shape shape = ShapeUtil::MakeShape(element_type, dimensions); local
821 << "Invalid index " << index << " for shape " << shape; local
    [all...]
index_util.cc 30 const Shape& shape, tensorflow::gtl::ArraySlice<int64> multi_index) {
31 DCHECK_EQ(shape.dimensions_size(), multi_index.size());
33 DCHECK_EQ(0, shape.layout().padded_dimensions_size());
37 DCHECK_LT(multi_index[i], shape.dimensions(i))
40 << "\n\tshape: " << ShapeUtil::HumanString(shape);
81 for (auto dimension : LayoutUtil::MinorToMajor(shape)) {
85 scale = shape.dimensions(dimension);
89 scale *= shape.dimensions(dimension);
96 const Shape& shape, int64 linear_index)
    [all...]
shape_layout.h 28 // A ShapeLayout object encapsulates the layout of a particular shape (including
30 // single array. ShapeLayout contains a Layout proto for each array in the shape
33 // shape with mutable layouts.
36 // Constructs a ShapeLayout of the given shape. Layouts are copied from the
37 // shape parameter.
38 explicit ShapeLayout(const Shape& shape) : shape_(shape) {}
41 // shape. 'to_shape' and the shape of the ShapeLayout object must b
62 const Shape& shape() const { return shape_; } function in class:xla::ShapeLayout
    [all...]
shape_util.h 37 // An index for specifying a particular nested subshape within a shape. Used in
41 // shape. For a non-nested tuple, an index has a single element. For example,
143 // Namespaced collection of (static) shape utilities.
149 // Returns the number of elements are contained within the provided shape;
152 // LayoutUtil::MaxSparseElements(shape) to obtain the maximum number of
153 // elements that can be stored in a sparse shape.
154 // Precondition: !IsTuple(shape)
155 static int64 ElementsIn(const Shape& shape);
157 // Returns true if 'shape' has zero elements
    [all...]
  /development/samples/ApiDemos/src/com/example/android/apis/animation/
ShapeHolder.java 22 import android.graphics.drawable.shapes.Shape;
26 * A data structure that holds a Shape and various properties that can be used to define
27 * how the shape is drawn.
31 private ShapeDrawable shape; field in class:ShapeHolder
57 shape = value;
60 return shape;
66 shape.getPaint().setColor(value);
78 shape.setAlpha((int)((alpha * 255f) + .5f));
82 return shape.getShape().getWidth();
85 Shape s = shape.getShape()
    [all...]
  /external/tensorflow/tensorflow/compiler/tests/
reverse_ops_test.py 34 shape = (7, 5, 9, 11)
35 for revdim in range(len(shape)):
36 self._AssertReverseEqual([revdim], shape)
39 shape = (7, 5, 9, 11)
41 itertools.combinations(range(len(shape)), k)
42 for k in range(2, len(shape)+1)):
43 self._AssertReverseEqual(revdims, shape)
45 def _AssertReverseEqual(self, revdims, shape):
47 pval = np.random.randint(0, 100, size=shape).astype(float)
50 p = array_ops.placeholder(dtypes.int32, shape=shape
    [all...]
  /external/tensorflow/tensorflow/python/ops/
partitioned_variables.py 42 <shape>, <slicing>, <initializer>, name=<optional-name>)
111 def _partitioner(shape, dtype):
115 shape: A `TensorShape`.
119 A tuple representing how much to slice each axis in shape.
122 ValueError: If shape is not a fully defined `TensorShape` or dtype is not
125 if not isinstance(shape, tensor_shape.TensorShape):
126 raise ValueError("shape is not a TensorShape: %s" % shape)
127 if not shape.is_fully_defined():
128 raise ValueError("shape is not fully defined: %s" % shape
    [all...]
  /external/tensorflow/tensorflow/contrib/distributions/python/ops/bijectors/
softmax_centered.py 102 ndims = array_ops.shape(input_shape)
104 # It is not possible for a negative shape so we need only check <= 1.
124 ndims = array_ops.shape(output_shape)[0]
126 # It is not possible for a negative shape so we need only check <= 1.
140 # Set shape hints.
141 if x.shape.ndims is not None:
142 shape = x.shape.as_list()
144 shape += [2]
145 elif shape[-1] is not None
    [all...]
  /external/tensorflow/tensorflow/compiler/xla/service/cpu/
cpu_hlo_support_checker.cc 29 ShapeUtil::ValidateShapeWithOptionalLayout(instruction->shape()));
31 instruction->shape(),
32 [&instruction](const Shape& subshape, const ShapeIndex&) {
35 "CPU backend does not support HLO instruction %s with shape "
38 ShapeUtil::HumanStringWithLayout(instruction->shape())
  /external/tensorflow/tensorflow/compiler/xla/service/gpu/
gpu_hlo_support_checker.cc 29 ShapeUtil::ValidateShapeWithOptionalLayout(instruction->shape()));
31 instruction->shape(),
32 [&instruction](const Shape& subshape, const ShapeIndex&) {
35 "GPU backend does not support HLO instruction %s with shape "
38 ShapeUtil::HumanStringWithLayout(instruction->shape())
  /external/tensorflow/tensorflow/compiler/xla/service/
zero_sized_hlo_elimination.cc 35 ShapeUtil::IsTuple(instruction->shape())) {
39 ShapeUtil::HasZeroElements(instruction->shape())) {
42 Literal::CreateFromShape(instruction->shape()))));

Completed in 1172 milliseconds

1 2 3 4 5 6 7 8 91011>>