HomeSort by relevance Sort by last modified time
    Searched refs:transpose_a (Results 1 - 25 of 56) sorted by null

1 2 3

  /external/tensorflow/tensorflow/compiler/tf2xla/lib/
triangular_solve.h 31 // depending on the value of the value of (left_side, transpose_a, conjugate_a)
53 // `transpose_a` is a boolean indicating whether the matrix `a` is transposed.
56 // transpose_a and conjugate_a are true the effect is a Hermitian adjoint.
62 xla::ComputationDataHandle b, bool left_side, bool lower, bool transpose_a,
67 const xla::ComputationDataHandle& b, bool transpose_a, bool conjugate_a);
triangular_solve.cc 34 xla::ComputationDataHandle b, bool left_side, bool lower, bool transpose_a,
136 b_param, transpose_a,
141 left_side, lower, transpose_a,
164 if (!left_side && lower == transpose_a) {
188 // a_slice_2 = T(a_slice_2) if transpose_a else a_slice_2
203 /*transpose_y=*/transpose_a,
214 } else if (left_side && lower != transpose_a) {
238 // a_slice_2 = T(a_slice_2) if transpose_a else a_slice_2
251 /*transpose_x=*/transpose_a,
262 } else if (!left_side && lower != transpose_a) {
    [all...]
  /external/tensorflow/tensorflow/python/ops/
matmul_benchmark.py 35 def build_graph(device, n, m, k, transpose_a, transpose_b, dtype):
43 transpose_a: boolean value to show if tensor A is transposed.
51 if not transpose_a:
60 z = math_ops.matmul(x, y, transpose_a=transpose_a, transpose_b=transpose_b)
67 def run_graph(self, device, n, m, k, transpose_a, transpose_b, num_iters,
76 transpose_a: boolean value to show if tensor A is transposed.
86 output = build_graph(device, n, m, k, transpose_a, transpose_b, dtype)
99 ',ta:' + str(transpose_a) + '.tb:' + str(transpose_b), num_iters,
109 str(transpose_a) + ',tb:' + str(transpose_b)).replace(' ', '')
    [all...]
matmul_benchmark_test.py 32 def BuildGraphTest(n, m, k, transpose_a, transpose_b, dtype):
37 (n, m, k, transpose_a, transpose_b))
40 (n, m, k, transpose_a, transpose_b))
41 self._VerifyBuildGraph(n, m, k, transpose_a, transpose_b, dtype)
46 def RunGraphTest(n, m, k, transpose_a, transpose_b, dtype):
51 (n, m, k, transpose_a, transpose_b))
54 (n, m, k, transpose_a, transpose_b))
55 self._VerifyRunGraph(n, m, k, transpose_a, transpose_b, dtype)
71 def _VerifyBuildGraph(self, n, m, k, transpose_a, transpose_b, dtype):
75 transpose_a, transpose_b, dtype
    [all...]
math_grad.py     [all...]
  /external/gemmlowp/eight_bit_int_gemm/
eight_bit_int_gemm.h 55 void EightBitIntGemm(bool transpose_a, bool transpose_b, bool transpose_c,
62 void EightBitIntGemm(bool transpose_a, bool transpose_b, bool transpose_c,
eight_bit_int_gemm.cc 71 template <bool transpose_a, bool transpose_b, bool transpose_c>
87 transpose_a ? MapOrder::RowMajor : MapOrder::ColMajor;
110 template <bool transpose_a, bool transpose_b, bool transpose_c>
122 transpose_a ? MapOrder::RowMajor : MapOrder::ColMajor;
220 bool CanHandleMetaFastpath(bool transpose_a, bool transpose_b, bool transpose_c,
229 if (!IsRowMajorOrVector(transpose_a, lda, m, k)) {
304 void EightBitIntGemm(bool transpose_a, bool transpose_b, bool transpose_c,
314 if (CanHandleMetaFastpath(transpose_a, transpose_b, transpose_c, m, n, k, lda,
323 if (transpose_a == ta && transpose_b == tb && transpose_c == tc) { \
341 void EightBitIntGemm(bool transpose_a, bool transpose_b, bool transpose_c
    [all...]
  /external/tensorflow/tensorflow/core/kernels/
quantized_matmul_op_test.cc 90 const bool transpose_a = true; local
104 .Attr("transpose_a", transpose_a)
113 // We have set the transpose_a flag to true, so the matrix is transposed, and
140 const bool transpose_a = true; local
154 .Attr("transpose_a", transpose_a)
180 const bool transpose_a = true; local
194 .Attr("transpose_a", transpose_a)
273 const bool transpose_a = true; local
    [all...]
matmul_op_test.cc 24 static Graph* Matmul(int m, int k, int n, bool transpose_a, bool transpose_b,
27 Tensor in0(type, transpose_a ? TensorShape({k, m}) : TensorShape({m, k}));
32 test::graph::Constant(g, in1), transpose_a, transpose_b);
reference_gemm.h 34 void ReferenceGemm(bool transpose_a, bool transpose_b, bool transpose_c,
40 if (transpose_a) {
matmul_op.cc 274 bool transpose_a = dim_pair[0].first == 0; local
276 auto blas_transpose_a = trans[transpose_a];
294 transpose_a, transpose_b, m, n, k, dtype, device_id,
317 transpose_b ? k : n, a_ptr, transpose_a ? m : k, beta,
335 transpose_b ? k : n, a_ptr, transpose_a ? m : k, 0.0,
348 LaunchBlasGemv<T>::Compute(ctx, stream, !transpose_a,
349 transpose_a ? m : k, transpose_a ? k : m,
376 transpose_b ? k : n, a_ptr, transpose_a ? m : k, beta,
409 LaunchBlasGemv<T>::Compute(ctx, stream, !transpose_a,
    [all...]
meta_support.h 66 // If transpose_a is false the lhs operand has row major layout, otherwise
70 void QuantizedGemm(OpKernelContext* context, bool transpose_a, bool transpose_b,
mkl_matmul_op.cc 42 OP_REQUIRES_OK(ctx, ctx->GetAttr("transpose_a", &transpose_a_));
89 bool transpose_a = dim_pair[0].first == 0; variable
96 MklBlasGemm(transpose_a, transpose_b, m, n, k, a_ptr, transpose_a ? m : k,
sparse_matmul_op_test.cc 51 Node* SparseMatMulNode(Graph* g, Node* in0, Node* in1, bool transpose_a,
57 .Attr("transpose_a", transpose_a)
68 bool transpose_a, bool transpose_b) {
72 auto left_shape = transpose_a ? TensorShape({d, m}) : TensorShape({m, d});
83 test::graph::Constant(g, right), transpose_a, transpose_b,
90 float sparsity_b, bool transpose_a,
94 transpose_a, transpose_b);
quantized_conv_ops.cc 376 const bool transpose_a = false; local
391 meta::QuantizedGemm(context, transpose_a, transpose_b, im2col_buffer,
413 !transpose_a ? gemmlowp::MapOrder::RowMajor
440 transpose_a, transpose_b, transpose_c, m, n, k, im2col_buffer,
  /external/tensorflow/tensorflow/contrib/factorization/python/kernel_tests/
masked_matmul_benchmark.py 63 transpose_a=False, transpose_b=False):
73 transpose_a: boolean, whether to transpose the a matrix.
83 a_shape = a_shape if not transpose_a else [a_shape[1], a_shape[0]]
92 a_ph, b_ph, mask_indices_ph, transpose_a, transpose_b)
116 tr_a=int(transpose_a),
133 for transpose_a in [False, True]:
139 self._run_graph(a_shape, b_shape, nnz, num_iters, sort, transpose_a,
masked_matmul_ops_test.py 60 def _runTestMaskedProduct(self, transpose_a, transpose_b):
62 a = self._a if not transpose_a else array_ops.transpose(self._a)
73 a, b, self._mask_ind, transpose_a, transpose_b)
  /external/tensorflow/tensorflow/python/kernel_tests/
matmul_op_test.py 152 math_ops.matmul(a, b, transpose_a=True)
236 for adjoint_a, transpose_a in trans_options:
240 transpose_a, adjoint_b, transpose_b)
247 transpose_a=transpose_a,
256 transpose_a=transpose_a,
sparse_matmul_op_test.py 57 transpose_a=tr_a,
147 transpose_a=tr_a,
  /external/tensorflow/tensorflow/python/tools/
print_selective_registration_header_test.py 47 attr: { key: "transpose_a" value: { b: false } }
56 attr: { key: "transpose_a" value: { b: false } }
  /external/tensorflow/tensorflow/contrib/factorization/kernels/
masked_matmul_ops.cc 74 const Tensor& transpose_a = context->input(3); variable
83 OP_REQUIRES(context, TensorShapeUtils::IsScalar(transpose_a.shape()),
84 InvalidArgument("Input transpose_a should be a scalar."));
88 const bool adj_a = transpose_a.scalar<bool>()();
  /external/tensorflow/tensorflow/python/ops/linalg/
linear_operator_util.py 200 transpose_a=False,
251 transpose_a: If `True`, `a` is transposed before multiplication.
276 ValueError: If transpose_a and adjoint_a, or transpose_b and adjoint_b
284 transpose_a=transpose_a,
  /external/tensorflow/tensorflow/contrib/rnn/python/ops/
gru_ops.py 90 d_w_ru = math_ops.matmul(x_h_prev, d_r_bar_u_bar, transpose_a=True)
94 d_w_c = math_ops.matmul(x_h_prevr, d_c_bar, transpose_a=True)
  /external/tensorflow/tensorflow/core/grappler/costs/
op_level_cost_estimator.cc 559 bool transpose_a = false; local
567 if (item.first == "transpose_a" && item.second.b() == true)
568 transpose_a = true;
572 VLOG(1) << "transpose_a:" << transpose_a;
578 if (transpose_a) {
683 AttrValue transpose_a; local
684 transpose_a.set_b(false);
686 transpose_a.set_b(op_features.attr().at("adj_x").b());
688 (*matmul_op_features.mutable_attr())["transpose_a"] = transpose_a
    [all...]
  /external/tensorflow/tensorflow/contrib/factorization/python/ops/
gmm_ops.py 62 cov = math_ops.matmul(x, x, transpose_a=True) / (num_points - 1)
382 self._w[shard_id], array_ops.squeeze(shard, [0]), transpose_a=True),
428 square_mean = math_ops.matmul(mean, mean, transpose_a=True)

Completed in 318 milliseconds

1 2 3