HomeSort by relevance Sort by last modified time
    Searched refs:bias_max (Results 1 - 5 of 5) sorted by null

  /external/tensorflow/tensorflow/core/kernels/
quantized_bias_add_op_test.cc 62 const float bias_max = 3.0f; local
67 FloatTensorToQuantized<quint8>(bias_float, bias_min, bias_max);
80 AddInputFromArray<float>(TensorShape({1}), {bias_max});
122 const float bias_max = 0.641057f; local
139 FloatTensorToQuantized<quint8>(bias_float, bias_min, bias_max);
162 AddInputFromArray<float>(TensorShape({1}), {bias_max});
quantized_bias_add_op.cc 44 const float bias_max = context->input(5).flat<float>()(0); variable
72 bias_max, &total_min, &total_max);
76 bias_min, bias_max, total_min, total_max,
81 input_max, bias, bias_min, bias_max, output, &total_min, &total_max);
meta_support.h 94 // [input_min, input_max], and [bias_min, bias_max] accordingly, as uint8
101 float bias_max, float output_min, float output_max,
meta_support.cc 349 float bias_max, float output_min, float output_max,
368 CalculateRangeScale<uint8_t>(bias_min, bias_max);
  /hardware/qcom/neuralnetworks/hvxservice/1.0/
HexagonOperationsPrepare.cpp 626 const hexagon_nn_input& bias_max = model->getQuantizationMax(ins[2]); local
632 OP_QuantizedConv2d_8x8to32, pad, {bias, bias_min, bias_max}, act,
683 const hexagon_nn_input& bias_max = model->getQuantizationMax(ins[2]); local
690 OP_QuantizedDepthwiseConv2d_8x8to32, pad, {bias, bias_min, bias_max}, act,
726 const hexagon_nn_input& bias_max = model->getQuantizationMax(ins[2]); local
730 OP_QuantizedMatMul_8x8to32, NN_PAD_NA, {bias, bias_min, bias_max}, act
    [all...]

Completed in 416 milliseconds