/external/tensorflow/tensorflow/core/kernels/ |
quantized_bias_add_op_test.cc | 62 const float bias_max = 3.0f; local 67 FloatTensorToQuantized<quint8>(bias_float, bias_min, bias_max); 80 AddInputFromArray<float>(TensorShape({1}), {bias_max}); 122 const float bias_max = 0.641057f; local 139 FloatTensorToQuantized<quint8>(bias_float, bias_min, bias_max); 162 AddInputFromArray<float>(TensorShape({1}), {bias_max});
|
quantized_bias_add_op.cc | 44 const float bias_max = context->input(5).flat<float>()(0); variable 72 bias_max, &total_min, &total_max); 76 bias_min, bias_max, total_min, total_max, 81 input_max, bias, bias_min, bias_max, output, &total_min, &total_max);
|
meta_support.h | 94 // [input_min, input_max], and [bias_min, bias_max] accordingly, as uint8 101 float bias_max, float output_min, float output_max,
|
meta_support.cc | 349 float bias_max, float output_min, float output_max, 368 CalculateRangeScale<uint8_t>(bias_min, bias_max);
|
/hardware/qcom/neuralnetworks/hvxservice/1.0/ |
HexagonOperationsPrepare.cpp | 626 const hexagon_nn_input& bias_max = model->getQuantizationMax(ins[2]); local 632 OP_QuantizedConv2d_8x8to32, pad, {bias, bias_min, bias_max}, act, 683 const hexagon_nn_input& bias_max = model->getQuantizationMax(ins[2]); local 690 OP_QuantizedDepthwiseConv2d_8x8to32, pad, {bias, bias_min, bias_max}, act, 726 const hexagon_nn_input& bias_max = model->getQuantizationMax(ins[2]); local 730 OP_QuantizedMatMul_8x8to32, NN_PAD_NA, {bias, bias_min, bias_max}, act [all...] |