/external/tensorflow/tensorflow/core/kernels/ |
quantized_reshape_op.cc | 36 Tensor* output_min = nullptr; variable 37 OP_REQUIRES_OK(ctx, ctx->allocate_output(1, TensorShape({}), &output_min)); 38 output_min->flat<float>()(0) = input_min_float;
|
quantize_op_test.cc | 292 const float output_min = GetOutput(1)->flat<float>()(0); local 294 EXPECT_NEAR(0.0f, output_min, 1e-5f); 314 const float output_min = GetOutput(1)->flat<float>()(0); local 316 EXPECT_NEAR(0.0f, output_min, 1e-5f); 336 const float output_min = GetOutput(1)->flat<float>()(0); local 338 EXPECT_NEAR(0.0f, output_min, 1e-5f); 358 const float output_min = GetOutput(1)->flat<float>()(0); local 360 EXPECT_NEAR(-0.3f, output_min, 1e-5f);
|
quantized_activation_ops.cc | 51 Tensor* output_min = nullptr; variable 52 OP_REQUIRES_OK(context, context->allocate_output(1, {}, &output_min)); 53 output_min->flat<float>()(0) = min_input; 89 Tensor* output_min = nullptr; variable 90 OP_REQUIRES_OK(context, context->allocate_output(1, {}, &output_min)); 91 output_min->flat<float>()(0) = min_input;
|
quantized_activation_ops_test.cc | 62 const float output_min = GetOutput(1)->flat<float>()(0); local 65 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max); 93 const float output_min = GetOutput(1)->flat<float>()(0); local 96 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
|
quantized_pooling_ops.cc | 94 Tensor* output_min = nullptr; variable 95 OP_REQUIRES_OK(context, context->allocate_output(1, {}, &output_min)); 96 output_min->flat<float>()(0) = min_input; 118 Tensor* output_min = nullptr; variable 119 OP_REQUIRES_OK(context, context->allocate_output(1, {}, &output_min)); 120 output_min->flat<float>()(0) = min_input;
|
requantize.cc | 49 Tensor* output_min = nullptr; variable 50 OP_REQUIRES_OK(ctx, ctx->allocate_output(1, TensorShape({}), &output_min)); 92 output_min->flat<float>().setConstant(requested_output_min_float);
|
quantize_down_and_shrink_range.cc | 47 Tensor* output_min = nullptr; variable 48 OP_REQUIRES_OK(ctx, ctx->allocate_output(1, TensorShape({}), &output_min)); 96 output_min->flat<float>().setConstant(actual_min_float);
|
quantized_bias_add_op.cc | 84 Tensor* output_min = nullptr; variable 85 OP_REQUIRES_OK(context, context->allocate_output(1, {}, &output_min)); 86 output_min->flat<float>()(0) = total_min;
|
quantized_bias_add_op_test.cc | 83 const float output_min = GetOutput(1)->flat<float>()(0); local 86 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max); 165 const float output_min = GetOutput(1)->flat<float>()(0); local 168 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max);
|
quantized_pooling_ops_test.cc | 76 const float output_min = GetOutput(1)->flat<float>()(0); local 79 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max); 121 const float output_min = GetOutput(1)->flat<float>()(0); local 124 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
|
requantization_range_op.cc | 51 Tensor* output_min = nullptr; variable 52 OP_REQUIRES_OK(ctx, ctx->allocate_output(0, TensorShape({}), &output_min)); 68 output_min->flat<float>().setConstant(used_min_float);
|
mkl_quantized_concat_op_test.cc | 150 const float output_min = GetOutput(1)->flat<float>()(0); local 153 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max); 224 const float output_min = GetOutput(1)->flat<float>()(0); local 227 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
|
mkl_quantized_pooling_ops_test.cc | 125 const float output_min = GetOutput(1)->flat<float>()(0); local 128 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max); 193 const float output_min = GetOutput(1)->flat<float>()(0); local 196 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
|
mkl_requantization_range_per_channel_op.cc | 102 Tensor* output_min = nullptr; variable 104 OP_REQUIRES_OK(ctx, ctx->allocate_output(kOutputMinIndex, {}, &output_min)); 106 output_min->flat<float>()(0) = is_non_negative ? 0.0f : -out_min_max;
|
mkl_requantize_ops_test.cc | 163 const float output_min = GetOutput(0)->flat<float>()(0); local 165 EXPECT_NEAR(-14.8217, output_min, 0.002); 224 const float output_min = GetOutput(0)->flat<float>()(0); local 226 EXPECT_NEAR(-6.0, output_min, 0.002); // Values are aligned with clip_value. 293 const float output_min = GetOutput(1)->flat<float>()(0); local 295 EXPECT_NEAR(range_op_output_min, output_min, 0.002);
|
quantized_conv_ops_test.cc | 127 const float output_min = GetOutput(1)->flat<float>()(0); local 130 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max); 318 const float output_min = GetOutput(1)->flat<float>()(0); local 321 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max);
|
quantized_matmul_op_test.cc | 351 const float output_min = GetOutput(1)->flat<float>()(0); local 354 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max);
|
mkl_quantized_conv_ops_test.cc | 198 const float output_min = GetOutput(1)->flat<float>()(0); local 201 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max);
|
mkl_requantize_per_channel_op.cc | 134 Tensor* output_min = nullptr; variable 137 ctx->allocate_output(kOutputMinIndex, {}, &output_min)); 141 output_min->flat<float>()(0) = input_requested_min_float;
|
quantized_batch_norm_op_test.cc | 127 const float output_min = GetOutput(1)->flat<float>()(0); local 130 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max); 235 const float output_min = GetOutput(1)->flat<float>()(0); local 238 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max);
|
quantized_concat_op_test.cc | 111 const float output_min = GetOutput(1)->flat<float>()(0); local 114 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max); 177 const float output_min = GetOutput(1)->flat<float>()(0); local 180 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max); 239 const float output_min = GetOutput(1)->flat<float>()(0); local 242 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
|
quantized_batch_norm_op.cc | 38 Tensor* output, float* output_min, float* output_max) { 49 *output_min = std::numeric_limits<float>::max(); 78 *output_min = std::min(output_value, *output_min); 82 FloatToQuantized<T2>(output_value, *output_min, *output_max); 101 Tensor* output, float* output_min, float* output_max) { 115 *output_min = -(1 << 20); 139 FloatToQuantized<T2>(scale_value, *output_min, *output_max); 141 FloatToQuantized<T2>(offset_value, *output_min, *output_max); 145 FloatToQuantized<T2>(1.0f, *output_min, *output_max) 210 float output_min; variable [all...] |
/external/tensorflow/tensorflow/lite/experimental/micro/kernels/ |
depthwise_conv_test.cc | 116 std::initializer_list<int> output_dims_data, float output_min, 135 output_min, output_max), 234 const float output_min = -127.0f; local 289 F2Q(71, output_min, output_max), 290 F2Q(-34, output_min, output_max), 291 F2Q(99, output_min, output_max), 292 F2Q(-20, output_min, output_max), 293 F2Q(91, output_min, output_max), 294 F2Q(-26, output_min, output_max), 295 F2Q(127, output_min, output_max) 342 const float output_min = -127.0f; local 421 const float output_min = -127.0f; local [all...] |
fully_connected_test.cc | 109 std::initializer_list<int> output_dims_data, float output_min, 128 output_min, output_max), 272 const float output_min = -127.0f; local 321 F2Q(24, output_min, output_max), 322 F2Q(25, output_min, output_max), 323 F2Q(26, output_min, output_max), 324 F2Q(58, output_min, output_max), 325 F2Q(59, output_min, output_max), 326 F2Q(60, output_min, output_max), 329 output_min, output_max, // Output quantization range 343 const float output_min = -127.0f; local 414 const float output_min = -63.5f; local 511 const float output_min = -127.0f; local 582 const float output_min = -63.5f; local [all...] |
softmax_test.cc | 94 float output_min, float output_max, 107 output_min, output_max), 193 const float output_min = 0.0f; local 209 F2Q(0.011656231, output_min, output_max), 210 F2Q(0.031684921, output_min, output_max), 211 F2Q(0.086128544, output_min, output_max), 212 F2Q(0.234121657, output_min, output_max), 213 F2Q(0.636408647, output_min, output_max), 216 output_min, output_max, // Output quantized range.
|