/external/tensorflow/tensorflow/core/kernels/ |
quantized_instance_norm_test.cc | 112 float out_min = outputs[1].flat<float>()(0); local 114 float out_scale = (out_max - out_min) / 255.0f; 117 (expected.flat<float>() - (out_min + out_scale * out.cast<float>()))
|
quantized_instance_norm.cc | 199 const float32x4_t out_min = vdupq_n_f32(minimum); local 235 vcvtq_s32_f32(vmulq_n_f32(vsubq_f32(normed, out_min), out_scale));
|
quantization_utils_test.cc | 161 const float out_min = -29.1234; local 166 qint32 low = FloatToQuantized<qint32>(out_min, in_min, in_max); 171 TestRequantizeMany(eigen_device, in_min, in_max, out_min, out_max, vals); [all...] |
quantized_resize_bilinear_op.cc | 705 Tensor* out_min = nullptr; variable 706 OP_REQUIRES_OK(context, context->allocate_output(1, {}, &out_min)); 707 out_min->flat<float>()(0) = in_min;
|