HomeSort by relevance Sort by last modified time
    Searched defs:output_min (Results 1 - 25 of 33) sorted by null

1 2

  /external/tensorflow/tensorflow/core/kernels/
quantized_reshape_op.cc 36 Tensor* output_min = nullptr; variable
37 OP_REQUIRES_OK(ctx, ctx->allocate_output(1, TensorShape({}), &output_min));
38 output_min->flat<float>()(0) = input_min_float;
quantize_op_test.cc 292 const float output_min = GetOutput(1)->flat<float>()(0); local
294 EXPECT_NEAR(0.0f, output_min, 1e-5f);
314 const float output_min = GetOutput(1)->flat<float>()(0); local
316 EXPECT_NEAR(0.0f, output_min, 1e-5f);
336 const float output_min = GetOutput(1)->flat<float>()(0); local
338 EXPECT_NEAR(0.0f, output_min, 1e-5f);
358 const float output_min = GetOutput(1)->flat<float>()(0); local
360 EXPECT_NEAR(-0.3f, output_min, 1e-5f);
quantized_activation_ops.cc 51 Tensor* output_min = nullptr; variable
52 OP_REQUIRES_OK(context, context->allocate_output(1, {}, &output_min));
53 output_min->flat<float>()(0) = min_input;
89 Tensor* output_min = nullptr; variable
90 OP_REQUIRES_OK(context, context->allocate_output(1, {}, &output_min));
91 output_min->flat<float>()(0) = min_input;
quantized_activation_ops_test.cc 62 const float output_min = GetOutput(1)->flat<float>()(0); local
65 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
93 const float output_min = GetOutput(1)->flat<float>()(0); local
96 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
quantized_pooling_ops.cc 94 Tensor* output_min = nullptr; variable
95 OP_REQUIRES_OK(context, context->allocate_output(1, {}, &output_min));
96 output_min->flat<float>()(0) = min_input;
118 Tensor* output_min = nullptr; variable
119 OP_REQUIRES_OK(context, context->allocate_output(1, {}, &output_min));
120 output_min->flat<float>()(0) = min_input;
requantize.cc 49 Tensor* output_min = nullptr; variable
50 OP_REQUIRES_OK(ctx, ctx->allocate_output(1, TensorShape({}), &output_min));
92 output_min->flat<float>().setConstant(requested_output_min_float);
quantize_down_and_shrink_range.cc 47 Tensor* output_min = nullptr; variable
48 OP_REQUIRES_OK(ctx, ctx->allocate_output(1, TensorShape({}), &output_min));
96 output_min->flat<float>().setConstant(actual_min_float);
quantized_bias_add_op.cc 84 Tensor* output_min = nullptr; variable
85 OP_REQUIRES_OK(context, context->allocate_output(1, {}, &output_min));
86 output_min->flat<float>()(0) = total_min;
quantized_bias_add_op_test.cc 83 const float output_min = GetOutput(1)->flat<float>()(0); local
86 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max);
165 const float output_min = GetOutput(1)->flat<float>()(0); local
168 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max);
quantized_pooling_ops_test.cc 76 const float output_min = GetOutput(1)->flat<float>()(0); local
79 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
121 const float output_min = GetOutput(1)->flat<float>()(0); local
124 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
requantization_range_op.cc 51 Tensor* output_min = nullptr; variable
52 OP_REQUIRES_OK(ctx, ctx->allocate_output(0, TensorShape({}), &output_min));
68 output_min->flat<float>().setConstant(used_min_float);
mkl_quantized_concat_op_test.cc 150 const float output_min = GetOutput(1)->flat<float>()(0); local
153 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
224 const float output_min = GetOutput(1)->flat<float>()(0); local
227 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
mkl_quantized_pooling_ops_test.cc 125 const float output_min = GetOutput(1)->flat<float>()(0); local
128 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
193 const float output_min = GetOutput(1)->flat<float>()(0); local
196 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
mkl_requantization_range_per_channel_op.cc 102 Tensor* output_min = nullptr; variable
104 OP_REQUIRES_OK(ctx, ctx->allocate_output(kOutputMinIndex, {}, &output_min));
106 output_min->flat<float>()(0) = is_non_negative ? 0.0f : -out_min_max;
mkl_requantize_ops_test.cc 163 const float output_min = GetOutput(0)->flat<float>()(0); local
165 EXPECT_NEAR(-14.8217, output_min, 0.002);
224 const float output_min = GetOutput(0)->flat<float>()(0); local
226 EXPECT_NEAR(-6.0, output_min, 0.002); // Values are aligned with clip_value.
293 const float output_min = GetOutput(1)->flat<float>()(0); local
295 EXPECT_NEAR(range_op_output_min, output_min, 0.002);
quantized_conv_ops_test.cc 127 const float output_min = GetOutput(1)->flat<float>()(0); local
130 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max);
318 const float output_min = GetOutput(1)->flat<float>()(0); local
321 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max);
quantized_matmul_op_test.cc 351 const float output_min = GetOutput(1)->flat<float>()(0); local
354 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max);
mkl_quantized_conv_ops_test.cc 198 const float output_min = GetOutput(1)->flat<float>()(0); local
201 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max);
mkl_requantize_per_channel_op.cc 134 Tensor* output_min = nullptr; variable
137 ctx->allocate_output(kOutputMinIndex, {}, &output_min));
141 output_min->flat<float>()(0) = input_requested_min_float;
quantized_batch_norm_op_test.cc 127 const float output_min = GetOutput(1)->flat<float>()(0); local
130 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max);
235 const float output_min = GetOutput(1)->flat<float>()(0); local
238 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max);
quantized_concat_op_test.cc 111 const float output_min = GetOutput(1)->flat<float>()(0); local
114 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
177 const float output_min = GetOutput(1)->flat<float>()(0); local
180 QuantizedTensorToFloat<qint32>(output_quantized, output_min, output_max);
239 const float output_min = GetOutput(1)->flat<float>()(0); local
242 QuantizedTensorToFloat<quint8>(output_quantized, output_min, output_max);
quantized_batch_norm_op.cc 38 Tensor* output, float* output_min, float* output_max) {
49 *output_min = std::numeric_limits<float>::max();
78 *output_min = std::min(output_value, *output_min);
82 FloatToQuantized<T2>(output_value, *output_min, *output_max);
101 Tensor* output, float* output_min, float* output_max) {
115 *output_min = -(1 << 20);
139 FloatToQuantized<T2>(scale_value, *output_min, *output_max);
141 FloatToQuantized<T2>(offset_value, *output_min, *output_max);
145 FloatToQuantized<T2>(1.0f, *output_min, *output_max)
210 float output_min; variable
    [all...]
  /external/tensorflow/tensorflow/lite/experimental/micro/kernels/
depthwise_conv_test.cc 116 std::initializer_list<int> output_dims_data, float output_min,
135 output_min, output_max),
234 const float output_min = -127.0f; local
289 F2Q(71, output_min, output_max),
290 F2Q(-34, output_min, output_max),
291 F2Q(99, output_min, output_max),
292 F2Q(-20, output_min, output_max),
293 F2Q(91, output_min, output_max),
294 F2Q(-26, output_min, output_max),
295 F2Q(127, output_min, output_max)
342 const float output_min = -127.0f; local
421 const float output_min = -127.0f; local
    [all...]
fully_connected_test.cc 109 std::initializer_list<int> output_dims_data, float output_min,
128 output_min, output_max),
272 const float output_min = -127.0f; local
321 F2Q(24, output_min, output_max),
322 F2Q(25, output_min, output_max),
323 F2Q(26, output_min, output_max),
324 F2Q(58, output_min, output_max),
325 F2Q(59, output_min, output_max),
326 F2Q(60, output_min, output_max),
329 output_min, output_max, // Output quantization range
343 const float output_min = -127.0f; local
414 const float output_min = -63.5f; local
511 const float output_min = -127.0f; local
582 const float output_min = -63.5f; local
    [all...]
softmax_test.cc 94 float output_min, float output_max,
107 output_min, output_max),
193 const float output_min = 0.0f; local
209 F2Q(0.011656231, output_min, output_max),
210 F2Q(0.031684921, output_min, output_max),
211 F2Q(0.086128544, output_min, output_max),
212 F2Q(0.234121657, output_min, output_max),
213 F2Q(0.636408647, output_min, output_max),
216 output_min, output_max, // Output quantized range.

Completed in 397 milliseconds

1 2