/external/tensorflow/tensorflow/core/kernels/ |
quantized_activation_ops_test.cc | 48 const int input_height = 4; local 49 Tensor input_float(DT_FLOAT, {input_height, input_width}); 53 Tensor expected_float(DT_FLOAT, {input_height, input_width}); 79 const int input_height = 4; local 80 Tensor input_float(DT_FLOAT, {input_height, input_width}); 84 Tensor expected_float(DT_FLOAT, {input_height, input_width});
|
quantized_pooling_ops_test.cc | 53 const int input_height = 4; local 56 Tensor input_float(DT_FLOAT, {1, input_height, input_width, input_channels}); 65 const int expected_height = input_height / stride; 98 const int input_height = 4; local 101 Tensor input_float(DT_FLOAT, {1, input_height, input_width, input_channels}); 110 const int expected_height = input_height / stride;
|
quantized_bias_add_op_test.cc | 53 const int input_height = 2; local 55 Tensor input_float(DT_FLOAT, {input_height, input_width}); 69 Tensor expected_float(DT_FLOAT, {input_height, input_width}); 103 const int input_height = 1; local 105 Tensor input_float(DT_FLOAT, {input_height, input_width}); 141 Tensor expected_float(DT_FLOAT, {input_height, input_width});
|
conv_ops_using_gemm.cc | 82 // input_data = [input_batches, input_height, input_width, input_depth] 89 int input_batches, int input_height, int input_width, 115 input_height + 1) / 121 ((output_height - 1) * stride_rows + filter_height - input_height) / 165 (in_y < input_height)) { 167 input_data[(batch * input_height * input_width * 212 int input_batches, int input_height, int input_width, 217 if ((input_batches <= 0) || (input_width <= 0) || (input_height <= 0) || 220 << input_batches << ", " << input_height << ", " 241 const int m = input_batches * input_height * input_width [all...] |
depthtospace_op_gpu.cu.cc | 37 const int input_height, const int input_width, 57 in_d + input_depth * (in_w + input_width * (in_h + input_height * b)); 152 const int input_height = input.dimension(1); local 164 input_height, input_width, input_depth, output_height, output_width, 179 const int input_height = input.dimension(2); local 182 const int input_area = input_width * input_height; 219 output_depth * input_height, output.data());
|
eigen_attention.h | 86 const Index input_height = in.dimension(2); local 89 eigen_assert(input_height > 0); 99 y *= input_height; 106 y += input_height / 2.0f; 138 } else if (offset_y + height_ >= input_height) { 139 glimpse_height = (std::max<Index>)(0, input_height - offset_y); 144 slice_extent[2] = std::min<Index>(input_height, slice_extent[2]); 168 DSizes<Index, 2> input_size(input_width, input_height);
|
spacetodepth_op_gpu.cu.cc | 35 const int input_height, const int input_width, 40 // inp_idx = d + input_depth * (w + input_width * (h + input_height * b)) 45 const int h = inp_idx3 % input_height; 46 const int b = inp_idx3 / input_height; 148 const int input_height = input.dimension(1); local 156 batch_size * input_height * input_width * input_depth; 160 input_height, input_width, input_depth, output_height, output_width,
|
quantized_batch_norm_op_test.cc | 64 const int input_height = 1; local 68 {input_batch, input_height, input_width, input_depth}); 122 TensorShape({input_batch, input_height, input_width, input_depth})); 161 const int input_height = 1; local 165 {input_batch, input_height, input_width, input_depth}); 219 TensorShape({input_batch, input_height, input_width, input_depth}));
|
quantized_conv_ops.cc | 48 // input_data = [input_batches, input_height, input_width, input_depth] 55 int input_batches, int input_height, int input_width, 93 ((output_height - 1) * stride + filter_height - input_height + 1) / 2; 98 ((output_height - 1) * stride + filter_height - input_height) / 2; 142 (in_y < input_height)) { 144 input_data[(batch * input_height * input_width * 201 int input_batches, int input_height, int input_width, 222 conv_functor(context, input_data, input_batches, input_height, 238 ((output_height - 1) * stride + filter_height - input_height + 1) / 2; 243 ((output_height - 1) * stride + filter_height - input_height) / 2 [all...] |
/external/tensorflow/tensorflow/examples/label_image/ |
label_image.py | 39 input_height=299, 59 resized = tf.image.resize_bilinear(dims_expander, [input_height, input_width]) 80 input_height = 299 variable 91 parser.add_argument("--input_height", type=int, help="input height") 105 if args.input_height: 106 input_height = args.input_height variable 121 input_height=input_height, variable
|
main.cc | 114 Status ReadTensorFromImageFile(const string& file_name, const int input_height, 166 Const(root.WithOpName("size"), {input_height, input_width})); 286 int32 input_height = 299; local 298 Flag("input_height", &input_height, 336 ReadTensorFromImageFile(image_path, input_height, input_width, input_mean,
|
/external/tensorflow/tensorflow/tools/benchmark/ |
benchmark_model_test.cc | 35 const int input_height = 10; local 37 input.shape = TensorShape({input_width, input_height}); 39 const TensorShape constant_shape({input_height, input_width});
|
/external/mesa3d/src/gallium/state_trackers/xvmc/tests/ |
xvmc_bench.c | 56 unsigned int input_height; member in struct:Config 72 config->input_height = DEFAULT_INPUT_HEIGHT; 88 if (sscanf(argv[++i], "%u", &config->input_height) != 1) 172 config->output_height = config->input_height; 202 mbh = align(config.input_height, MACROBLOCK_HEIGHT) / MACROBLOCK_HEIGHT; 210 config.input_height, 238 assert(XvMCCreateContext(display, port_num, surface_type_id, config.input_width, config.input_height, XVMC_DIRECT, &context) == Success); 278 assert(XvMCPutSurface(display, &surface, window, 0, 0, config.input_width, config.input_height, 0, 0, config.output_width, config.output_height, XVMC_FRAME_PICTURE) == Success); 287 printf("Input: %u,%u\nOutput: %u,%u\n", config.input_width, config.input_height, config.output_width, config.output_height);
|
/external/libxcam/tests/ |
test-image-blend.cpp | 42 static uint32_t input_height = 960; variable 211 input_height = atoi(optarg); 251 printf ("input/output height:%d\n", input_height); 258 output_height = input_height; 259 input_buf_info0.init (input_format, input_width0, input_height); 260 input_buf_info1.init (input_format, input_width1, input_height); 289 input_format, input_width0, input_height, XCAM_ALIGN_UP (input_width0, 16), XCAM_ALIGN_UP(input_height, 16)); 291 input_format, input_width1, input_height, XCAM_ALIGN_UP (input_width1, 16), XCAM_ALIGN_UP(input_height, 16)) [all...] |
test-video-stabilization.cpp | 76 uint32_t input_height = 1080; local 119 input_height = atoi(optarg); 120 output_height = input_height; 154 printf ("input height:%d\n", input_height); 202 input_buf_info.init (input_format, input_width, input_height);
|
/hardware/qcom/media/msm8974/mm-video-legacy/vidc/venc/inc/ |
video_encoder_test.h | 45 unsigned long input_height; member in struct:video_encoder_context
|
/hardware/qcom/media/msm8974/mm-video-v4l2/vidc/venc/inc/ |
video_encoder_test.h | 44 unsigned long input_height; member in struct:video_encoder_context
|
/hardware/qcom/media/msm8996/mm-video-v4l2/vidc/venc/inc/ |
video_encoder_test.h | 44 unsigned long input_height; member in struct:video_encoder_context
|
/hardware/qcom/media/msm8998/mm-video-v4l2/vidc/venc/inc/ |
video_encoder_test.h | 44 unsigned long input_height; member in struct:video_encoder_context
|
/packages/apps/Camera2/jni/ |
tinyplanet.cc | 93 int input_width, int input_height, 96 ImageRGBA input(input_image, input_width, input_height); 120 float py = (phi / PI_F) * input_height; 124 py = wrap(py, input_height);
|
/packages/apps/Gallery2/jni/filters/ |
tinyplanet.cc | 92 int input_width, int input_height, 95 ImageRGBA input(input_image, input_width, input_height); 119 float py = (phi / PI_F) * input_height; 123 py = wrap(py, input_height);
|
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/optimized/ |
multithreaded_conv.h | 117 int input_height, int input_width, int input_depth, 137 } else if (filter_height == input_height && filter_width == input_width && 153 ConstEigenTensor input(input_data, input_batches, input_height, 175 const int input_height = ArraySize(input_dims, 2); local 182 conv_functor(input_data, im2col_data, batches, input_height, input_width,
|
/external/tensorflow/tensorflow/contrib/lite/kernels/ |
space_to_depth.cc | 57 const int input_height = input->dims->data[1]; local 59 int output_height = input_height / block_size; 62 TF_LITE_ENSURE_EQ(context, input_height, output_height * block_size);
|
/external/tensorflow/tensorflow/contrib/lite/kernels/internal/reference/ |
depthwiseconv_float.h | 35 const int input_height = ArraySize(input_dims, 2); local 60 (in_y < input_height)) {
|
/external/opencv/ml/src/ |
mlcnn.cpp | 75 length(X)==<n_input_planes>*<input_height>*<input_width>, 129 if( img_size != first_layer->input_height*first_layer->input_width ) \ 228 const int img_height = first_layer->input_height; 341 img_height = first_layer->input_height; 514 if( layer->input_height != 1 || layer->output_height != 1 || 521 prev_layer->output_height != layer->input_height || 575 int n_input_planes, int input_height, int input_width, 589 input_height < 1 || input_width < 1 || 591 input_height < output_height || 608 layer->input_height = input_height [all...] |