/external/libhevc/decoder/ |
ihevcd_ittiam_logo.h | 116 UWORD32 y_pos, 122 #define INSERT_LOGO(buf_y, buf_u, buf_v, stride, x_pos, y_pos, yuv_fmt,disp_wd,disp_ht) ihevcd_insert_logo(buf_y, buf_u, buf_v, stride, x_pos, y_pos, yuv_fmt,disp_wd,disp_ht); 124 #define INSERT_LOGO(buf_y, buf_u, buf_v, stride, x_pos, y_pos, yuv_fmt,disp_wd,disp_ht)
|
/external/tensorflow/tensorflow/examples/android/jni/object_tracking/ |
keypoint_detector.cc | 52 const int y_pos = keypoint->pos_.y * 2; local 54 const int curr_color[] = {u_data[y_pos][x_pos], v_data[y_pos][x_pos]}; 122 const int y_pos = candidate.pos_.y; local 124 MarkImage(x_pos, y_pos, distance, interest_map_.get()); 134 const int y_pos = candidate.pos_.y; local 137 !interest_map_->ValidPixel(x_pos, y_pos)) { 141 if (!(*interest_map_)[y_pos][x_pos]) { 146 MarkImage(x_pos, y_pos, distance, interest_map_.get());
|
optical_flow.cc | 100 const float y_pos = Clip(src_top_real + y, 0.0f, real_y_max); local 105 *vals_I_ptr++ = img_I.GetPixelInterp(x_pos, y_pos); 106 *vals_I_x_ptr++ = I_x.GetPixelInterp(x_pos, y_pos); 107 *vals_I_y_ptr++ = I_y.GetPixelInterp(x_pos, y_pos); 151 const float y_pos = Clip(top_real + win_y, 0.0f, real_y_max); local 154 *vals_J_ptr++ = img_J.GetPixelInterp(x_pos, y_pos);
|
/external/webrtc/talk/media/base/ |
fakevideorenderer.h | 123 const uint8_t* y_pos = y_plane; local 127 uint8_t y_value = *(y_pos + j); 132 y_pos += y_pitch;
|
/external/webp/src/dec/ |
webpi_dec.h | 31 typedef int (*OutputRowFunc)(WebPDecParams* const p, int y_pos,
|
vp8l_dec.c | 588 static void ConvertToYUVA(const uint32_t* const src, int width, int y_pos, 593 WebPConvertARGBToY(src, buf->y + y_pos * buf->y_stride, width); 597 uint8_t* const u = buf->u + (y_pos >> 1) * buf->u_stride; 598 uint8_t* const v = buf->v + (y_pos >> 1) * buf->v_stride; 601 WebPConvertARGBToUV(src, u, v, width, !(y_pos & 1)); 605 uint8_t* const a = buf->a + y_pos * buf->a_stride; 614 static int ExportYUVA(const VP8LDecoder* const dec, int y_pos) { 622 ConvertToYUVA(src, dst_width, y_pos, dec->output_); 623 ++y_pos; 632 int y_pos = dec->last_out_row_ local 651 int y_pos = dec->last_out_row_; local [all...] |
io_dec.c | 355 static int ExportRGB(WebPDecParams* const p, int y_pos) { 359 uint8_t* dst = buf->rgba + y_pos * buf->stride; 365 assert(y_pos + num_lines_out < p->output->height); 404 static int ExportAlpha(WebPDecParams* const p, int y_pos, int max_lines_out) { 406 uint8_t* const base_rgba = buf->rgba + y_pos * buf->stride; 418 assert(y_pos + num_lines_out < p->output->height); 431 static int ExportAlphaRGBA4444(WebPDecParams* const p, int y_pos, 434 uint8_t* const base_rgba = buf->rgba + y_pos * buf->stride; 449 assert(y_pos + num_lines_out < p->output->height);
|
/frameworks/av/media/libstagefright/codecs/m4v_h263/dec/src/ |
dcac_prediction.cpp | 42 int y_pos = video->mbnum_row; local 83 if (!comp && y_pos && !(video->headerInfo.Mode[mbnum-nMBPerRow]&INTRA_MASK)) /* not intra */ 102 if (x_pos == 0 && y_pos == 0) 116 else if (y_pos == 0) 218 int y_pos = video->mbnum_row; local 261 if (x_pos == 0 && y_pos == 0) 274 else if (y_pos == 0)
|
combined_decode.cpp | 553 int y_pos = video->mbnum_row; local 555 int32 offset = (int32)(y_pos << 4) * width + (x_pos << 4); 563 pp_mod[0] = video->pstprcTypCur + (y_pos << 1) * (MB_in_width << 1) + (x_pos << 1);
|
datapart_decode.cpp | 637 int y_pos = video->mbnum_row; local 653 pp_mod[0] = video->pstprcTypCur + (y_pos << 1) * (MB_in_width << 1) + (x_pos << 1); 715 offset = (int32)(y_pos << 4) * width + (x_pos << 4);
|
block_idct.cpp | 230 int y_pos = video->mbnum_row; local 235 offset = (int32)(y_pos << 4) * width + (x_pos << 4);
|
vlc_decode.cpp | 145 int y_pos = video->mbnum_row; local 147 int pos = (x_pos + y_pos * doubleWidth) << 1; [all...] |
/external/freetype/src/sfnt/ |
ttsbit.c | 693 FT_Int y_pos, 702 FT_Int y_pos, 711 FT_Int y_pos, 736 y_pos < 0 || (FT_UInt)( y_pos + height ) > bit_height ) 752 line += y_pos * pitch + ( x_pos >> 3 ); 853 FT_Int y_pos, 879 y_pos < 0 || (FT_UInt)( y_pos + height ) > bit_height ) 903 line += y_pos * pitch + ( x_pos >> 3 ) [all...] |
/device/google/cuttlefish_common/host/frontend/vnc_server/ |
vnc_client_connection.cpp | 165 return lhs.x_pos == rhs.x_pos && lhs.y_pos == rhs.y_pos && 502 auto y_pos = uint16_tAt(&msg[3]); local 506 std::tie(x_pos, y_pos) = 507 std::make_pair(ActualScreenWidth() - y_pos, x_pos); 510 virtual_inputs_->HandlePointerEvent(button_mask, x_pos, y_pos);
|
vnc_client_connection.h | 63 std::uint16_t y_pos; member in struct:cvd::vnc::VncClientConnection::FrameBufferUpdateRequest
|
/hardware/libhardware/include/hardware/ |
hwcomposer.h | [all...] |
/external/freetype/src/pfr/ |
pfrgload.c | 568 FT_Int x_pos, y_pos; local 628 y_pos = 0; 669 y_pos = PFR_NEXT_SHORT( p ); 674 y_pos += PFR_NEXT_INT8( p ); 682 subglyph->y_delta = y_pos;
|
/external/tensorflow/tensorflow/cc/framework/ |
gradient_checker.cc | 237 std::vector<Tensor> y_pos; local 238 TF_RETURN_IF_ERROR(EvaluateGraph(&session, xs, ys, x_datas, &y_pos)); 247 auto y_pos_flat = y_pos[y_idx].flat<Y_T>();
|
/external/tensorflow/tensorflow/python/ops/ |
gradient_checker.py | 184 y_pos = y.eval(feed_dict=_extra_feeds(extra_feed_dict, {x: x_pos})) 187 diff = (y_pos - y_neg) / scale
|
/external/libxcam/modules/ocl/ |
cl_utils.cpp | 32 float y_pos; member in struct:XCam::NV12Pixel 39 : x_pos (0.0f), y_pos (0.0f)
|
/frameworks/av/media/libstagefright/codecs/m4v_h263/enc/src/ |
vlc_encode.cpp | 2027 Int y_pos = video->outputMB->mb_y; local [all...] |
/external/libmpeg2/decoder/ |
impeg2d_api_main.c | 77 #define INSERT_LOGO(buf_y, buf_u, buf_v, stride, x_pos, y_pos, yuv_fmt,disp_wd,disp_ht) impeg2_insert_logo(buf_y, buf_u, buf_v, stride, x_pos, y_pos, yuv_fmt,disp_wd,disp_ht); 79 #define INSERT_LOGO(buf_y, buf_u, buf_v, stride, x_pos, y_pos, yuv_fmt,disp_wd,disp_ht) [all...] |
/external/tensorflow/tensorflow/contrib/cudnn_rnn/python/kernel_tests/ |
cudnn_rnn_test.py | [all...] |