/external/libvpx/libvpx/vp8/encoder/ |
segmentation.c | 15 int mb_row, mb_col; local 29 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) {
|
mr_dissim.c | 77 int mb_col; local 84 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 111 if (mb_col < (cm->mb_cols - 1)) { 125 if (mb_col < (cm->mb_cols - 1) && mb_row < (cm->mb_rows - 1)) { 135 if (mb_col < (cm->mb_cols - 1)) { 149 if (mb_col < (cm->mb_cols - 1) && mb_row < (cm->mb_rows - 1)) {
|
ethreading.c | 77 int mb_col; local 109 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 110 if (((mb_col - 1) % nsync) == 0) { 111 vpx_atomic_store_release(current_mb_col, mb_col - 1); 114 if (mb_row && !(mb_col & (nsync - 1))) { 115 vp8_atomic_spin_wait(mb_col, last_row_current_mb_col, nsync); 126 xd->mb_to_left_edge = -((mb_col * 16) << 3); 127 xd->mb_to_right_edge = ((cm->mb_cols - 1 - mb_col) * 16) << 3 [all...] |
temporal_filter.c | 191 int mb_col, mb_row; local 226 for (mb_col = 0; mb_col < mb_cols; ++mb_col) { 234 cpi->mb.mv_col_min = -((mb_col * 16) + (16 - 5)); 235 cpi->mb.mv_col_max = ((cpi->common.mb_cols - 1 - mb_col) * 16) + (16 - 5);
|
encodeframe.c | 104 int mb_row, int mb_col) { 108 int use_dc_pred = (mb_col || mb_row) && (!mb_col || !mb_row); 183 int mb_row, mb_col; local 200 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 245 int mb_row, mb_col; local 257 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) 329 int mb_col; local [all...] |
bitstream.c | 457 int mb_col = -1; local 459 while (++mb_col < pc->mb_cols) { 470 xd->mb_to_left_edge = -((mb_col * 16) << 3); 471 xd->mb_to_right_edge = ((pc->mb_cols - 1 - mb_col) * 16) << 3; 621 int mb_col = -1; local 623 while (++mb_col < c->mb_cols) { [all...] |
firstpass.c | 473 int mb_row, mb_col; local 560 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 563 int use_dc_pred = (mb_col || mb_row) && (!mb_col || !mb_row); 568 xd->left_available = (mb_col != 0); 593 x->mv_col_min = -((mb_col * 16) + (VP8BORDERINPIXELS - 16)); 595 ((cm->mb_cols - 1 - mb_col) * 16) + (VP8BORDERINPIXELS - 16); 709 if (mb_col < cm->mb_cols / 2) [all...] |
onyx_if.c | 612 int mb_row, mb_col, num_bl; local 625 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) { 626 const int bl_index = mb_row * cm->mb_cols + mb_col; 643 for (mb_col = 1; mb_col < cm->mb_cols - 1; mb_col++) { 644 const int bl_index = mb_row * cm->mb_cols + mb_col; 651 int bl_neighbor_index = (mb_row + mi) * cm->mb_cols + mb_col + mj 4324 int mb_col; local 4350 int mb_col; local [all...] |
/external/libvpx/libvpx/vp8/common/ |
debugmodes.c | 17 int mb_col; local 26 for (mb_col = 0; mb_col < cols; ++mb_col) { 42 for (mb_col = 0; mb_col < cols; ++mb_col) { 59 for (mb_col = 0; mb_col < cols; ++mb_col) { [all...] |
vp8_skin_detection.c | 72 int i, j, mb_row, mb_col, num_bl; local 92 for (mb_col = 0; mb_col < cm->mb_cols; mb_col += 1) {
|
vp8_loopfilter.c | 171 int mb_col; local 177 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 195 if (mb_col > 0) 225 int mb_col; local 230 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 242 if (mb_col > 0 273 int mb_col; local 394 int mb_col; local 484 int mb_col; local [all...] |
mfqe.c | 238 int mb_col; local 256 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) {
|
/external/libvpx/libvpx/vp8/decoder/ |
onyxd_if.c | 429 int mb_row, mb_col; local 432 for (mb_col = 0; mb_col < oci->mb_cols; mb_col++, mi++) {
|
decodemv.c | 530 int mb_col = -1; local 535 while (++mb_col < pbi->common.mb_cols) { 537 int mb_num = mb_row * pbi->common.mb_cols + mb_col;
|
error_concealment.c | 96 * (mb_row, mb_col) in macroblocks, which are being overlapped by a given 103 int mb_col, int first_blk_row, 105 /* Find the blocks within this MB (defined by mb_row, mb_col) which are 111 const int rel_ol_blk_col = first_blk_col - mb_col * 4; 124 int end_col = VPXMIN(4 + mb_col * 4 - first_blk_col, 2); 137 if (new_col < (mb_col * 16) << 3) end_col = 1; 277 int mb_row, int mb_col, int mb_rows, 285 4 * mb_row + sub_row, 4 * mb_col + sub_col); 295 int mb_row, mb_col; local 299 for (mb_col = 0; mb_col < mb_cols; ++mb_col) [all...] |
threading.c | 296 int mb_col; local 359 for (mb_col = 0; mb_col < pc->mb_cols; ++mb_col) { 360 if (((mb_col - 1) % nsync) == 0) { 361 vpx_atomic_store_release(current_mb_col, mb_col - 1); 364 if (mb_row && !(mb_col & (nsync - 1))) { 365 vp8_atomic_spin_wait(mb_col, last_row_current_mb_col, nsync); 372 xd->mb_to_left_edge = -((mb_col * 16) << 3); 373 xd->mb_to_right_edge = ((pc->mb_cols - 1 - mb_col) * 16) << 3 [all...] |
decodeframe.c | 446 int mb_row, mb_col; local 526 for (mb_col = 0; mb_col < pc->mb_cols; ++mb_col) { 531 xd->mb_to_left_edge = -((mb_col * 16) << 3); 532 xd->mb_to_right_edge = ((pc->mb_cols - 1 - mb_col) * 16) << 3; 551 vp8_interpolate_motion(xd, mb_row, mb_col, pc->mb_rows, pc->mb_cols); [all...] |
/external/libvpx/libvpx/vp9/encoder/ |
vp9_mbgraph.c | 27 int mb_col) { 70 vp9_build_inter_predictors_sby(xd, mb_row, mb_col, BLOCK_16X16); 77 int_mv *dst_mv, int mb_row, int mb_col) { 91 tmp_err = do_16x16_motion_iteration(cpi, ref_mv, &tmp_mv, mb_row, mb_col); 104 do_16x16_motion_iteration(cpi, &zero_ref_mv, &tmp_mv, mb_row, mb_col); 163 int mb_col) { 188 &stats->ref[GOLDEN_FRAME].m.mv, mb_row, mb_col); 220 int mb_col, mb_row, offset = 0; local 255 for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) 289 int mb_col, mb_row, offset, i; local [all...] |
vp9_temporal_filter.c | 295 int mb_col; local 341 for (mb_col = mb_col_start; mb_col < mb_col_end; mb_col++) { 349 td->mb.mv_limits.col_min = -((mb_col * 16) + (17 - 2 * VP9_INTERP_EXTEND)); 351 ((mb_cols - 1 - mb_col) * 16) + (17 - 2 * VP9_INTERP_EXTEND); 405 mb_col * 16, mb_row * 16);
|
vp9_firstpass.c | 467 static BLOCK_SIZE get_bsize(const VP9_COMMON *cm, int mb_row, int mb_col) { 468 if (2 * mb_col + 1 < cm->mi_cols) { 696 int mb_row, mb_col; local 701 for (mb_col = 0; mb_col < cm->mb_cols; ++mb_col) { 702 const int mb_index = mb_row * cm->mb_cols + mb_col; 832 int mb_col; local [all...] |
vp9_encodeframe.c | 3209 int mb_col = mi_col >> 1; local 3304 int mb_col = mi_col >> 1; local [all...] |