Home | History | Annotate | Download | only in decoder

Lines Matching refs:mb_row

90 static void decode_macroblock(VP8D_COMP *pbi, MACROBLOCKD *xd, int mb_row, int mb_col)
117 /*mt_skip_recon_mb(pbi, xd, mb_row, mb_col);*/
120 vp8mt_build_intra_predictors_mbuv_s(pbi, xd, mb_row, mb_col);
121 vp8mt_build_intra_predictors_mby_s(pbi, xd, mb_row, mb_col);
136 vp8mt_build_intra_predictors_mbuv(pbi, xd, mb_row, mb_col);
140 vp8mt_build_intra_predictors_mby(pbi, xd, mb_row, mb_col);
142 vp8mt_intra_prediction_down_copy(pbi, xd, mb_row, mb_col);
185 vp8mt_predict_intra4x4(pbi, xd, b->bmi.mode, b->predictor, mb_row, mb_col, i);
239 int mb_row;
244 for (mb_row = ithread+1; mb_row < pc->mb_rows; mb_row += (pbi->decoding_thread_count + 1))
259 pbi->mb_row_di[ithread].mb_row = mb_row;
260 pbi->mb_row_di[ithread].mbd.current_bc = &pbi->mbc[mb_row%num_part];
262 last_row_current_mb_col = &pbi->mt_current_mb_col[mb_row -1];
264 recon_yoffset = mb_row * recon_y_stride * 16;
265 recon_uvoffset = mb_row * recon_uv_stride * 8;
271 xd->up_available = (mb_row != 0);
273 xd->mb_to_top_edge = -((mb_row * 16)) << 3;
274 xd->mb_to_bottom_edge = ((pc->mb_rows - 1 - mb_row) * 16) << 3;
321 decode_macroblock(pbi, xd, mb_row, mb_col);
325 if( mb_row != pc->mb_rows-1 )
328 vpx_memcpy((pbi->mt_yabove_row[mb_row + 1] + 32 + mb_col*16), (xd->dst.y_buffer + 15 * recon_y_stride), 16);
329 vpx_memcpy((pbi->mt_uabove_row[mb_row + 1] + 16 + mb_col*8), (xd->dst.u_buffer + 7 * recon_uv_stride), 8);
330 vpx_memcpy((pbi->mt_vabove_row[mb_row + 1] + 16 + mb_col*8), (xd->dst.v_buffer + 7 * recon_uv_stride), 8);
341 pbi->mt_yleft_col[mb_row][i] = xd->dst.y_buffer [i* recon_y_stride + 15];
344 pbi->mt_uleft_col[mb_row][i] = xd->dst.u_buffer [i* recon_uv_stride + 7];
345 pbi->mt_vleft_col[mb_row][i] = xd->dst.v_buffer [i* recon_uv_stride + 7];
369 if (mb_row > 0)
385 pbi->mt_current_mb_col[mb_row] = mb_col;
391 mb_row != pc->mb_rows-1)
398 pbi->mt_yabove_row[mb_row +1][lasty + i] = pbi->mt_yabove_row[mb_row +1][lasty -1];
399 pbi->mt_uabove_row[mb_row +1][lastuv + i] = pbi->mt_uabove_row[mb_row +1][lastuv -1];
400 pbi->mt_vabove_row[mb_row +1][lastuv + i] = pbi->mt_vabove_row[mb_row +1][lastuv -1];
414 if ((mbrd->mb_row == pbi->common.mb_rows-1) || ((mbrd->mb_row == pbi->common.mb_rows-2) && (pbi->common.mb_rows % (pbi->decoding_thread_count+1))==1))
643 /*int mb_row;
683 int mb_row;
726 for (mb_row = 0; mb_row < pc->mb_rows; mb_row += (pbi->decoding_thread_count + 1))
729 xd->current_bc = &pbi->mbc[mb_row%num_part];
731 /* vp8_decode_mb_row(pbi, pc, mb_row, xd); */
742 if (mb_row > 0)
743 last_row_current_mb_col = &pbi->mt_current_mb_col[mb_row -1];
746 recon_yoffset = mb_row * recon_y_stride * 16;
747 recon_uvoffset = mb_row * recon_uv_stride * 8;
751 xd->up_available = (mb_row != 0);
753 xd->mb_to_top_edge = -((mb_row * 16)) << 3;
754 xd->mb_to_bottom_edge = ((pc->mb_rows - 1 - mb_row) * 16) << 3;
758 if ( mb_row > 0 && (mb_col & (nsync-1)) == 0){
806 decode_macroblock(pbi, xd, mb_row, mb_col);
814 if(mb_row != pc->mb_rows-1)
816 vpx_memcpy((pbi->mt_yabove_row[mb_row +1] + 32 + mb_col*16), (xd->dst.y_buffer + 15 * recon_y_stride), 16);
817 vpx_memcpy((pbi->mt_uabove_row[mb_row +1] + 16 + mb_col*8), (xd->dst.u_buffer + 7 * recon_uv_stride), 8);
818 vpx_memcpy((pbi->mt_vabove_row[mb_row +1] + 16 + mb_col*8), (xd->dst.v_buffer + 7 * recon_uv_stride), 8);
829 pbi->mt_yleft_col[mb_row][i] = xd->dst.y_buffer [i* recon_y_stride + 15];
832 pbi->mt_uleft_col[mb_row][i] = xd->dst.u_buffer [i* recon_uv_stride + 7];
833 pbi->mt_vleft_col[mb_row][i] = xd->dst.v_buffer [i* recon_uv_stride + 7];
857 if (mb_row > 0)
872 pbi->mt_current_mb_col[mb_row] = mb_col;
878 if(mb_row != pc->mb_rows-1)
885 pbi->mt_yabove_row[mb_row +1][lasty + i] = pbi->mt_yabove_row[mb_row +1][lasty -1];
886 pbi->mt_uabove_row[mb_row +1][lastuv + i] = pbi->mt_uabove_row[mb_row +1][lastuv -1];
887 pbi->mt_vabove_row[mb_row +1][lastuv + i] = pbi->mt_vabove_row[mb_row +1][lastuv -1];