Home | History | Annotate | Download | only in encoder

Lines Matching defs:xd

265     MACROBLOCKD *xd = &x->e_mbd;
283 xd->up_available = (mb_row != 0);
290 xd->dst.y_buffer = new_yv12->y_buffer + recon_yoffset;
291 xd->left_available = (mb_col != 0);
319 vp8_extend_mb_row(new_yv12, xd->dst.y_buffer + 16,
320 xd->dst.u_buffer + 8, xd->dst.v_buffer + 8);
365 MACROBLOCKD *xd,
404 xd->above_context = cm->above_context;
406 xd->up_available = (mb_row != 0);
416 xd->mb_to_top_edge = -((mb_row * 16) << 3);
417 xd->mb_to_bottom_edge = ((cm->mb_rows - 1 - mb_row) * 16) << 3;
440 xd->mb_to_left_edge = -((mb_col * 16) << 3);
441 xd->mb_to_right_edge = ((cm->mb_cols - 1 - mb_col) * 16) << 3;
450 xd->dst.y_buffer = cm->yv12_fb[dst_fb_idx].y_buffer + recon_yoffset;
451 xd->dst.u_buffer = cm->yv12_fb[dst_fb_idx].u_buffer + recon_uvoffset;
452 xd->dst.v_buffer = cm->yv12_fb[dst_fb_idx].v_buffer + recon_uvoffset;
453 xd->left_available = (mb_col != 0);
482 if (xd->segmentation_enabled)
484 /* Code to set segment id in xd->mbmi.segment_id for current MB
488 xd->mode_info_context->mbmi.segment_id = cpi->segmentation_map[map_index+mb_col];
490 xd->mode_info_context->mbmi.segment_id = 0;
496 xd->mode_info_context->mbmi.segment_id = 0;
504 y_modes[xd->mbmi.mode] ++;
512 inter_y_modes[xd->mbmi.mode] ++;
514 if (xd->mbmi.mode == SPLITMV)
518 for (b = 0; b < xd->mbmi.partition_count; b++)
529 if (xd->mode_info_context->mbmi.mode == ZEROMV &&
530 xd->mode_info_context->mbmi.ref_frame == LAST_FRAME) {
540 * If cyclic update enabled then copy xd->mbmi.segment_id; (which
547 xd->segmentation_enabled))
549 cpi->segmentation_map[map_index+mb_col] = xd->mode_info_context->mbmi.segment_id;
558 if (xd->mode_info_context->mbmi.segment_id)
560 else if ((xd->mode_info_context->mbmi.mode == ZEROMV) && (xd->mode_info_context->mbmi.ref_frame == LAST_FRAME))
595 segment_counts[xd->mode_info_context->mbmi.segment_id] ++;
598 xd->mode_info_context++;
600 xd->above_context++;
605 xd->dst.y_buffer + 16,
606 xd->dst.u_buffer + 8,
607 xd->dst.v_buffer + 8);
615 xd->mode_info_context++;
623 MACROBLOCKD *const xd = & x->e_mbd;
635 xd->mode_info_context = cm->mi;
636 xd->mode_info_stride = cm->mode_info_stride;
638 xd->frame_type = cm->frame_type;
646 xd->pre = cm->yv12_fb[cm->lst_fb_idx];
647 xd->dst = cm->yv12_fb[cm->new_fb_idx];
654 xd->mode_info_context->mbmi.mode = DC_PRED;
655 xd->mode_info_context->mbmi.uv_mode = DC_PRED;
657 xd->left_context = &cm->left_context;
684 xd->fullpixel_mask = 0xffffffff;
686 xd->fullpixel_mask = 0xfffffff8;
731 MACROBLOCKD *const xd = & x->e_mbd;
754 xd->subpixel_predict = vp8_sixtap_predict4x4;
755 xd->subpixel_predict8x4 = vp8_sixtap_predict8x4;
756 xd->subpixel_predict8x8 = vp8_sixtap_predict8x8;
757 xd->subpixel_predict16x16 = vp8_sixtap_predict16x16;
761 xd->subpixel_predict = vp8_bilinear_predict4x4;
762 xd->subpixel_predict8x4 = vp8_bilinear_predict8x4;
763 xd->subpixel_predict8x8 = vp8_bilinear_predict8x8;
764 xd->subpixel_predict16x16 = vp8_bilinear_predict16x16;
776 xd->mode_info_context = cm->mi;
842 encode_mb_row(cpi, cm, mb_row, x, xd, &tp, segment_counts, &totalrate);
849 xd->mode_info_context += xd->mode_info_stride * cpi->encoding_thread_count;
850 x->partition_info += xd->mode_info_stride * cpi->encoding_thread_count;
867 if (xd->segmentation_enabled)
871 if (xd->segmentation_enabled)
935 encode_mb_row(cpi, cm, mb_row, x, xd, &tp, segment_counts, &totalrate);
964 if (xd->segmentation_enabled && xd->update_mb_segmentation_map)
970 vpx_memset(xd->mb_segment_tree_probs, 255 , sizeof(xd->mb_segment_tree_probs));
976 xd->mb_segment_tree_probs[0] = ((segment_counts[0] + segment_counts[1]) * 255) / tot_count;
982 xd->mb_segment_tree_probs[1] = (segment_counts[0] * 255) / tot_count;
988 xd->mb_segment_tree_probs[2] = (segment_counts[2] * 255) / tot_count;
993 if (xd->mb_segment_tree_probs[i] == 0)
994 xd->mb_segment_tree_probs[i] = 1;
1124 const MACROBLOCKD *xd = & x->e_mbd;
1125 const MB_PREDICTION_MODE m = xd->mode_info_context->mbmi.mode;
1126 const MB_PREDICTION_MODE uvm = xd->mode_info_context->mbmi.uv_mode;
1141 ++ bct[xd->block[b].bmi.mode];
1179 MACROBLOCKD *xd = &x->e_mbd;
1204 if (xd->mode_info_context->mbmi.mode != B_PRED)
1205 vp8_inverse_transform_mby(xd);
1208 (xd->qcoeff+16*16, xd->dequant_uv,
1209 xd->dst.u_buffer, xd->dst.v_buffer,
1210 xd->dst.uv_stride, xd->eobs+16);
1226 MACROBLOCKD *const xd = &x->e_mbd;
1233 if (xd->segmentation_enabled)
1234 x->encode_breakout = cpi->segment_encode_breakout[xd->mode_info_context->mbmi.segment_id];
1297 if (xd->segmentation_enabled)
1303 if ((xd->mode_info_context->mbmi.segment_id == 1) &&
1304 ((xd->mode_info_context->mbmi.ref_frame != LAST_FRAME) || (xd->mode_info_context->mbmi.mode != ZEROMV)))
1306 xd->mode_info_context->mbmi.segment_id = 0;
1322 if ( xd->mode_info_context->mbmi.ref_frame != INTRA_FRAME )
1324 if (xd->mode_info_context->mbmi.mode == ZEROMV)
1326 if (xd->mode_info_context->mbmi.ref_frame != LAST_FRAME &&
1332 else if (xd->mode_info_context->mbmi.mode == SPLITMV)
1345 x->count_mb_ref_frame_usage[xd->mode_info_context->mbmi.ref_frame] ++;
1347 if (xd->mode_info_context->mbmi.ref_frame == INTRA_FRAME)
1351 if (xd->mode_info_context->mbmi.mode == B_PRED)
1366 if (xd->mode_info_context->mbmi.ref_frame == LAST_FRAME)
1368 else if (xd->mode_info_context->mbmi.ref_frame == GOLDEN_FRAME)
1373 xd->pre.y_buffer = cpi->common.yv12_fb[ref_fb_idx].y_buffer + recon_yoffset;
1374 xd->pre.u_buffer = cpi->common.yv12_fb[ref_fb_idx].u_buffer + recon_uvoffset;
1375 xd->pre.v_buffer = cpi->common.yv12_fb[ref_fb_idx].v_buffer + recon_uvoffset;
1382 vp8_build_inter16x16_predictors_mb(xd, xd->dst.y_buffer,
1383 xd->dst.u_buffer, xd->dst.v_buffer,
1384 xd->dst.y_stride, xd->dst.uv_stride);
1392 if (xd->mode_info_context->mbmi.mode != B_PRED)
1393 vp8_inverse_transform_mby(xd);
1396 (xd->qcoeff+16*16, xd->dequant_uv,
1397 xd->dst.u_buffer, xd->dst.v_buffer,
1398 xd->dst.uv_stride, xd->eobs+16);
1403 xd->mode_info_context->mbmi.mb_skip_coeff = 1;
1408 vp8_fix_contexts(xd);