/external/libvpx/libvpx/vp9/common/ |
vp9_seg_common.h | 53 int segment_id, 55 return seg->enabled && (seg->feature_mask[segment_id] & (1 << feature_id)); 60 void vp9_enable_segfeature(struct segmentation *seg, int segment_id, 67 void vp9_set_segdata(struct segmentation *seg, int segment_id, 70 static INLINE int get_segdata(const struct segmentation *seg, int segment_id, 72 return seg->feature_data[segment_id][feature_id];
|
vp9_seg_common.c | 34 void vp9_enable_segfeature(struct segmentation *seg, int segment_id, 36 seg->feature_mask[segment_id] |= 1 << feature_id; 47 void vp9_set_segdata(struct segmentation *seg, int segment_id, 55 seg->feature_data[segment_id][feature_id] = seg_data;
|
vp9_quant_common.h | 29 int vp9_get_qindex(const struct segmentation *seg, int segment_id,
|
vp9_pred_common.h | 30 int x, y, segment_id = MAX_SEGMENTS; local 34 segment_id = 35 VPXMIN(segment_id, segment_ids[mi_offset + y * cm->mi_cols + x]); 37 assert(segment_id >= 0 && segment_id < MAX_SEGMENTS); 38 return segment_id;
|
vp9_quant_common.c | 196 int vp9_get_qindex(const struct segmentation *seg, int segment_id, 198 if (segfeature_active(seg, segment_id, SEG_LVL_ALT_Q)) { 199 const int data = get_segdata(seg, segment_id, SEG_LVL_ALT_Q);
|
vp9_blockd.h | 72 int8_t segment_id; member in struct:MODE_INFO
|
/external/libvpx/libvpx/vp9/encoder/ |
vp9_segmentation.h | 24 void vp9_disable_segfeature(struct segmentation *seg, int segment_id, 26 void vp9_clear_segdata(struct segmentation *seg, int segment_id,
|
vp9_segmentation.c | 39 void vp9_disable_segfeature(struct segmentation *seg, int segment_id, 41 seg->feature_mask[segment_id] &= ~(1 << feature_id); 44 void vp9_clear_segdata(struct segmentation *seg, int segment_id, 46 seg->feature_data[segment_id][feature_id] = 0; 110 int segment_id; local 115 segment_id = xd->mi[0]->segment_id; 120 no_pred_segcounts[segment_id]++; 128 const int pred_flag = pred_segment_id == segment_id; 137 if (!pred_flag) t_unpred_seg_counts[segment_id]++ [all...] |
vp9_aq_cyclicrefresh.h | 92 // check if we should reset the segment_id, and update the cyclic_refresh map 128 static INLINE int cyclic_refresh_segment_id_boosted(int segment_id) { 129 return segment_id == CR_SEGMENT_ID_BOOST1 || 130 segment_id == CR_SEGMENT_ID_BOOST2; 133 static INLINE int cyclic_refresh_segment_id(int segment_id) { 134 if (segment_id == CR_SEGMENT_ID_BOOST1) 136 else if (segment_id == CR_SEGMENT_ID_BOOST2)
|
vp9_aq_variance.c | 32 static const int segment_id[ENERGY_SPAN] = { 0, 1, 1, 2, 3, 4 }; variable 34 #define SEGMENT_ID(i) segment_id[(i)-ENERGY_MIN] 43 return SEGMENT_ID(energy);
|
vp9_aq_cyclicrefresh.c | 146 // check if we should reset the segment_id, and update the cyclic_refresh map 178 // segment_id. 179 if (cyclic_refresh_segment_id_boosted(mi->segment_id)) { 180 mi->segment_id = refresh_this_block; 181 // Reset segment_id if it will be skipped. 182 if (skip) mi->segment_id = CR_SEGMENT_ID_BASE; 189 if (cyclic_refresh_segment_id_boosted(mi->segment_id)) { 202 // copy mbmi->segment_id into global segmentation map. 207 cpi->segmentation_map[map_offset] = mi->segment_id; 231 mi->segment_id <= CR_SEGMENT_ID_BOOST2) [all...] |
vp9_bitstream.c | 99 int segment_id, const MODE_INFO *mi, vpx_writer *w) { 100 if (segfeature_active(&cm->seg, segment_id, SEG_LVL_SKIP)) { 192 int segment_id) { 194 vp9_write_tree(w, vp9_segment_tree, seg->tree_probs, segment_id, 3, 0); 202 const int segment_id = mi->segment_id; local 206 if (segfeature_active(&cm->seg, segment_id, SEG_LVL_REF_FRAME)) { 209 get_segdata(&cm->seg, segment_id, SEG_LVL_REF_FRAME)); 243 const int segment_id = mi->segment_id; local [all...] |
vp9_rd.c | 227 int i, bsize, segment_id; local 229 for (segment_id = 0; segment_id < MAX_SEGMENTS; ++segment_id) { 231 clamp(vp9_get_qindex(&cm->seg, segment_id, cm->base_qindex) + 244 rd->threshes[segment_id][bsize][i] = rd->thresh_mult[i] < thresh_max 249 rd->threshes[segment_id][bsize][i] =
|
vp9_encodeframe.c | 265 mi->segment_id = get_segment_id(cm, map, bsize, mi_row, mi_col); 269 x->encode_breakout = cpi->segment_encode_breakout[mi->segment_id]; 271 mi->segment_id = 0; 1255 int segment_id; local 3591 int segment_id = get_segment_id(cm, map, BLOCK_64X64, mi_row, mi_col); local 4380 int segment_id = get_segment_id(cm, map, BLOCK_64X64, mi_row, mi_col); local 4868 int segment_id; local [all...] |
vp9_quantize.c | 277 const int segment_id = xd->mi[0]->segment_id; local 278 const int qindex = vp9_get_qindex(&cm->seg, segment_id, cm->base_qindex); 308 x->skip_block = segfeature_active(&cm->seg, segment_id, SEG_LVL_SKIP);
|
vp9_rdopt.c | 3040 unsigned char segment_id = mi->segment_id; local 3709 unsigned char segment_id = mi->segment_id; local 3809 unsigned char segment_id = mi->segment_id; local [all...] |
vp9_pickmode.c | 397 cyclic_refresh_segment_id_boosted(xd->mi[0]->segment_id)) 575 cyclic_refresh_segment_id_boosted(xd->mi[0]->segment_id)) 2436 unsigned char segment_id = mi->segment_id; local [all...] |
/external/libvpx/libvpx/vp9/decoder/ |
vp9_decodemv.c | 93 int x, y, segment_id = INT_MAX; local 97 segment_id = 98 VPXMIN(segment_id, segment_ids[mi_offset + y * cm->mi_cols + x]); 100 assert(segment_id >= 0 && segment_id < MAX_SEGMENTS); 101 return segment_id; 105 int segment_id) { 108 assert(segment_id >= 0 && segment_id < MAX_SEGMENTS); 112 cm->current_frame_seg_map[mi_offset + y * cm->mi_cols + x] = segment_id; 131 int segment_id; local 151 int predicted_segment_id, segment_id; local [all...] |
/external/libvpx/libvpx/vp8/common/ |
vp8_loopfilter.c | 183 const int seg = mode_info_context->mbmi.segment_id; 236 const int seg = mode_info_context->mbmi.segment_id; 303 const int seg = mode_info_context->mbmi.segment_id; 353 const int seg = mode_info_context->mbmi.segment_id; 424 const int seg = mode_info_context->mbmi.segment_id; 522 const int seg = mode_info_context->mbmi.segment_id;
|
blockd.h | 150 uint8_t segment_id; member in struct:__anon26109
|
/external/libvpx/libvpx/vp8/encoder/ |
ethreading.c | 157 /* Code to set segment id in xd->mbmi.segment_id for 161 xd->mode_info_context->mbmi.segment_id = 164 xd->mode_info_context->mbmi.segment_id = 0; 170 xd->mode_info_context->mbmi.segment_id = 0; 220 * xd->mbmi.segment_id; (which may have been updated 229 cpi->segmentation_map[map_index + mb_col] = mbmi->segment_id; 239 if (mbmi->segment_id) { 276 segment_counts[xd->mode_info_context->mbmi.segment_id]++;
|
encodeframe.c | 434 /* Code to set segment id in xd->mbmi.segment_id for current MB 438 xd->mode_info_context->mbmi.segment_id = 441 xd->mode_info_context->mbmi.segment_id = 0; 447 xd->mode_info_context->mbmi.segment_id = 0; 497 * If cyclic update enabled then copy xd->mbmi.segment_id; (which 505 xd->mode_info_context->mbmi.segment_id; 514 if (xd->mode_info_context->mbmi.segment_id) { 551 segment_counts[xd->mode_info_context->mbmi.segment_id]++; [all...] |
/external/libvpx/libvpx/vp8/decoder/ |
onyxd_if.c | 374 pbi->common.mi[i].mbmi.segment_id = 375 pbi->common.prev_mi[i].mbmi.segment_id;
|
decodemv.c | 479 mi->segment_id = 482 mi->segment_id = 499 mi->mbmi.segment_id = 0;
|
/hardware/intel/common/libva/va/ |
va_dec_vp8.h | 121 * probabilities of the segment_id decoding tree and same as
|