Home | History | Annotate | Download | only in encoder
      1 /*
      2  *  Copyright (c) 2014 The WebM project authors. All Rights Reserved.
      3  *
      4  *  Use of this source code is governed by a BSD-style license
      5  *  that can be found in the LICENSE file in the root of the source
      6  *  tree. An additional intellectual property rights grant can be found
      7  *  in the file PATENTS.  All contributing project authors may
      8  *  be found in the AUTHORS file in the root of the source tree.
      9  */
     10 
     11 #include "vp9/encoder/vp9_encodeframe.h"
     12 #include "vp9/encoder/vp9_encoder.h"
     13 #include "vp9/encoder/vp9_ethread.h"
     14 #include "vp9/encoder/vp9_firstpass.h"
     15 #include "vp9/encoder/vp9_multi_thread.h"
     16 #include "vp9/encoder/vp9_temporal_filter.h"
     17 #include "vpx_dsp/vpx_dsp_common.h"
     18 
     19 static void accumulate_rd_opt(ThreadData *td, ThreadData *td_t) {
     20   int i, j, k, l, m, n;
     21 
     22   for (i = 0; i < REFERENCE_MODES; i++)
     23     td->rd_counts.comp_pred_diff[i] += td_t->rd_counts.comp_pred_diff[i];
     24 
     25   for (i = 0; i < SWITCHABLE_FILTER_CONTEXTS; i++)
     26     td->rd_counts.filter_diff[i] += td_t->rd_counts.filter_diff[i];
     27 
     28   for (i = 0; i < TX_SIZES; i++)
     29     for (j = 0; j < PLANE_TYPES; j++)
     30       for (k = 0; k < REF_TYPES; k++)
     31         for (l = 0; l < COEF_BANDS; l++)
     32           for (m = 0; m < COEFF_CONTEXTS; m++)
     33             for (n = 0; n < ENTROPY_TOKENS; n++)
     34               td->rd_counts.coef_counts[i][j][k][l][m][n] +=
     35                   td_t->rd_counts.coef_counts[i][j][k][l][m][n];
     36 }
     37 
     38 static int enc_worker_hook(EncWorkerData *const thread_data, void *unused) {
     39   VP9_COMP *const cpi = thread_data->cpi;
     40   const VP9_COMMON *const cm = &cpi->common;
     41   const int tile_cols = 1 << cm->log2_tile_cols;
     42   const int tile_rows = 1 << cm->log2_tile_rows;
     43   int t;
     44 
     45   (void)unused;
     46 
     47   for (t = thread_data->start; t < tile_rows * tile_cols;
     48        t += cpi->num_workers) {
     49     int tile_row = t / tile_cols;
     50     int tile_col = t % tile_cols;
     51 
     52     vp9_encode_tile(cpi, thread_data->td, tile_row, tile_col);
     53   }
     54 
     55   return 0;
     56 }
     57 
     58 static int get_max_tile_cols(VP9_COMP *cpi) {
     59   const int aligned_width = ALIGN_POWER_OF_TWO(cpi->oxcf.width, MI_SIZE_LOG2);
     60   int mi_cols = aligned_width >> MI_SIZE_LOG2;
     61   int min_log2_tile_cols, max_log2_tile_cols;
     62   int log2_tile_cols;
     63 
     64   vp9_get_tile_n_bits(mi_cols, &min_log2_tile_cols, &max_log2_tile_cols);
     65   log2_tile_cols =
     66       clamp(cpi->oxcf.tile_columns, min_log2_tile_cols, max_log2_tile_cols);
     67   return (1 << log2_tile_cols);
     68 }
     69 
     70 static void create_enc_workers(VP9_COMP *cpi, int num_workers) {
     71   VP9_COMMON *const cm = &cpi->common;
     72   const VPxWorkerInterface *const winterface = vpx_get_worker_interface();
     73   int i;
     74 
     75   // Only run once to create threads and allocate thread data.
     76   if (cpi->num_workers == 0) {
     77     int allocated_workers = num_workers;
     78 
     79     // While using SVC, we need to allocate threads according to the highest
     80     // resolution. When row based multithreading is enabled, it is OK to
     81     // allocate more threads than the number of max tile columns.
     82     if (cpi->use_svc && !cpi->row_mt) {
     83       int max_tile_cols = get_max_tile_cols(cpi);
     84       allocated_workers = VPXMIN(cpi->oxcf.max_threads, max_tile_cols);
     85     }
     86 
     87     CHECK_MEM_ERROR(cm, cpi->workers,
     88                     vpx_malloc(allocated_workers * sizeof(*cpi->workers)));
     89 
     90     CHECK_MEM_ERROR(cm, cpi->tile_thr_data,
     91                     vpx_calloc(allocated_workers, sizeof(*cpi->tile_thr_data)));
     92 
     93     for (i = 0; i < allocated_workers; i++) {
     94       VPxWorker *const worker = &cpi->workers[i];
     95       EncWorkerData *thread_data = &cpi->tile_thr_data[i];
     96 
     97       ++cpi->num_workers;
     98       winterface->init(worker);
     99 
    100       if (i < allocated_workers - 1) {
    101         thread_data->cpi = cpi;
    102 
    103         // Allocate thread data.
    104         CHECK_MEM_ERROR(cm, thread_data->td,
    105                         vpx_memalign(32, sizeof(*thread_data->td)));
    106         vp9_zero(*thread_data->td);
    107 
    108         // Set up pc_tree.
    109         thread_data->td->leaf_tree = NULL;
    110         thread_data->td->pc_tree = NULL;
    111         vp9_setup_pc_tree(cm, thread_data->td);
    112 
    113         // Allocate frame counters in thread data.
    114         CHECK_MEM_ERROR(cm, thread_data->td->counts,
    115                         vpx_calloc(1, sizeof(*thread_data->td->counts)));
    116 
    117         // Create threads
    118         if (!winterface->reset(worker))
    119           vpx_internal_error(&cm->error, VPX_CODEC_ERROR,
    120                              "Tile encoder thread creation failed");
    121       } else {
    122         // Main thread acts as a worker and uses the thread data in cpi.
    123         thread_data->cpi = cpi;
    124         thread_data->td = &cpi->td;
    125       }
    126       winterface->sync(worker);
    127     }
    128   }
    129 }
    130 
    131 static void launch_enc_workers(VP9_COMP *cpi, VPxWorkerHook hook, void *data2,
    132                                int num_workers) {
    133   const VPxWorkerInterface *const winterface = vpx_get_worker_interface();
    134   int i;
    135 
    136   for (i = 0; i < num_workers; i++) {
    137     VPxWorker *const worker = &cpi->workers[i];
    138     worker->hook = (VPxWorkerHook)hook;
    139     worker->data1 = &cpi->tile_thr_data[i];
    140     worker->data2 = data2;
    141   }
    142 
    143   // Encode a frame
    144   for (i = 0; i < num_workers; i++) {
    145     VPxWorker *const worker = &cpi->workers[i];
    146     EncWorkerData *const thread_data = (EncWorkerData *)worker->data1;
    147 
    148     // Set the starting tile for each thread.
    149     thread_data->start = i;
    150 
    151     if (i == cpi->num_workers - 1)
    152       winterface->execute(worker);
    153     else
    154       winterface->launch(worker);
    155   }
    156 
    157   // Encoding ends.
    158   for (i = 0; i < num_workers; i++) {
    159     VPxWorker *const worker = &cpi->workers[i];
    160     winterface->sync(worker);
    161   }
    162 }
    163 
    164 void vp9_encode_tiles_mt(VP9_COMP *cpi) {
    165   VP9_COMMON *const cm = &cpi->common;
    166   const int tile_cols = 1 << cm->log2_tile_cols;
    167   const int num_workers = VPXMIN(cpi->oxcf.max_threads, tile_cols);
    168   int i;
    169 
    170   vp9_init_tile_data(cpi);
    171 
    172   create_enc_workers(cpi, num_workers);
    173 
    174   for (i = 0; i < num_workers; i++) {
    175     EncWorkerData *thread_data;
    176     thread_data = &cpi->tile_thr_data[i];
    177 
    178     // Before encoding a frame, copy the thread data from cpi.
    179     if (thread_data->td != &cpi->td) {
    180       thread_data->td->mb = cpi->td.mb;
    181       thread_data->td->rd_counts = cpi->td.rd_counts;
    182     }
    183     if (thread_data->td->counts != &cpi->common.counts) {
    184       memcpy(thread_data->td->counts, &cpi->common.counts,
    185              sizeof(cpi->common.counts));
    186     }
    187 
    188     // Handle use_nonrd_pick_mode case.
    189     if (cpi->sf.use_nonrd_pick_mode) {
    190       MACROBLOCK *const x = &thread_data->td->mb;
    191       MACROBLOCKD *const xd = &x->e_mbd;
    192       struct macroblock_plane *const p = x->plane;
    193       struct macroblockd_plane *const pd = xd->plane;
    194       PICK_MODE_CONTEXT *ctx = &thread_data->td->pc_root->none;
    195       int j;
    196 
    197       for (j = 0; j < MAX_MB_PLANE; ++j) {
    198         p[j].coeff = ctx->coeff_pbuf[j][0];
    199         p[j].qcoeff = ctx->qcoeff_pbuf[j][0];
    200         pd[j].dqcoeff = ctx->dqcoeff_pbuf[j][0];
    201         p[j].eobs = ctx->eobs_pbuf[j][0];
    202       }
    203     }
    204   }
    205 
    206   launch_enc_workers(cpi, (VPxWorkerHook)enc_worker_hook, NULL, num_workers);
    207 
    208   for (i = 0; i < num_workers; i++) {
    209     VPxWorker *const worker = &cpi->workers[i];
    210     EncWorkerData *const thread_data = (EncWorkerData *)worker->data1;
    211 
    212     // Accumulate counters.
    213     if (i < cpi->num_workers - 1) {
    214       vp9_accumulate_frame_counts(&cm->counts, thread_data->td->counts, 0);
    215       accumulate_rd_opt(&cpi->td, thread_data->td);
    216     }
    217   }
    218 }
    219 
    220 static void accumulate_fp_tile_stat(TileDataEnc *tile_data,
    221                                     TileDataEnc *tile_data_t) {
    222   tile_data->fp_data.intra_factor += tile_data_t->fp_data.intra_factor;
    223   tile_data->fp_data.brightness_factor +=
    224       tile_data_t->fp_data.brightness_factor;
    225   tile_data->fp_data.coded_error += tile_data_t->fp_data.coded_error;
    226   tile_data->fp_data.sr_coded_error += tile_data_t->fp_data.sr_coded_error;
    227   tile_data->fp_data.frame_noise_energy +=
    228       tile_data_t->fp_data.frame_noise_energy;
    229   tile_data->fp_data.intra_error += tile_data_t->fp_data.intra_error;
    230   tile_data->fp_data.intercount += tile_data_t->fp_data.intercount;
    231   tile_data->fp_data.second_ref_count += tile_data_t->fp_data.second_ref_count;
    232   tile_data->fp_data.neutral_count += tile_data_t->fp_data.neutral_count;
    233   tile_data->fp_data.intra_count_low += tile_data_t->fp_data.intra_count_low;
    234   tile_data->fp_data.intra_count_high += tile_data_t->fp_data.intra_count_high;
    235   tile_data->fp_data.intra_skip_count += tile_data_t->fp_data.intra_skip_count;
    236   tile_data->fp_data.mvcount += tile_data_t->fp_data.mvcount;
    237   tile_data->fp_data.sum_mvr += tile_data_t->fp_data.sum_mvr;
    238   tile_data->fp_data.sum_mvr_abs += tile_data_t->fp_data.sum_mvr_abs;
    239   tile_data->fp_data.sum_mvc += tile_data_t->fp_data.sum_mvc;
    240   tile_data->fp_data.sum_mvc_abs += tile_data_t->fp_data.sum_mvc_abs;
    241   tile_data->fp_data.sum_mvrs += tile_data_t->fp_data.sum_mvrs;
    242   tile_data->fp_data.sum_mvcs += tile_data_t->fp_data.sum_mvcs;
    243   tile_data->fp_data.sum_in_vectors += tile_data_t->fp_data.sum_in_vectors;
    244   tile_data->fp_data.intra_smooth_count +=
    245       tile_data_t->fp_data.intra_smooth_count;
    246   tile_data->fp_data.image_data_start_row =
    247       VPXMIN(tile_data->fp_data.image_data_start_row,
    248              tile_data_t->fp_data.image_data_start_row) == INVALID_ROW
    249           ? VPXMAX(tile_data->fp_data.image_data_start_row,
    250                    tile_data_t->fp_data.image_data_start_row)
    251           : VPXMIN(tile_data->fp_data.image_data_start_row,
    252                    tile_data_t->fp_data.image_data_start_row);
    253 }
    254 
    255 // Allocate memory for row synchronization
    256 void vp9_row_mt_sync_mem_alloc(VP9RowMTSync *row_mt_sync, VP9_COMMON *cm,
    257                                int rows) {
    258   row_mt_sync->rows = rows;
    259 #if CONFIG_MULTITHREAD
    260   {
    261     int i;
    262 
    263     CHECK_MEM_ERROR(cm, row_mt_sync->mutex_,
    264                     vpx_malloc(sizeof(*row_mt_sync->mutex_) * rows));
    265     if (row_mt_sync->mutex_) {
    266       for (i = 0; i < rows; ++i) {
    267         pthread_mutex_init(&row_mt_sync->mutex_[i], NULL);
    268       }
    269     }
    270 
    271     CHECK_MEM_ERROR(cm, row_mt_sync->cond_,
    272                     vpx_malloc(sizeof(*row_mt_sync->cond_) * rows));
    273     if (row_mt_sync->cond_) {
    274       for (i = 0; i < rows; ++i) {
    275         pthread_cond_init(&row_mt_sync->cond_[i], NULL);
    276       }
    277     }
    278   }
    279 #endif  // CONFIG_MULTITHREAD
    280 
    281   CHECK_MEM_ERROR(cm, row_mt_sync->cur_col,
    282                   vpx_malloc(sizeof(*row_mt_sync->cur_col) * rows));
    283 
    284   // Set up nsync.
    285   row_mt_sync->sync_range = 1;
    286 }
    287 
    288 // Deallocate row based multi-threading synchronization related mutex and data
    289 void vp9_row_mt_sync_mem_dealloc(VP9RowMTSync *row_mt_sync) {
    290   if (row_mt_sync != NULL) {
    291 #if CONFIG_MULTITHREAD
    292     int i;
    293 
    294     if (row_mt_sync->mutex_ != NULL) {
    295       for (i = 0; i < row_mt_sync->rows; ++i) {
    296         pthread_mutex_destroy(&row_mt_sync->mutex_[i]);
    297       }
    298       vpx_free(row_mt_sync->mutex_);
    299     }
    300     if (row_mt_sync->cond_ != NULL) {
    301       for (i = 0; i < row_mt_sync->rows; ++i) {
    302         pthread_cond_destroy(&row_mt_sync->cond_[i]);
    303       }
    304       vpx_free(row_mt_sync->cond_);
    305     }
    306 #endif  // CONFIG_MULTITHREAD
    307     vpx_free(row_mt_sync->cur_col);
    308     // clear the structure as the source of this call may be dynamic change
    309     // in tiles in which case this call will be followed by an _alloc()
    310     // which may fail.
    311     vp9_zero(*row_mt_sync);
    312   }
    313 }
    314 
    315 void vp9_row_mt_sync_read(VP9RowMTSync *const row_mt_sync, int r, int c) {
    316 #if CONFIG_MULTITHREAD
    317   const int nsync = row_mt_sync->sync_range;
    318 
    319   if (r && !(c & (nsync - 1))) {
    320     pthread_mutex_t *const mutex = &row_mt_sync->mutex_[r - 1];
    321     pthread_mutex_lock(mutex);
    322 
    323     while (c > row_mt_sync->cur_col[r - 1] - nsync + 1) {
    324       pthread_cond_wait(&row_mt_sync->cond_[r - 1], mutex);
    325     }
    326     pthread_mutex_unlock(mutex);
    327   }
    328 #else
    329   (void)row_mt_sync;
    330   (void)r;
    331   (void)c;
    332 #endif  // CONFIG_MULTITHREAD
    333 }
    334 
    335 void vp9_row_mt_sync_read_dummy(VP9RowMTSync *const row_mt_sync, int r, int c) {
    336   (void)row_mt_sync;
    337   (void)r;
    338   (void)c;
    339   return;
    340 }
    341 
    342 void vp9_row_mt_sync_write(VP9RowMTSync *const row_mt_sync, int r, int c,
    343                            const int cols) {
    344 #if CONFIG_MULTITHREAD
    345   const int nsync = row_mt_sync->sync_range;
    346   int cur;
    347   // Only signal when there are enough encoded blocks for next row to run.
    348   int sig = 1;
    349 
    350   if (c < cols - 1) {
    351     cur = c;
    352     if (c % nsync != nsync - 1) sig = 0;
    353   } else {
    354     cur = cols + nsync;
    355   }
    356 
    357   if (sig) {
    358     pthread_mutex_lock(&row_mt_sync->mutex_[r]);
    359 
    360     row_mt_sync->cur_col[r] = cur;
    361 
    362     pthread_cond_signal(&row_mt_sync->cond_[r]);
    363     pthread_mutex_unlock(&row_mt_sync->mutex_[r]);
    364   }
    365 #else
    366   (void)row_mt_sync;
    367   (void)r;
    368   (void)c;
    369   (void)cols;
    370 #endif  // CONFIG_MULTITHREAD
    371 }
    372 
    373 void vp9_row_mt_sync_write_dummy(VP9RowMTSync *const row_mt_sync, int r, int c,
    374                                  const int cols) {
    375   (void)row_mt_sync;
    376   (void)r;
    377   (void)c;
    378   (void)cols;
    379   return;
    380 }
    381 
    382 static int first_pass_worker_hook(EncWorkerData *const thread_data,
    383                                   MultiThreadHandle *multi_thread_ctxt) {
    384   VP9_COMP *const cpi = thread_data->cpi;
    385   const VP9_COMMON *const cm = &cpi->common;
    386   const int tile_cols = 1 << cm->log2_tile_cols;
    387   int tile_row, tile_col;
    388   TileDataEnc *this_tile;
    389   int end_of_frame;
    390   int thread_id = thread_data->thread_id;
    391   int cur_tile_id = multi_thread_ctxt->thread_id_to_tile_id[thread_id];
    392   JobNode *proc_job = NULL;
    393   FIRSTPASS_DATA fp_acc_data;
    394   MV zero_mv = { 0, 0 };
    395   MV best_ref_mv;
    396   int mb_row;
    397 
    398   end_of_frame = 0;
    399   while (0 == end_of_frame) {
    400     // Get the next job in the queue
    401     proc_job =
    402         (JobNode *)vp9_enc_grp_get_next_job(multi_thread_ctxt, cur_tile_id);
    403     if (NULL == proc_job) {
    404       // Query for the status of other tiles
    405       end_of_frame = vp9_get_tiles_proc_status(
    406           multi_thread_ctxt, thread_data->tile_completion_status, &cur_tile_id,
    407           tile_cols);
    408     } else {
    409       tile_col = proc_job->tile_col_id;
    410       tile_row = proc_job->tile_row_id;
    411 
    412       this_tile = &cpi->tile_data[tile_row * tile_cols + tile_col];
    413       mb_row = proc_job->vert_unit_row_num;
    414 
    415       best_ref_mv = zero_mv;
    416       vp9_zero(fp_acc_data);
    417       fp_acc_data.image_data_start_row = INVALID_ROW;
    418       vp9_first_pass_encode_tile_mb_row(cpi, thread_data->td, &fp_acc_data,
    419                                         this_tile, &best_ref_mv, mb_row);
    420     }
    421   }
    422   return 0;
    423 }
    424 
    425 void vp9_encode_fp_row_mt(VP9_COMP *cpi) {
    426   VP9_COMMON *const cm = &cpi->common;
    427   const int tile_cols = 1 << cm->log2_tile_cols;
    428   const int tile_rows = 1 << cm->log2_tile_rows;
    429   MultiThreadHandle *multi_thread_ctxt = &cpi->multi_thread_ctxt;
    430   TileDataEnc *first_tile_col;
    431   int num_workers = VPXMAX(cpi->oxcf.max_threads, 1);
    432   int i;
    433 
    434   if (multi_thread_ctxt->allocated_tile_cols < tile_cols ||
    435       multi_thread_ctxt->allocated_tile_rows < tile_rows ||
    436       multi_thread_ctxt->allocated_vert_unit_rows < cm->mb_rows) {
    437     vp9_row_mt_mem_dealloc(cpi);
    438     vp9_init_tile_data(cpi);
    439     vp9_row_mt_mem_alloc(cpi);
    440   } else {
    441     vp9_init_tile_data(cpi);
    442   }
    443 
    444   create_enc_workers(cpi, num_workers);
    445 
    446   vp9_assign_tile_to_thread(multi_thread_ctxt, tile_cols, cpi->num_workers);
    447 
    448   vp9_prepare_job_queue(cpi, FIRST_PASS_JOB);
    449 
    450   vp9_multi_thread_tile_init(cpi);
    451 
    452   for (i = 0; i < num_workers; i++) {
    453     EncWorkerData *thread_data;
    454     thread_data = &cpi->tile_thr_data[i];
    455 
    456     // Before encoding a frame, copy the thread data from cpi.
    457     if (thread_data->td != &cpi->td) {
    458       thread_data->td->mb = cpi->td.mb;
    459     }
    460   }
    461 
    462   launch_enc_workers(cpi, (VPxWorkerHook)first_pass_worker_hook,
    463                      multi_thread_ctxt, num_workers);
    464 
    465   first_tile_col = &cpi->tile_data[0];
    466   for (i = 1; i < tile_cols; i++) {
    467     TileDataEnc *this_tile = &cpi->tile_data[i];
    468     accumulate_fp_tile_stat(first_tile_col, this_tile);
    469   }
    470 }
    471 
    472 static int temporal_filter_worker_hook(EncWorkerData *const thread_data,
    473                                        MultiThreadHandle *multi_thread_ctxt) {
    474   VP9_COMP *const cpi = thread_data->cpi;
    475   const VP9_COMMON *const cm = &cpi->common;
    476   const int tile_cols = 1 << cm->log2_tile_cols;
    477   int tile_row, tile_col;
    478   int mb_col_start, mb_col_end;
    479   TileDataEnc *this_tile;
    480   int end_of_frame;
    481   int thread_id = thread_data->thread_id;
    482   int cur_tile_id = multi_thread_ctxt->thread_id_to_tile_id[thread_id];
    483   JobNode *proc_job = NULL;
    484   int mb_row;
    485 
    486   end_of_frame = 0;
    487   while (0 == end_of_frame) {
    488     // Get the next job in the queue
    489     proc_job =
    490         (JobNode *)vp9_enc_grp_get_next_job(multi_thread_ctxt, cur_tile_id);
    491     if (NULL == proc_job) {
    492       // Query for the status of other tiles
    493       end_of_frame = vp9_get_tiles_proc_status(
    494           multi_thread_ctxt, thread_data->tile_completion_status, &cur_tile_id,
    495           tile_cols);
    496     } else {
    497       tile_col = proc_job->tile_col_id;
    498       tile_row = proc_job->tile_row_id;
    499       this_tile = &cpi->tile_data[tile_row * tile_cols + tile_col];
    500       mb_col_start = (this_tile->tile_info.mi_col_start) >> 1;
    501       mb_col_end = (this_tile->tile_info.mi_col_end + 1) >> 1;
    502       mb_row = proc_job->vert_unit_row_num;
    503 
    504       vp9_temporal_filter_iterate_row_c(cpi, thread_data->td, mb_row,
    505                                         mb_col_start, mb_col_end);
    506     }
    507   }
    508   return 0;
    509 }
    510 
    511 void vp9_temporal_filter_row_mt(VP9_COMP *cpi) {
    512   VP9_COMMON *const cm = &cpi->common;
    513   const int tile_cols = 1 << cm->log2_tile_cols;
    514   const int tile_rows = 1 << cm->log2_tile_rows;
    515   MultiThreadHandle *multi_thread_ctxt = &cpi->multi_thread_ctxt;
    516   int num_workers = cpi->num_workers ? cpi->num_workers : 1;
    517   int i;
    518 
    519   if (multi_thread_ctxt->allocated_tile_cols < tile_cols ||
    520       multi_thread_ctxt->allocated_tile_rows < tile_rows ||
    521       multi_thread_ctxt->allocated_vert_unit_rows < cm->mb_rows) {
    522     vp9_row_mt_mem_dealloc(cpi);
    523     vp9_init_tile_data(cpi);
    524     vp9_row_mt_mem_alloc(cpi);
    525   } else {
    526     vp9_init_tile_data(cpi);
    527   }
    528 
    529   create_enc_workers(cpi, num_workers);
    530 
    531   vp9_assign_tile_to_thread(multi_thread_ctxt, tile_cols, cpi->num_workers);
    532 
    533   vp9_prepare_job_queue(cpi, ARNR_JOB);
    534 
    535   for (i = 0; i < num_workers; i++) {
    536     EncWorkerData *thread_data;
    537     thread_data = &cpi->tile_thr_data[i];
    538 
    539     // Before encoding a frame, copy the thread data from cpi.
    540     if (thread_data->td != &cpi->td) {
    541       thread_data->td->mb = cpi->td.mb;
    542     }
    543   }
    544 
    545   launch_enc_workers(cpi, (VPxWorkerHook)temporal_filter_worker_hook,
    546                      multi_thread_ctxt, num_workers);
    547 }
    548 
    549 static int enc_row_mt_worker_hook(EncWorkerData *const thread_data,
    550                                   MultiThreadHandle *multi_thread_ctxt) {
    551   VP9_COMP *const cpi = thread_data->cpi;
    552   const VP9_COMMON *const cm = &cpi->common;
    553   const int tile_cols = 1 << cm->log2_tile_cols;
    554   int tile_row, tile_col;
    555   int end_of_frame;
    556   int thread_id = thread_data->thread_id;
    557   int cur_tile_id = multi_thread_ctxt->thread_id_to_tile_id[thread_id];
    558   JobNode *proc_job = NULL;
    559   int mi_row;
    560 
    561   end_of_frame = 0;
    562   while (0 == end_of_frame) {
    563     // Get the next job in the queue
    564     proc_job =
    565         (JobNode *)vp9_enc_grp_get_next_job(multi_thread_ctxt, cur_tile_id);
    566     if (NULL == proc_job) {
    567       // Query for the status of other tiles
    568       end_of_frame = vp9_get_tiles_proc_status(
    569           multi_thread_ctxt, thread_data->tile_completion_status, &cur_tile_id,
    570           tile_cols);
    571     } else {
    572       tile_col = proc_job->tile_col_id;
    573       tile_row = proc_job->tile_row_id;
    574       mi_row = proc_job->vert_unit_row_num * MI_BLOCK_SIZE;
    575 
    576       vp9_encode_sb_row(cpi, thread_data->td, tile_row, tile_col, mi_row);
    577     }
    578   }
    579   return 0;
    580 }
    581 
    582 void vp9_encode_tiles_row_mt(VP9_COMP *cpi) {
    583   VP9_COMMON *const cm = &cpi->common;
    584   const int tile_cols = 1 << cm->log2_tile_cols;
    585   const int tile_rows = 1 << cm->log2_tile_rows;
    586   MultiThreadHandle *multi_thread_ctxt = &cpi->multi_thread_ctxt;
    587   int num_workers = VPXMAX(cpi->oxcf.max_threads, 1);
    588   int i;
    589 
    590   if (multi_thread_ctxt->allocated_tile_cols < tile_cols ||
    591       multi_thread_ctxt->allocated_tile_rows < tile_rows ||
    592       multi_thread_ctxt->allocated_vert_unit_rows < cm->mb_rows) {
    593     vp9_row_mt_mem_dealloc(cpi);
    594     vp9_init_tile_data(cpi);
    595     vp9_row_mt_mem_alloc(cpi);
    596   } else {
    597     vp9_init_tile_data(cpi);
    598   }
    599 
    600   create_enc_workers(cpi, num_workers);
    601 
    602   vp9_assign_tile_to_thread(multi_thread_ctxt, tile_cols, cpi->num_workers);
    603 
    604   vp9_prepare_job_queue(cpi, ENCODE_JOB);
    605 
    606   vp9_multi_thread_tile_init(cpi);
    607 
    608   for (i = 0; i < num_workers; i++) {
    609     EncWorkerData *thread_data;
    610     thread_data = &cpi->tile_thr_data[i];
    611     // Before encoding a frame, copy the thread data from cpi.
    612     if (thread_data->td != &cpi->td) {
    613       thread_data->td->mb = cpi->td.mb;
    614       thread_data->td->rd_counts = cpi->td.rd_counts;
    615     }
    616     if (thread_data->td->counts != &cpi->common.counts) {
    617       memcpy(thread_data->td->counts, &cpi->common.counts,
    618              sizeof(cpi->common.counts));
    619     }
    620 
    621     // Handle use_nonrd_pick_mode case.
    622     if (cpi->sf.use_nonrd_pick_mode) {
    623       MACROBLOCK *const x = &thread_data->td->mb;
    624       MACROBLOCKD *const xd = &x->e_mbd;
    625       struct macroblock_plane *const p = x->plane;
    626       struct macroblockd_plane *const pd = xd->plane;
    627       PICK_MODE_CONTEXT *ctx = &thread_data->td->pc_root->none;
    628       int j;
    629 
    630       for (j = 0; j < MAX_MB_PLANE; ++j) {
    631         p[j].coeff = ctx->coeff_pbuf[j][0];
    632         p[j].qcoeff = ctx->qcoeff_pbuf[j][0];
    633         pd[j].dqcoeff = ctx->dqcoeff_pbuf[j][0];
    634         p[j].eobs = ctx->eobs_pbuf[j][0];
    635       }
    636     }
    637   }
    638 
    639   launch_enc_workers(cpi, (VPxWorkerHook)enc_row_mt_worker_hook,
    640                      multi_thread_ctxt, num_workers);
    641 
    642   for (i = 0; i < num_workers; i++) {
    643     VPxWorker *const worker = &cpi->workers[i];
    644     EncWorkerData *const thread_data = (EncWorkerData *)worker->data1;
    645 
    646     // Accumulate counters.
    647     if (i < cpi->num_workers - 1) {
    648       vp9_accumulate_frame_counts(&cm->counts, thread_data->td->counts, 0);
    649       accumulate_rd_opt(&cpi->td, thread_data->td);
    650     }
    651   }
    652 }
    653