/src/libvpx/vp9/encoder/vp9_mbgraph.c
Line  | Count  | Source (jump to first uncovered line)  | 
1  |  | /*  | 
2  |  |  *  Copyright (c) 2010 The WebM project authors. All Rights Reserved.  | 
3  |  |  *  | 
4  |  |  *  Use of this source code is governed by a BSD-style license  | 
5  |  |  *  that can be found in the LICENSE file in the root of the source  | 
6  |  |  *  tree. An additional intellectual property rights grant can be found  | 
7  |  |  *  in the file PATENTS.  All contributing project authors may  | 
8  |  |  *  be found in the AUTHORS file in the root of the source tree.  | 
9  |  |  */  | 
10  |  |  | 
11  |  | #include <limits.h>  | 
12  |  |  | 
13  |  | #include "./vp9_rtcd.h"  | 
14  |  | #include "./vpx_dsp_rtcd.h"  | 
15  |  |  | 
16  |  | #include "vpx_dsp/vpx_dsp_common.h"  | 
17  |  | #include "vpx_mem/vpx_mem.h"  | 
18  |  | #include "vpx_ports/system_state.h"  | 
19  |  | #include "vp9/encoder/vp9_segmentation.h"  | 
20  |  | #include "vp9/encoder/vp9_mcomp.h"  | 
21  |  | #include "vp9/common/vp9_blockd.h"  | 
22  |  | #include "vp9/common/vp9_reconinter.h"  | 
23  |  | #include "vp9/common/vp9_reconintra.h"  | 
24  |  |  | 
25  |  | static unsigned int do_16x16_motion_iteration(VP9_COMP *cpi, const MV *ref_mv,  | 
26  |  |                                               MV *dst_mv, int mb_row,  | 
27  | 0  |                                               int mb_col) { | 
28  | 0  |   MACROBLOCK *const x = &cpi->td.mb;  | 
29  | 0  |   MACROBLOCKD *const xd = &x->e_mbd;  | 
30  | 0  |   MV_SPEED_FEATURES *const mv_sf = &cpi->sf.mv;  | 
31  | 0  |   const SEARCH_METHODS old_search_method = mv_sf->search_method;  | 
32  | 0  |   const vp9_variance_fn_ptr_t v_fn_ptr = cpi->fn_ptr[BLOCK_16X16];  | 
33  | 0  |   const MvLimits tmp_mv_limits = x->mv_limits;  | 
34  | 0  |   MV ref_full;  | 
35  | 0  |   int cost_list[5];  | 
36  |  |  | 
37  |  |   // Further step/diamond searches as necessary  | 
38  | 0  |   int step_param = mv_sf->reduce_first_step_size;  | 
39  | 0  |   step_param = VPXMIN(step_param, MAX_MVSEARCH_STEPS - 2);  | 
40  |  | 
  | 
41  | 0  |   vp9_set_mv_search_range(&x->mv_limits, ref_mv);  | 
42  |  | 
  | 
43  | 0  |   ref_full.col = ref_mv->col >> 3;  | 
44  | 0  |   ref_full.row = ref_mv->row >> 3;  | 
45  |  | 
  | 
46  | 0  |   mv_sf->search_method = HEX;  | 
47  | 0  |   vp9_full_pixel_search(cpi, x, BLOCK_16X16, &ref_full, step_param,  | 
48  | 0  |                         cpi->sf.mv.search_method, x->errorperbit,  | 
49  | 0  |                         cond_cost_list(cpi, cost_list), ref_mv, dst_mv, 0, 0);  | 
50  | 0  |   mv_sf->search_method = old_search_method;  | 
51  |  |  | 
52  |  |   /* restore UMV window */  | 
53  | 0  |   x->mv_limits = tmp_mv_limits;  | 
54  |  |  | 
55  |  |   // Try sub-pixel MC  | 
56  |  |   // if (bestsme > error_thresh && bestsme < INT_MAX)  | 
57  | 0  |   { | 
58  | 0  |     uint32_t distortion;  | 
59  | 0  |     uint32_t sse;  | 
60  |  |     // TODO(yunqing): may use higher tap interp filter than 2 taps if needed.  | 
61  | 0  |     cpi->find_fractional_mv_step(  | 
62  | 0  |         x, dst_mv, ref_mv, cpi->common.allow_high_precision_mv, x->errorperbit,  | 
63  | 0  |         &v_fn_ptr, 0, mv_sf->subpel_search_level,  | 
64  | 0  |         cond_cost_list(cpi, cost_list), NULL, NULL, &distortion, &sse, NULL, 0,  | 
65  | 0  |         0, USE_2_TAPS);  | 
66  | 0  |   }  | 
67  |  | 
  | 
68  | 0  |   xd->mi[0]->mode = NEWMV;  | 
69  | 0  |   xd->mi[0]->mv[0].as_mv = *dst_mv;  | 
70  |  | 
  | 
71  | 0  |   vp9_build_inter_predictors_sby(xd, mb_row, mb_col, BLOCK_16X16);  | 
72  |  | 
  | 
73  | 0  |   return vpx_sad16x16(x->plane[0].src.buf, x->plane[0].src.stride,  | 
74  | 0  |                       xd->plane[0].dst.buf, xd->plane[0].dst.stride);  | 
75  | 0  | }  | 
76  |  |  | 
77  |  | static int do_16x16_motion_search(VP9_COMP *cpi, const MV *ref_mv,  | 
78  | 0  |                                   int_mv *dst_mv, int mb_row, int mb_col) { | 
79  | 0  |   MACROBLOCK *const x = &cpi->td.mb;  | 
80  | 0  |   MACROBLOCKD *const xd = &x->e_mbd;  | 
81  | 0  |   unsigned int err, tmp_err;  | 
82  | 0  |   MV tmp_mv;  | 
83  |  |  | 
84  |  |   // Try zero MV first  | 
85  |  |   // FIXME should really use something like near/nearest MV and/or MV prediction  | 
86  | 0  |   err = vpx_sad16x16(x->plane[0].src.buf, x->plane[0].src.stride,  | 
87  | 0  |                      xd->plane[0].pre[0].buf, xd->plane[0].pre[0].stride);  | 
88  | 0  |   dst_mv->as_int = 0;  | 
89  |  |  | 
90  |  |   // Test last reference frame using the previous best mv as the  | 
91  |  |   // starting point (best reference) for the search  | 
92  | 0  |   tmp_err = do_16x16_motion_iteration(cpi, ref_mv, &tmp_mv, mb_row, mb_col);  | 
93  | 0  |   if (tmp_err < err) { | 
94  | 0  |     err = tmp_err;  | 
95  | 0  |     dst_mv->as_mv = tmp_mv;  | 
96  | 0  |   }  | 
97  |  |  | 
98  |  |   // If the current best reference mv is not centered on 0,0 then do a 0,0  | 
99  |  |   // based search as well.  | 
100  | 0  |   if (ref_mv->row != 0 || ref_mv->col != 0) { | 
101  | 0  |     MV zero_ref_mv = { 0, 0 }; | 
102  |  | 
  | 
103  | 0  |     tmp_err =  | 
104  | 0  |         do_16x16_motion_iteration(cpi, &zero_ref_mv, &tmp_mv, mb_row, mb_col);  | 
105  | 0  |     if (tmp_err < err) { | 
106  | 0  |       dst_mv->as_mv = tmp_mv;  | 
107  | 0  |       err = tmp_err;  | 
108  | 0  |     }  | 
109  | 0  |   }  | 
110  |  | 
  | 
111  | 0  |   return err;  | 
112  | 0  | }  | 
113  |  |  | 
114  | 0  | static int do_16x16_zerozero_search(VP9_COMP *cpi, int_mv *dst_mv) { | 
115  | 0  |   MACROBLOCK *const x = &cpi->td.mb;  | 
116  | 0  |   MACROBLOCKD *const xd = &x->e_mbd;  | 
117  | 0  |   unsigned int err;  | 
118  |  |  | 
119  |  |   // Try zero MV first  | 
120  |  |   // FIXME should really use something like near/nearest MV and/or MV prediction  | 
121  | 0  |   err = vpx_sad16x16(x->plane[0].src.buf, x->plane[0].src.stride,  | 
122  | 0  |                      xd->plane[0].pre[0].buf, xd->plane[0].pre[0].stride);  | 
123  |  | 
  | 
124  | 0  |   dst_mv->as_int = 0;  | 
125  |  | 
  | 
126  | 0  |   return err;  | 
127  | 0  | }  | 
128  | 0  | static int find_best_16x16_intra(VP9_COMP *cpi, PREDICTION_MODE *pbest_mode) { | 
129  | 0  |   MACROBLOCK *const x = &cpi->td.mb;  | 
130  | 0  |   MACROBLOCKD *const xd = &x->e_mbd;  | 
131  | 0  |   PREDICTION_MODE best_mode = -1, mode;  | 
132  | 0  |   unsigned int best_err = INT_MAX;  | 
133  |  |  | 
134  |  |   // calculate SATD for each intra prediction mode;  | 
135  |  |   // we're intentionally not doing 4x4, we just want a rough estimate  | 
136  | 0  |   for (mode = DC_PRED; mode <= TM_PRED; mode++) { | 
137  | 0  |     unsigned int err;  | 
138  |  | 
  | 
139  | 0  |     xd->mi[0]->mode = mode;  | 
140  | 0  |     vp9_predict_intra_block(xd, 2, TX_16X16, mode, x->plane[0].src.buf,  | 
141  | 0  |                             x->plane[0].src.stride, xd->plane[0].dst.buf,  | 
142  | 0  |                             xd->plane[0].dst.stride, 0, 0, 0);  | 
143  | 0  |     err = vpx_sad16x16(x->plane[0].src.buf, x->plane[0].src.stride,  | 
144  | 0  |                        xd->plane[0].dst.buf, xd->plane[0].dst.stride);  | 
145  |  |  | 
146  |  |     // find best  | 
147  | 0  |     if (err < best_err) { | 
148  | 0  |       best_err = err;  | 
149  | 0  |       best_mode = mode;  | 
150  | 0  |     }  | 
151  | 0  |   }  | 
152  |  | 
  | 
153  | 0  |   if (pbest_mode) *pbest_mode = best_mode;  | 
154  |  | 
  | 
155  | 0  |   return best_err;  | 
156  | 0  | }  | 
157  |  |  | 
158  |  | static void update_mbgraph_mb_stats(VP9_COMP *cpi, MBGRAPH_MB_STATS *stats,  | 
159  |  |                                     YV12_BUFFER_CONFIG *buf, int mb_y_offset,  | 
160  |  |                                     YV12_BUFFER_CONFIG *golden_ref,  | 
161  |  |                                     const MV *prev_golden_ref_mv,  | 
162  |  |                                     YV12_BUFFER_CONFIG *alt_ref, int mb_row,  | 
163  | 0  |                                     int mb_col) { | 
164  | 0  |   MACROBLOCK *const x = &cpi->td.mb;  | 
165  | 0  |   MACROBLOCKD *const xd = &x->e_mbd;  | 
166  | 0  |   int intra_error;  | 
167  | 0  |   VP9_COMMON *cm = &cpi->common;  | 
168  |  |  | 
169  |  |   // FIXME in practice we're completely ignoring chroma here  | 
170  | 0  |   x->plane[0].src.buf = buf->y_buffer + mb_y_offset;  | 
171  | 0  |   x->plane[0].src.stride = buf->y_stride;  | 
172  |  | 
  | 
173  | 0  |   xd->plane[0].dst.buf = get_frame_new_buffer(cm)->y_buffer + mb_y_offset;  | 
174  | 0  |   xd->plane[0].dst.stride = get_frame_new_buffer(cm)->y_stride;  | 
175  |  |  | 
176  |  |   // do intra 16x16 prediction  | 
177  | 0  |   intra_error = find_best_16x16_intra(cpi, &stats->ref[INTRA_FRAME].m.mode);  | 
178  | 0  |   if (intra_error <= 0) intra_error = 1;  | 
179  | 0  |   stats->ref[INTRA_FRAME].err = intra_error;  | 
180  |  |  | 
181  |  |   // Golden frame MV search, if it exists and is different than last frame  | 
182  | 0  |   if (golden_ref) { | 
183  | 0  |     int g_motion_error;  | 
184  | 0  |     xd->plane[0].pre[0].buf = golden_ref->y_buffer + mb_y_offset;  | 
185  | 0  |     xd->plane[0].pre[0].stride = golden_ref->y_stride;  | 
186  | 0  |     g_motion_error =  | 
187  | 0  |         do_16x16_motion_search(cpi, prev_golden_ref_mv,  | 
188  | 0  |                                &stats->ref[GOLDEN_FRAME].m.mv, mb_row, mb_col);  | 
189  | 0  |     stats->ref[GOLDEN_FRAME].err = g_motion_error;  | 
190  | 0  |   } else { | 
191  | 0  |     stats->ref[GOLDEN_FRAME].err = INT_MAX;  | 
192  | 0  |     stats->ref[GOLDEN_FRAME].m.mv.as_int = 0;  | 
193  | 0  |   }  | 
194  |  |  | 
195  |  |   // Do an Alt-ref frame MV search, if it exists and is different than  | 
196  |  |   // last/golden frame.  | 
197  | 0  |   if (alt_ref) { | 
198  | 0  |     int a_motion_error;  | 
199  | 0  |     xd->plane[0].pre[0].buf = alt_ref->y_buffer + mb_y_offset;  | 
200  | 0  |     xd->plane[0].pre[0].stride = alt_ref->y_stride;  | 
201  | 0  |     a_motion_error =  | 
202  | 0  |         do_16x16_zerozero_search(cpi, &stats->ref[ALTREF_FRAME].m.mv);  | 
203  |  | 
  | 
204  | 0  |     stats->ref[ALTREF_FRAME].err = a_motion_error;  | 
205  | 0  |   } else { | 
206  | 0  |     stats->ref[ALTREF_FRAME].err = INT_MAX;  | 
207  | 0  |     stats->ref[ALTREF_FRAME].m.mv.as_int = 0;  | 
208  | 0  |   }  | 
209  | 0  | }  | 
210  |  |  | 
211  |  | static void update_mbgraph_frame_stats(VP9_COMP *cpi,  | 
212  |  |                                        MBGRAPH_FRAME_STATS *stats,  | 
213  |  |                                        YV12_BUFFER_CONFIG *buf,  | 
214  |  |                                        YV12_BUFFER_CONFIG *golden_ref,  | 
215  | 0  |                                        YV12_BUFFER_CONFIG *alt_ref) { | 
216  | 0  |   MACROBLOCK *const x = &cpi->td.mb;  | 
217  | 0  |   MACROBLOCKD *const xd = &x->e_mbd;  | 
218  | 0  |   VP9_COMMON *const cm = &cpi->common;  | 
219  |  | 
  | 
220  | 0  |   int mb_col, mb_row, offset = 0;  | 
221  | 0  |   int mb_y_offset = 0;  | 
222  | 0  |   MV gld_top_mv = { 0, 0 }; | 
223  | 0  |   MODE_INFO mi_local;  | 
224  | 0  |   MODE_INFO mi_above, mi_left;  | 
225  |  | 
  | 
226  | 0  |   vp9_zero(mi_local);  | 
227  |  |   // Set up limit values for motion vectors to prevent them extending outside  | 
228  |  |   // the UMV borders.  | 
229  | 0  |   x->mv_limits.row_min = -BORDER_MV_PIXELS_B16;  | 
230  | 0  |   x->mv_limits.row_max = (cm->mb_rows - 1) * 8 + BORDER_MV_PIXELS_B16;  | 
231  |  |   // Signal to vp9_predict_intra_block() that above is not available  | 
232  | 0  |   xd->above_mi = NULL;  | 
233  |  | 
  | 
234  | 0  |   xd->plane[0].dst.stride = buf->y_stride;  | 
235  | 0  |   xd->plane[0].pre[0].stride = buf->y_stride;  | 
236  | 0  |   xd->plane[1].dst.stride = buf->uv_stride;  | 
237  | 0  |   xd->mi[0] = &mi_local;  | 
238  | 0  |   mi_local.sb_type = BLOCK_16X16;  | 
239  | 0  |   mi_local.ref_frame[0] = LAST_FRAME;  | 
240  | 0  |   mi_local.ref_frame[1] = NO_REF_FRAME;  | 
241  |  | 
  | 
242  | 0  |   for (mb_row = 0; mb_row < cm->mb_rows; mb_row++) { | 
243  | 0  |     MV gld_left_mv = gld_top_mv;  | 
244  | 0  |     int mb_y_in_offset = mb_y_offset;  | 
245  |  |  | 
246  |  |     // Set up limit values for motion vectors to prevent them extending outside  | 
247  |  |     // the UMV borders.  | 
248  | 0  |     x->mv_limits.col_min = -BORDER_MV_PIXELS_B16;  | 
249  | 0  |     x->mv_limits.col_max = (cm->mb_cols - 1) * 8 + BORDER_MV_PIXELS_B16;  | 
250  |  |     // Signal to vp9_predict_intra_block() that left is not available  | 
251  | 0  |     xd->left_mi = NULL;  | 
252  |  | 
  | 
253  | 0  |     for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) { | 
254  | 0  |       MBGRAPH_MB_STATS *mb_stats = &stats->mb_stats[offset + mb_col];  | 
255  |  | 
  | 
256  | 0  |       update_mbgraph_mb_stats(cpi, mb_stats, buf, mb_y_in_offset, golden_ref,  | 
257  | 0  |                               &gld_left_mv, alt_ref, mb_row, mb_col);  | 
258  | 0  |       gld_left_mv = mb_stats->ref[GOLDEN_FRAME].m.mv.as_mv;  | 
259  | 0  |       if (mb_col == 0) { | 
260  | 0  |         gld_top_mv = gld_left_mv;  | 
261  | 0  |       }  | 
262  |  |       // Signal to vp9_predict_intra_block() that left is available  | 
263  | 0  |       xd->left_mi = &mi_left;  | 
264  |  | 
  | 
265  | 0  |       mb_y_in_offset += 16;  | 
266  | 0  |       x->mv_limits.col_min -= 16;  | 
267  | 0  |       x->mv_limits.col_max -= 16;  | 
268  | 0  |     }  | 
269  |  |  | 
270  |  |     // Signal to vp9_predict_intra_block() that above is available  | 
271  | 0  |     xd->above_mi = &mi_above;  | 
272  |  | 
  | 
273  | 0  |     mb_y_offset += buf->y_stride * 16;  | 
274  | 0  |     x->mv_limits.row_min -= 16;  | 
275  | 0  |     x->mv_limits.row_max -= 16;  | 
276  | 0  |     offset += cm->mb_cols;  | 
277  | 0  |   }  | 
278  | 0  | }  | 
279  |  |  | 
280  |  | // void separate_arf_mbs_byzz  | 
281  | 0  | static void separate_arf_mbs(VP9_COMP *cpi) { | 
282  | 0  |   VP9_COMMON *const cm = &cpi->common;  | 
283  | 0  |   int mb_col, mb_row, offset, i;  | 
284  | 0  |   int mi_row, mi_col;  | 
285  | 0  |   int ncnt[4] = { 0 }; | 
286  | 0  |   int n_frames = cpi->mbgraph_n_frames;  | 
287  |  | 
  | 
288  | 0  |   int *arf_not_zz;  | 
289  |  | 
  | 
290  | 0  |   CHECK_MEM_ERROR(  | 
291  | 0  |       &cm->error, arf_not_zz,  | 
292  | 0  |       vpx_calloc(cm->mb_rows * cm->mb_cols * sizeof(*arf_not_zz), 1));  | 
293  |  |  | 
294  |  |   // We are not interested in results beyond the alt ref itself.  | 
295  | 0  |   if (n_frames > cpi->rc.frames_till_gf_update_due)  | 
296  | 0  |     n_frames = cpi->rc.frames_till_gf_update_due;  | 
297  |  |  | 
298  |  |   // defer cost to reference frames  | 
299  | 0  |   for (i = n_frames - 1; i >= 0; i--) { | 
300  | 0  |     MBGRAPH_FRAME_STATS *frame_stats = &cpi->mbgraph_stats[i];  | 
301  |  | 
  | 
302  | 0  |     for (offset = 0, mb_row = 0; mb_row < cm->mb_rows;  | 
303  | 0  |          offset += cm->mb_cols, mb_row++) { | 
304  | 0  |       for (mb_col = 0; mb_col < cm->mb_cols; mb_col++) { | 
305  | 0  |         MBGRAPH_MB_STATS *mb_stats = &frame_stats->mb_stats[offset + mb_col];  | 
306  |  | 
  | 
307  | 0  |         int altref_err = mb_stats->ref[ALTREF_FRAME].err;  | 
308  | 0  |         int intra_err = mb_stats->ref[INTRA_FRAME].err;  | 
309  | 0  |         int golden_err = mb_stats->ref[GOLDEN_FRAME].err;  | 
310  |  |  | 
311  |  |         // Test for altref vs intra and gf and that its mv was 0,0.  | 
312  | 0  |         if (altref_err > 1000 || altref_err > intra_err ||  | 
313  | 0  |             altref_err > golden_err) { | 
314  | 0  |           arf_not_zz[offset + mb_col]++;  | 
315  | 0  |         }  | 
316  | 0  |       }  | 
317  | 0  |     }  | 
318  | 0  |   }  | 
319  |  |  | 
320  |  |   // arf_not_zz is indexed by MB, but this loop is indexed by MI to avoid out  | 
321  |  |   // of bound access in segmentation_map  | 
322  | 0  |   for (mi_row = 0; mi_row < cm->mi_rows; mi_row++) { | 
323  | 0  |     for (mi_col = 0; mi_col < cm->mi_cols; mi_col++) { | 
324  |  |       // If any of the blocks in the sequence failed then the MB  | 
325  |  |       // goes in segment 0  | 
326  | 0  |       if (arf_not_zz[mi_row / 2 * cm->mb_cols + mi_col / 2]) { | 
327  | 0  |         ncnt[0]++;  | 
328  | 0  |         cpi->segmentation_map[mi_row * cm->mi_cols + mi_col] = 0;  | 
329  | 0  |       } else { | 
330  | 0  |         cpi->segmentation_map[mi_row * cm->mi_cols + mi_col] = 1;  | 
331  | 0  |         ncnt[1]++;  | 
332  | 0  |       }  | 
333  | 0  |     }  | 
334  | 0  |   }  | 
335  |  |  | 
336  |  |   // Note % of blocks that are marked as static  | 
337  | 0  |   if (cm->MBs)  | 
338  | 0  |     cpi->static_mb_pct = (ncnt[1] * 100) / (cm->mi_rows * cm->mi_cols);  | 
339  |  |  | 
340  |  |   // This error case should not be reachable as this function should  | 
341  |  |   // never be called with the common data structure uninitialized.  | 
342  | 0  |   else  | 
343  | 0  |     cpi->static_mb_pct = 0;  | 
344  |  | 
  | 
345  | 0  |   vp9_enable_segmentation(&cm->seg);  | 
346  |  |  | 
347  |  |   // Free localy allocated storage  | 
348  | 0  |   vpx_free(arf_not_zz);  | 
349  | 0  | }  | 
350  |  |  | 
351  | 0  | void vp9_update_mbgraph_stats(VP9_COMP *cpi) { | 
352  | 0  |   VP9_COMMON *const cm = &cpi->common;  | 
353  | 0  |   int i, n_frames = vp9_lookahead_depth(cpi->lookahead);  | 
354  | 0  |   YV12_BUFFER_CONFIG *golden_ref = get_ref_frame_buffer(cpi, GOLDEN_FRAME);  | 
355  |  | 
  | 
356  | 0  |   assert(golden_ref != NULL);  | 
357  |  |  | 
358  |  |   // we need to look ahead beyond where the ARF transitions into  | 
359  |  |   // being a GF - so exit if we don't look ahead beyond that  | 
360  | 0  |   if (n_frames <= cpi->rc.frames_till_gf_update_due) return;  | 
361  |  |  | 
362  | 0  |   if (n_frames > MAX_LAG_BUFFERS) n_frames = MAX_LAG_BUFFERS;  | 
363  |  | 
  | 
364  | 0  |   cpi->mbgraph_n_frames = n_frames;  | 
365  | 0  |   for (i = 0; i < n_frames; i++) { | 
366  | 0  |     MBGRAPH_FRAME_STATS *frame_stats = &cpi->mbgraph_stats[i];  | 
367  | 0  |     memset(frame_stats->mb_stats, 0,  | 
368  | 0  |            cm->mb_rows * cm->mb_cols * sizeof(*cpi->mbgraph_stats[i].mb_stats));  | 
369  | 0  |   }  | 
370  |  |  | 
371  |  |   // do motion search to find contribution of each reference to data  | 
372  |  |   // later on in this GF group  | 
373  |  |   // FIXME really, the GF/last MC search should be done forward, and  | 
374  |  |   // the ARF MC search backwards, to get optimal results for MV caching  | 
375  | 0  |   for (i = 0; i < n_frames; i++) { | 
376  | 0  |     MBGRAPH_FRAME_STATS *frame_stats = &cpi->mbgraph_stats[i];  | 
377  | 0  |     struct lookahead_entry *q_cur = vp9_lookahead_peek(cpi->lookahead, i);  | 
378  |  | 
  | 
379  | 0  |     assert(q_cur != NULL);  | 
380  |  | 
  | 
381  | 0  |     update_mbgraph_frame_stats(cpi, frame_stats, &q_cur->img, golden_ref,  | 
382  | 0  |                                cpi->Source);  | 
383  | 0  |   }  | 
384  |  | 
  | 
385  | 0  |   vpx_clear_system_state();  | 
386  |  | 
  | 
387  | 0  |   separate_arf_mbs(cpi);  | 
388  | 0  | }  |