/src/aom/av1/encoder/encoder_alloc.h
Line | Count | Source (jump to first uncovered line) |
1 | | /* |
2 | | * Copyright (c) 2020, Alliance for Open Media. All rights reserved. |
3 | | * |
4 | | * This source code is subject to the terms of the BSD 2 Clause License and |
5 | | * the Alliance for Open Media Patent License 1.0. If the BSD 2 Clause License |
6 | | * was not distributed with this source code in the LICENSE file, you can |
7 | | * obtain it at www.aomedia.org/license/software. If the Alliance for Open |
8 | | * Media Patent License 1.0 was not distributed with this source code in the |
9 | | * PATENTS file, you can obtain it at www.aomedia.org/license/patent. |
10 | | */ |
11 | | |
12 | | #ifndef AOM_AV1_ENCODER_ENCODER_ALLOC_H_ |
13 | | #define AOM_AV1_ENCODER_ENCODER_ALLOC_H_ |
14 | | |
15 | | #include "av1/encoder/block.h" |
16 | | #include "av1/encoder/encodeframe_utils.h" |
17 | | #include "av1/encoder/encoder.h" |
18 | | #include "av1/encoder/encodetxb.h" |
19 | | #include "av1/encoder/ethread.h" |
20 | | #include "av1/encoder/global_motion_facade.h" |
21 | | #include "av1/encoder/intra_mode_search_utils.h" |
22 | | #include "av1/encoder/pickcdef.h" |
23 | | |
24 | | #ifdef __cplusplus |
25 | | extern "C" { |
26 | | #endif |
27 | | |
28 | | static inline void dealloc_context_buffers_ext( |
29 | 0 | MBMIExtFrameBufferInfo *mbmi_ext_info) { |
30 | 0 | aom_free(mbmi_ext_info->frame_base); |
31 | 0 | mbmi_ext_info->frame_base = NULL; |
32 | 0 | mbmi_ext_info->alloc_size = 0; |
33 | 0 | } Unexecuted instantiation: av1_cx_iface.c:dealloc_context_buffers_ext Unexecuted instantiation: encoder.c:dealloc_context_buffers_ext Unexecuted instantiation: encoder_utils.c:dealloc_context_buffers_ext Unexecuted instantiation: ethread.c:dealloc_context_buffers_ext Unexecuted instantiation: superres_scale.c:dealloc_context_buffers_ext Unexecuted instantiation: svc_layercontext.c:dealloc_context_buffers_ext Unexecuted instantiation: compound_type.c:dealloc_context_buffers_ext Unexecuted instantiation: encode_strategy.c:dealloc_context_buffers_ext |
34 | | |
35 | | static inline void alloc_context_buffers_ext( |
36 | 0 | AV1_COMMON *cm, MBMIExtFrameBufferInfo *mbmi_ext_info) { |
37 | 0 | const CommonModeInfoParams *const mi_params = &cm->mi_params; |
38 | |
|
39 | 0 | const int mi_alloc_size_1d = mi_size_wide[mi_params->mi_alloc_bsize]; |
40 | 0 | const int mi_alloc_rows = |
41 | 0 | (mi_params->mi_rows + mi_alloc_size_1d - 1) / mi_alloc_size_1d; |
42 | 0 | const int mi_alloc_cols = |
43 | 0 | (mi_params->mi_cols + mi_alloc_size_1d - 1) / mi_alloc_size_1d; |
44 | 0 | const int new_ext_mi_size = mi_alloc_rows * mi_alloc_cols; |
45 | |
|
46 | 0 | if (new_ext_mi_size > mbmi_ext_info->alloc_size) { |
47 | 0 | dealloc_context_buffers_ext(mbmi_ext_info); |
48 | 0 | CHECK_MEM_ERROR( |
49 | 0 | cm, mbmi_ext_info->frame_base, |
50 | 0 | aom_malloc(new_ext_mi_size * sizeof(*mbmi_ext_info->frame_base))); |
51 | 0 | mbmi_ext_info->alloc_size = new_ext_mi_size; |
52 | 0 | } |
53 | | // The stride needs to be updated regardless of whether new allocation |
54 | | // happened or not. |
55 | 0 | mbmi_ext_info->stride = mi_alloc_cols; |
56 | 0 | } Unexecuted instantiation: av1_cx_iface.c:alloc_context_buffers_ext Unexecuted instantiation: encoder.c:alloc_context_buffers_ext Unexecuted instantiation: encoder_utils.c:alloc_context_buffers_ext Unexecuted instantiation: ethread.c:alloc_context_buffers_ext Unexecuted instantiation: superres_scale.c:alloc_context_buffers_ext Unexecuted instantiation: svc_layercontext.c:alloc_context_buffers_ext Unexecuted instantiation: compound_type.c:alloc_context_buffers_ext Unexecuted instantiation: encode_strategy.c:alloc_context_buffers_ext |
57 | | |
58 | 0 | static inline void alloc_compressor_data(AV1_COMP *cpi) { |
59 | 0 | AV1_COMMON *cm = &cpi->common; |
60 | 0 | CommonModeInfoParams *const mi_params = &cm->mi_params; |
61 | | |
62 | | // Setup mi_params |
63 | 0 | mi_params->set_mb_mi(mi_params, cm->width, cm->height, |
64 | 0 | cpi->sf.part_sf.default_min_partition_size); |
65 | |
|
66 | 0 | if (!is_stat_generation_stage(cpi)) av1_alloc_txb_buf(cpi); |
67 | |
|
68 | 0 | aom_free(cpi->td.mv_costs_alloc); |
69 | 0 | cpi->td.mv_costs_alloc = NULL; |
70 | | // Avoid the memory allocation of 'mv_costs_alloc' for allintra encoding |
71 | | // mode. |
72 | 0 | if (cpi->oxcf.kf_cfg.key_freq_max != 0) { |
73 | 0 | CHECK_MEM_ERROR(cm, cpi->td.mv_costs_alloc, |
74 | 0 | (MvCosts *)aom_calloc(1, sizeof(*cpi->td.mv_costs_alloc))); |
75 | 0 | cpi->td.mb.mv_costs = cpi->td.mv_costs_alloc; |
76 | 0 | } |
77 | |
|
78 | 0 | av1_setup_shared_coeff_buffer(cm->seq_params, &cpi->td.shared_coeff_buf, |
79 | 0 | cm->error); |
80 | 0 | if (av1_setup_sms_tree(cpi, &cpi->td)) { |
81 | 0 | aom_internal_error(cm->error, AOM_CODEC_MEM_ERROR, |
82 | 0 | "Failed to allocate SMS tree"); |
83 | 0 | } |
84 | 0 | cpi->td.firstpass_ctx = |
85 | 0 | av1_alloc_pmc(cpi, BLOCK_16X16, &cpi->td.shared_coeff_buf); |
86 | 0 | if (!cpi->td.firstpass_ctx) |
87 | 0 | aom_internal_error(cm->error, AOM_CODEC_MEM_ERROR, |
88 | 0 | "Failed to allocate PICK_MODE_CONTEXT"); |
89 | 0 | } Unexecuted instantiation: av1_cx_iface.c:alloc_compressor_data Unexecuted instantiation: encoder.c:alloc_compressor_data Unexecuted instantiation: encoder_utils.c:alloc_compressor_data Unexecuted instantiation: ethread.c:alloc_compressor_data Unexecuted instantiation: superres_scale.c:alloc_compressor_data Unexecuted instantiation: svc_layercontext.c:alloc_compressor_data Unexecuted instantiation: compound_type.c:alloc_compressor_data Unexecuted instantiation: encode_strategy.c:alloc_compressor_data |
90 | | |
91 | | // Allocate mbmi buffers which are used to store mode information at block |
92 | | // level. |
93 | 0 | static inline void alloc_mb_mode_info_buffers(AV1_COMP *const cpi) { |
94 | 0 | AV1_COMMON *const cm = &cpi->common; |
95 | 0 | if (av1_alloc_context_buffers(cm, cm->width, cm->height, |
96 | 0 | cpi->sf.part_sf.default_min_partition_size)) { |
97 | 0 | aom_internal_error(cm->error, AOM_CODEC_MEM_ERROR, |
98 | 0 | "Failed to allocate context buffers"); |
99 | 0 | } |
100 | |
|
101 | 0 | if (!is_stat_generation_stage(cpi)) |
102 | 0 | alloc_context_buffers_ext(cm, &cpi->mbmi_ext_info); |
103 | 0 | } Unexecuted instantiation: av1_cx_iface.c:alloc_mb_mode_info_buffers Unexecuted instantiation: encoder.c:alloc_mb_mode_info_buffers Unexecuted instantiation: encoder_utils.c:alloc_mb_mode_info_buffers Unexecuted instantiation: ethread.c:alloc_mb_mode_info_buffers Unexecuted instantiation: superres_scale.c:alloc_mb_mode_info_buffers Unexecuted instantiation: svc_layercontext.c:alloc_mb_mode_info_buffers Unexecuted instantiation: compound_type.c:alloc_mb_mode_info_buffers Unexecuted instantiation: encode_strategy.c:alloc_mb_mode_info_buffers |
104 | | |
105 | 0 | static inline void realloc_segmentation_maps(AV1_COMP *cpi) { |
106 | 0 | AV1_COMMON *const cm = &cpi->common; |
107 | 0 | CommonModeInfoParams *const mi_params = &cm->mi_params; |
108 | | |
109 | | // Create the encoder segmentation map and set all entries to 0 |
110 | 0 | aom_free(cpi->enc_seg.map); |
111 | 0 | CHECK_MEM_ERROR(cm, cpi->enc_seg.map, |
112 | 0 | aom_calloc(mi_params->mi_rows * mi_params->mi_cols, 1)); |
113 | | |
114 | | // Create a map used for cyclic background refresh. |
115 | 0 | if (cpi->cyclic_refresh) av1_cyclic_refresh_free(cpi->cyclic_refresh); |
116 | 0 | CHECK_MEM_ERROR( |
117 | 0 | cm, cpi->cyclic_refresh, |
118 | 0 | av1_cyclic_refresh_alloc(mi_params->mi_rows, mi_params->mi_cols)); |
119 | | |
120 | | // Create a map used to mark inactive areas. |
121 | 0 | aom_free(cpi->active_map.map); |
122 | 0 | CHECK_MEM_ERROR(cm, cpi->active_map.map, |
123 | 0 | aom_calloc(mi_params->mi_rows * mi_params->mi_cols, 1)); |
124 | 0 | } Unexecuted instantiation: av1_cx_iface.c:realloc_segmentation_maps Unexecuted instantiation: encoder.c:realloc_segmentation_maps Unexecuted instantiation: encoder_utils.c:realloc_segmentation_maps Unexecuted instantiation: ethread.c:realloc_segmentation_maps Unexecuted instantiation: superres_scale.c:realloc_segmentation_maps Unexecuted instantiation: svc_layercontext.c:realloc_segmentation_maps Unexecuted instantiation: compound_type.c:realloc_segmentation_maps Unexecuted instantiation: encode_strategy.c:realloc_segmentation_maps |
125 | | |
126 | | static inline void alloc_obmc_buffers(OBMCBuffer *obmc_buffer, |
127 | 0 | struct aom_internal_error_info *error) { |
128 | 0 | AOM_CHECK_MEM_ERROR( |
129 | 0 | error, obmc_buffer->wsrc, |
130 | 0 | (int32_t *)aom_memalign(16, MAX_SB_SQUARE * sizeof(*obmc_buffer->wsrc))); |
131 | 0 | AOM_CHECK_MEM_ERROR( |
132 | 0 | error, obmc_buffer->mask, |
133 | 0 | (int32_t *)aom_memalign(16, MAX_SB_SQUARE * sizeof(*obmc_buffer->mask))); |
134 | 0 | AOM_CHECK_MEM_ERROR( |
135 | 0 | error, obmc_buffer->above_pred, |
136 | 0 | (uint8_t *)aom_memalign( |
137 | 0 | 16, MAX_MB_PLANE * MAX_SB_SQUARE * sizeof(*obmc_buffer->above_pred))); |
138 | 0 | AOM_CHECK_MEM_ERROR( |
139 | 0 | error, obmc_buffer->left_pred, |
140 | 0 | (uint8_t *)aom_memalign( |
141 | 0 | 16, MAX_MB_PLANE * MAX_SB_SQUARE * sizeof(*obmc_buffer->left_pred))); |
142 | 0 | } Unexecuted instantiation: av1_cx_iface.c:alloc_obmc_buffers Unexecuted instantiation: encoder.c:alloc_obmc_buffers Unexecuted instantiation: encoder_utils.c:alloc_obmc_buffers Unexecuted instantiation: ethread.c:alloc_obmc_buffers Unexecuted instantiation: superres_scale.c:alloc_obmc_buffers Unexecuted instantiation: svc_layercontext.c:alloc_obmc_buffers Unexecuted instantiation: compound_type.c:alloc_obmc_buffers Unexecuted instantiation: encode_strategy.c:alloc_obmc_buffers |
143 | | |
144 | 0 | static inline void release_obmc_buffers(OBMCBuffer *obmc_buffer) { |
145 | 0 | aom_free(obmc_buffer->mask); |
146 | 0 | aom_free(obmc_buffer->above_pred); |
147 | 0 | aom_free(obmc_buffer->left_pred); |
148 | 0 | aom_free(obmc_buffer->wsrc); |
149 | |
|
150 | 0 | obmc_buffer->mask = NULL; |
151 | 0 | obmc_buffer->above_pred = NULL; |
152 | 0 | obmc_buffer->left_pred = NULL; |
153 | 0 | obmc_buffer->wsrc = NULL; |
154 | 0 | } Unexecuted instantiation: av1_cx_iface.c:release_obmc_buffers Unexecuted instantiation: encoder.c:release_obmc_buffers Unexecuted instantiation: encoder_utils.c:release_obmc_buffers Unexecuted instantiation: ethread.c:release_obmc_buffers Unexecuted instantiation: superres_scale.c:release_obmc_buffers Unexecuted instantiation: svc_layercontext.c:release_obmc_buffers Unexecuted instantiation: compound_type.c:release_obmc_buffers Unexecuted instantiation: encode_strategy.c:release_obmc_buffers |
155 | | |
156 | | static inline void alloc_compound_type_rd_buffers( |
157 | 0 | struct aom_internal_error_info *error, CompoundTypeRdBuffers *const bufs) { |
158 | 0 | AOM_CHECK_MEM_ERROR( |
159 | 0 | error, bufs->pred0, |
160 | 0 | (uint8_t *)aom_memalign(16, 2 * MAX_SB_SQUARE * sizeof(*bufs->pred0))); |
161 | 0 | AOM_CHECK_MEM_ERROR( |
162 | 0 | error, bufs->pred1, |
163 | 0 | (uint8_t *)aom_memalign(16, 2 * MAX_SB_SQUARE * sizeof(*bufs->pred1))); |
164 | 0 | AOM_CHECK_MEM_ERROR( |
165 | 0 | error, bufs->residual1, |
166 | 0 | (int16_t *)aom_memalign(32, MAX_SB_SQUARE * sizeof(*bufs->residual1))); |
167 | 0 | AOM_CHECK_MEM_ERROR( |
168 | 0 | error, bufs->diff10, |
169 | 0 | (int16_t *)aom_memalign(32, MAX_SB_SQUARE * sizeof(*bufs->diff10))); |
170 | 0 | AOM_CHECK_MEM_ERROR(error, bufs->tmp_best_mask_buf, |
171 | 0 | (uint8_t *)aom_malloc(2 * MAX_SB_SQUARE * |
172 | 0 | sizeof(*bufs->tmp_best_mask_buf))); |
173 | 0 | } Unexecuted instantiation: av1_cx_iface.c:alloc_compound_type_rd_buffers Unexecuted instantiation: encoder.c:alloc_compound_type_rd_buffers Unexecuted instantiation: encoder_utils.c:alloc_compound_type_rd_buffers Unexecuted instantiation: ethread.c:alloc_compound_type_rd_buffers Unexecuted instantiation: superres_scale.c:alloc_compound_type_rd_buffers Unexecuted instantiation: svc_layercontext.c:alloc_compound_type_rd_buffers Unexecuted instantiation: compound_type.c:alloc_compound_type_rd_buffers Unexecuted instantiation: encode_strategy.c:alloc_compound_type_rd_buffers |
174 | | |
175 | | static inline void release_compound_type_rd_buffers( |
176 | 0 | CompoundTypeRdBuffers *const bufs) { |
177 | 0 | aom_free(bufs->pred0); |
178 | 0 | aom_free(bufs->pred1); |
179 | 0 | aom_free(bufs->residual1); |
180 | 0 | aom_free(bufs->diff10); |
181 | 0 | aom_free(bufs->tmp_best_mask_buf); |
182 | 0 | av1_zero(*bufs); // Set all pointers to NULL for safety. |
183 | 0 | } Unexecuted instantiation: av1_cx_iface.c:release_compound_type_rd_buffers Unexecuted instantiation: encoder.c:release_compound_type_rd_buffers Unexecuted instantiation: encoder_utils.c:release_compound_type_rd_buffers Unexecuted instantiation: ethread.c:release_compound_type_rd_buffers Unexecuted instantiation: superres_scale.c:release_compound_type_rd_buffers Unexecuted instantiation: svc_layercontext.c:release_compound_type_rd_buffers Unexecuted instantiation: compound_type.c:release_compound_type_rd_buffers Unexecuted instantiation: encode_strategy.c:release_compound_type_rd_buffers |
184 | | |
185 | 0 | static inline void dealloc_compressor_data(AV1_COMP *cpi) { |
186 | 0 | AV1_COMMON *const cm = &cpi->common; |
187 | 0 | TokenInfo *token_info = &cpi->token_info; |
188 | 0 | AV1EncRowMultiThreadInfo *const enc_row_mt = &cpi->mt_info.enc_row_mt; |
189 | 0 | const int num_planes = av1_num_planes(cm); |
190 | 0 | dealloc_context_buffers_ext(&cpi->mbmi_ext_info); |
191 | |
|
192 | 0 | aom_free(cpi->tile_data); |
193 | 0 | cpi->tile_data = NULL; |
194 | 0 | cpi->allocated_tiles = 0; |
195 | 0 | enc_row_mt->allocated_tile_cols = 0; |
196 | 0 | enc_row_mt->allocated_tile_rows = 0; |
197 | | |
198 | | // Delete sementation map |
199 | 0 | aom_free(cpi->enc_seg.map); |
200 | 0 | cpi->enc_seg.map = NULL; |
201 | |
|
202 | 0 | av1_cyclic_refresh_free(cpi->cyclic_refresh); |
203 | 0 | cpi->cyclic_refresh = NULL; |
204 | |
|
205 | 0 | aom_free(cpi->active_map.map); |
206 | 0 | cpi->active_map.map = NULL; |
207 | |
|
208 | 0 | aom_free(cpi->ssim_rdmult_scaling_factors); |
209 | 0 | cpi->ssim_rdmult_scaling_factors = NULL; |
210 | |
|
211 | 0 | aom_free(cpi->tpl_rdmult_scaling_factors); |
212 | 0 | cpi->tpl_rdmult_scaling_factors = NULL; |
213 | |
|
214 | | #if CONFIG_TUNE_VMAF |
215 | | aom_free(cpi->vmaf_info.rdmult_scaling_factors); |
216 | | cpi->vmaf_info.rdmult_scaling_factors = NULL; |
217 | | aom_close_vmaf_model(cpi->vmaf_info.vmaf_model); |
218 | | #endif |
219 | |
|
220 | | #if CONFIG_TUNE_BUTTERAUGLI |
221 | | aom_free(cpi->butteraugli_info.rdmult_scaling_factors); |
222 | | cpi->butteraugli_info.rdmult_scaling_factors = NULL; |
223 | | aom_free_frame_buffer(&cpi->butteraugli_info.source); |
224 | | aom_free_frame_buffer(&cpi->butteraugli_info.resized_source); |
225 | | #endif |
226 | |
|
227 | | #if CONFIG_SALIENCY_MAP |
228 | | aom_free(cpi->saliency_map); |
229 | | aom_free(cpi->sm_scaling_factor); |
230 | | #endif |
231 | |
|
232 | 0 | release_obmc_buffers(&cpi->td.mb.obmc_buffer); |
233 | |
|
234 | 0 | aom_free(cpi->td.mv_costs_alloc); |
235 | 0 | cpi->td.mv_costs_alloc = NULL; |
236 | 0 | aom_free(cpi->td.dv_costs_alloc); |
237 | 0 | cpi->td.dv_costs_alloc = NULL; |
238 | |
|
239 | 0 | aom_free(cpi->td.mb.sb_stats_cache); |
240 | 0 | cpi->td.mb.sb_stats_cache = NULL; |
241 | |
|
242 | 0 | aom_free(cpi->td.mb.sb_fp_stats); |
243 | 0 | cpi->td.mb.sb_fp_stats = NULL; |
244 | |
|
245 | | #if CONFIG_PARTITION_SEARCH_ORDER |
246 | | aom_free(cpi->td.mb.rdcost); |
247 | | cpi->td.mb.rdcost = NULL; |
248 | | #endif |
249 | |
|
250 | 0 | av1_free_pc_tree_recursive(cpi->td.pc_root, num_planes, 0, 0, |
251 | 0 | cpi->sf.part_sf.partition_search_type); |
252 | 0 | cpi->td.pc_root = NULL; |
253 | |
|
254 | 0 | for (int i = 0; i < 2; i++) |
255 | 0 | for (int j = 0; j < 2; j++) { |
256 | 0 | aom_free(cpi->td.mb.intrabc_hash_info.hash_value_buffer[i][j]); |
257 | 0 | cpi->td.mb.intrabc_hash_info.hash_value_buffer[i][j] = NULL; |
258 | 0 | } |
259 | |
|
260 | 0 | av1_hash_table_destroy(&cpi->td.mb.intrabc_hash_info.intrabc_hash_table); |
261 | |
|
262 | 0 | aom_free(cm->tpl_mvs); |
263 | 0 | cm->tpl_mvs = NULL; |
264 | |
|
265 | 0 | aom_free(cpi->td.pixel_gradient_info); |
266 | 0 | cpi->td.pixel_gradient_info = NULL; |
267 | |
|
268 | 0 | aom_free(cpi->td.src_var_info_of_4x4_sub_blocks); |
269 | 0 | cpi->td.src_var_info_of_4x4_sub_blocks = NULL; |
270 | |
|
271 | 0 | aom_free(cpi->td.vt64x64); |
272 | 0 | cpi->td.vt64x64 = NULL; |
273 | |
|
274 | 0 | av1_free_pmc(cpi->td.firstpass_ctx, num_planes); |
275 | 0 | cpi->td.firstpass_ctx = NULL; |
276 | |
|
277 | 0 | const int is_highbitdepth = cpi->tf_ctx.is_highbitdepth; |
278 | | // This call ensures that the buffers allocated by tf_alloc_and_reset_data() |
279 | | // in av1_temporal_filter() for single-threaded encode are freed in case an |
280 | | // error is encountered during temporal filtering (due to early termination |
281 | | // tf_dealloc_data() in av1_temporal_filter() would not be invoked). |
282 | 0 | tf_dealloc_data(&cpi->td.tf_data, is_highbitdepth); |
283 | | |
284 | | // This call ensures that tpl_tmp_buffers for single-threaded encode are freed |
285 | | // in case of an error during tpl. |
286 | 0 | tpl_dealloc_temp_buffers(&cpi->td.tpl_tmp_buffers); |
287 | | |
288 | | // This call ensures that the global motion (gm) data buffers for |
289 | | // single-threaded encode are freed in case of an error during gm. |
290 | 0 | gm_dealloc_data(&cpi->td.gm_data); |
291 | | |
292 | | // This call ensures that CDEF search context buffers are deallocated in case |
293 | | // of an error during cdef search. |
294 | 0 | av1_cdef_dealloc_data(cpi->cdef_search_ctx); |
295 | 0 | aom_free(cpi->cdef_search_ctx); |
296 | 0 | cpi->cdef_search_ctx = NULL; |
297 | |
|
298 | 0 | av1_dealloc_mb_data(&cpi->td.mb, num_planes); |
299 | |
|
300 | 0 | av1_dealloc_mb_wiener_var_pred_buf(&cpi->td); |
301 | |
|
302 | 0 | av1_free_txb_buf(cpi); |
303 | 0 | av1_free_context_buffers(cm); |
304 | |
|
305 | 0 | aom_free_frame_buffer(&cpi->last_frame_uf); |
306 | 0 | #if !CONFIG_REALTIME_ONLY |
307 | 0 | av1_free_restoration_buffers(cm); |
308 | 0 | av1_free_firstpass_data(&cpi->firstpass_data); |
309 | 0 | #endif |
310 | |
|
311 | 0 | if (!is_stat_generation_stage(cpi)) { |
312 | 0 | av1_free_cdef_buffers(cm, &cpi->ppi->p_mt_info.cdef_worker, |
313 | 0 | &cpi->mt_info.cdef_sync); |
314 | 0 | } |
315 | |
|
316 | 0 | for (int plane = 0; plane < num_planes; plane++) { |
317 | 0 | aom_free(cpi->pick_lr_ctxt.rusi[plane]); |
318 | 0 | cpi->pick_lr_ctxt.rusi[plane] = NULL; |
319 | 0 | } |
320 | 0 | aom_free(cpi->pick_lr_ctxt.dgd_avg); |
321 | 0 | cpi->pick_lr_ctxt.dgd_avg = NULL; |
322 | |
|
323 | 0 | aom_free_frame_buffer(&cpi->trial_frame_rst); |
324 | 0 | aom_free_frame_buffer(&cpi->scaled_source); |
325 | 0 | aom_free_frame_buffer(&cpi->scaled_last_source); |
326 | 0 | aom_free_frame_buffer(&cpi->orig_source); |
327 | 0 | aom_free_frame_buffer(&cpi->svc.source_last_TL0); |
328 | |
|
329 | 0 | free_token_info(token_info); |
330 | |
|
331 | 0 | av1_free_shared_coeff_buffer(&cpi->td.shared_coeff_buf); |
332 | 0 | av1_free_sms_tree(&cpi->td); |
333 | |
|
334 | 0 | aom_free(cpi->td.mb.palette_buffer); |
335 | 0 | release_compound_type_rd_buffers(&cpi->td.mb.comp_rd_buffer); |
336 | 0 | aom_free(cpi->td.mb.tmp_conv_dst); |
337 | 0 | for (int j = 0; j < 2; ++j) { |
338 | 0 | aom_free(cpi->td.mb.tmp_pred_bufs[j]); |
339 | 0 | } |
340 | |
|
341 | 0 | #if CONFIG_DENOISE && !CONFIG_REALTIME_ONLY |
342 | 0 | if (cpi->denoise_and_model) { |
343 | 0 | aom_denoise_and_model_free(cpi->denoise_and_model); |
344 | 0 | cpi->denoise_and_model = NULL; |
345 | 0 | } |
346 | 0 | #endif |
347 | 0 | #if !CONFIG_REALTIME_ONLY |
348 | 0 | if (cpi->film_grain_table) { |
349 | 0 | aom_film_grain_table_free(cpi->film_grain_table); |
350 | 0 | aom_free(cpi->film_grain_table); |
351 | 0 | cpi->film_grain_table = NULL; |
352 | 0 | } |
353 | 0 | #endif |
354 | |
|
355 | 0 | if (cpi->ppi->use_svc) av1_free_svc_cyclic_refresh(cpi); |
356 | 0 | aom_free(cpi->svc.layer_context); |
357 | 0 | cpi->svc.layer_context = NULL; |
358 | |
|
359 | 0 | aom_free(cpi->consec_zero_mv); |
360 | 0 | cpi->consec_zero_mv = NULL; |
361 | 0 | cpi->consec_zero_mv_alloc_size = 0; |
362 | |
|
363 | 0 | aom_free(cpi->src_sad_blk_64x64); |
364 | 0 | cpi->src_sad_blk_64x64 = NULL; |
365 | |
|
366 | 0 | aom_free(cpi->mb_weber_stats); |
367 | 0 | cpi->mb_weber_stats = NULL; |
368 | |
|
369 | 0 | if (cpi->oxcf.enable_rate_guide_deltaq) { |
370 | 0 | aom_free(cpi->prep_rate_estimates); |
371 | 0 | cpi->prep_rate_estimates = NULL; |
372 | |
|
373 | 0 | aom_free(cpi->ext_rate_distribution); |
374 | 0 | cpi->ext_rate_distribution = NULL; |
375 | 0 | } |
376 | |
|
377 | 0 | aom_free(cpi->mb_delta_q); |
378 | 0 | cpi->mb_delta_q = NULL; |
379 | 0 | } Unexecuted instantiation: av1_cx_iface.c:dealloc_compressor_data Unexecuted instantiation: encoder.c:dealloc_compressor_data Unexecuted instantiation: encoder_utils.c:dealloc_compressor_data Unexecuted instantiation: ethread.c:dealloc_compressor_data Unexecuted instantiation: superres_scale.c:dealloc_compressor_data Unexecuted instantiation: svc_layercontext.c:dealloc_compressor_data Unexecuted instantiation: compound_type.c:dealloc_compressor_data Unexecuted instantiation: encode_strategy.c:dealloc_compressor_data |
380 | | |
381 | 0 | static inline void allocate_gradient_info_for_hog(AV1_COMP *cpi) { |
382 | 0 | if (!is_gradient_caching_for_hog_enabled(cpi)) return; |
383 | | |
384 | 0 | PixelLevelGradientInfo *pixel_gradient_info = cpi->td.pixel_gradient_info; |
385 | 0 | if (!pixel_gradient_info) { |
386 | 0 | const AV1_COMMON *const cm = &cpi->common; |
387 | 0 | const int plane_types = PLANE_TYPES >> cm->seq_params->monochrome; |
388 | 0 | CHECK_MEM_ERROR( |
389 | 0 | cm, pixel_gradient_info, |
390 | 0 | aom_malloc(sizeof(*pixel_gradient_info) * plane_types * MAX_SB_SQUARE)); |
391 | 0 | cpi->td.pixel_gradient_info = pixel_gradient_info; |
392 | 0 | } |
393 | |
|
394 | 0 | cpi->td.mb.pixel_gradient_info = pixel_gradient_info; |
395 | 0 | } Unexecuted instantiation: av1_cx_iface.c:allocate_gradient_info_for_hog Unexecuted instantiation: encoder.c:allocate_gradient_info_for_hog Unexecuted instantiation: encoder_utils.c:allocate_gradient_info_for_hog Unexecuted instantiation: ethread.c:allocate_gradient_info_for_hog Unexecuted instantiation: superres_scale.c:allocate_gradient_info_for_hog Unexecuted instantiation: svc_layercontext.c:allocate_gradient_info_for_hog Unexecuted instantiation: compound_type.c:allocate_gradient_info_for_hog Unexecuted instantiation: encode_strategy.c:allocate_gradient_info_for_hog |
396 | | |
397 | 0 | static inline void allocate_src_var_of_4x4_sub_block_buf(AV1_COMP *cpi) { |
398 | 0 | if (!is_src_var_for_4x4_sub_blocks_caching_enabled(cpi)) return; |
399 | | |
400 | 0 | Block4x4VarInfo *source_variance_info = |
401 | 0 | cpi->td.src_var_info_of_4x4_sub_blocks; |
402 | 0 | if (!source_variance_info) { |
403 | 0 | const AV1_COMMON *const cm = &cpi->common; |
404 | 0 | const BLOCK_SIZE sb_size = cm->seq_params->sb_size; |
405 | 0 | const int mi_count_in_sb = mi_size_wide[sb_size] * mi_size_high[sb_size]; |
406 | 0 | CHECK_MEM_ERROR(cm, source_variance_info, |
407 | 0 | aom_malloc(sizeof(*source_variance_info) * mi_count_in_sb)); |
408 | 0 | cpi->td.src_var_info_of_4x4_sub_blocks = source_variance_info; |
409 | 0 | } |
410 | |
|
411 | 0 | cpi->td.mb.src_var_info_of_4x4_sub_blocks = source_variance_info; |
412 | 0 | } Unexecuted instantiation: av1_cx_iface.c:allocate_src_var_of_4x4_sub_block_buf Unexecuted instantiation: encoder.c:allocate_src_var_of_4x4_sub_block_buf Unexecuted instantiation: encoder_utils.c:allocate_src_var_of_4x4_sub_block_buf Unexecuted instantiation: ethread.c:allocate_src_var_of_4x4_sub_block_buf Unexecuted instantiation: superres_scale.c:allocate_src_var_of_4x4_sub_block_buf Unexecuted instantiation: svc_layercontext.c:allocate_src_var_of_4x4_sub_block_buf Unexecuted instantiation: compound_type.c:allocate_src_var_of_4x4_sub_block_buf Unexecuted instantiation: encode_strategy.c:allocate_src_var_of_4x4_sub_block_buf |
413 | | |
414 | 0 | static inline void variance_partition_alloc(AV1_COMP *cpi) { |
415 | 0 | AV1_COMMON *const cm = &cpi->common; |
416 | 0 | const int num_64x64_blocks = (cm->seq_params->sb_size == BLOCK_64X64) ? 1 : 4; |
417 | 0 | if (cpi->td.vt64x64) { |
418 | 0 | if (num_64x64_blocks != cpi->td.num_64x64_blocks) { |
419 | 0 | aom_free(cpi->td.vt64x64); |
420 | 0 | cpi->td.vt64x64 = NULL; |
421 | 0 | } |
422 | 0 | } |
423 | 0 | if (!cpi->td.vt64x64) { |
424 | 0 | CHECK_MEM_ERROR(cm, cpi->td.vt64x64, |
425 | 0 | aom_malloc(sizeof(*cpi->td.vt64x64) * num_64x64_blocks)); |
426 | 0 | cpi->td.num_64x64_blocks = num_64x64_blocks; |
427 | 0 | } |
428 | 0 | } Unexecuted instantiation: av1_cx_iface.c:variance_partition_alloc Unexecuted instantiation: encoder.c:variance_partition_alloc Unexecuted instantiation: encoder_utils.c:variance_partition_alloc Unexecuted instantiation: ethread.c:variance_partition_alloc Unexecuted instantiation: superres_scale.c:variance_partition_alloc Unexecuted instantiation: svc_layercontext.c:variance_partition_alloc Unexecuted instantiation: compound_type.c:variance_partition_alloc Unexecuted instantiation: encode_strategy.c:variance_partition_alloc |
429 | | |
430 | | static inline YV12_BUFFER_CONFIG *realloc_and_scale_source(AV1_COMP *cpi, |
431 | | int scaled_width, |
432 | 0 | int scaled_height) { |
433 | 0 | AV1_COMMON *cm = &cpi->common; |
434 | 0 | const int num_planes = av1_num_planes(cm); |
435 | |
|
436 | 0 | if (scaled_width == cpi->unscaled_source->y_crop_width && |
437 | 0 | scaled_height == cpi->unscaled_source->y_crop_height) { |
438 | 0 | return cpi->unscaled_source; |
439 | 0 | } |
440 | | |
441 | 0 | if (aom_realloc_frame_buffer( |
442 | 0 | &cpi->scaled_source, scaled_width, scaled_height, |
443 | 0 | cm->seq_params->subsampling_x, cm->seq_params->subsampling_y, |
444 | 0 | cm->seq_params->use_highbitdepth, AOM_BORDER_IN_PIXELS, |
445 | 0 | cm->features.byte_alignment, NULL, NULL, NULL, cpi->alloc_pyramid, 0)) |
446 | 0 | aom_internal_error(cm->error, AOM_CODEC_MEM_ERROR, |
447 | 0 | "Failed to reallocate scaled source buffer"); |
448 | 0 | assert(cpi->scaled_source.y_crop_width == scaled_width); |
449 | 0 | assert(cpi->scaled_source.y_crop_height == scaled_height); |
450 | 0 | if (!av1_resize_and_extend_frame_nonnormative( |
451 | 0 | cpi->unscaled_source, &cpi->scaled_source, |
452 | 0 | (int)cm->seq_params->bit_depth, num_planes)) |
453 | 0 | aom_internal_error(cm->error, AOM_CODEC_MEM_ERROR, |
454 | 0 | "Failed to reallocate buffers during resize"); |
455 | 0 | return &cpi->scaled_source; |
456 | 0 | } Unexecuted instantiation: av1_cx_iface.c:realloc_and_scale_source Unexecuted instantiation: encoder.c:realloc_and_scale_source Unexecuted instantiation: encoder_utils.c:realloc_and_scale_source Unexecuted instantiation: ethread.c:realloc_and_scale_source Unexecuted instantiation: superres_scale.c:realloc_and_scale_source Unexecuted instantiation: svc_layercontext.c:realloc_and_scale_source Unexecuted instantiation: compound_type.c:realloc_and_scale_source Unexecuted instantiation: encode_strategy.c:realloc_and_scale_source |
457 | | |
458 | | // Deallocate allocated thread_data. |
459 | 0 | static inline void free_thread_data(AV1_PRIMARY *ppi) { |
460 | 0 | PrimaryMultiThreadInfo *const p_mt_info = &ppi->p_mt_info; |
461 | 0 | const int num_tf_workers = |
462 | 0 | AOMMIN(p_mt_info->num_mod_workers[MOD_TF], p_mt_info->num_workers); |
463 | 0 | const int num_tpl_workers = |
464 | 0 | AOMMIN(p_mt_info->num_mod_workers[MOD_TPL], p_mt_info->num_workers); |
465 | 0 | const int is_highbitdepth = ppi->seq_params.use_highbitdepth; |
466 | 0 | const int num_planes = ppi->seq_params.monochrome ? 1 : MAX_MB_PLANE; |
467 | 0 | for (int t = 1; t < p_mt_info->num_workers; ++t) { |
468 | 0 | EncWorkerData *const thread_data = &p_mt_info->tile_thr_data[t]; |
469 | 0 | thread_data->td = thread_data->original_td; |
470 | 0 | ThreadData *const td = thread_data->td; |
471 | 0 | if (!td) continue; |
472 | 0 | aom_free(td->tctx); |
473 | 0 | aom_free(td->palette_buffer); |
474 | 0 | aom_free(td->tmp_conv_dst); |
475 | 0 | release_compound_type_rd_buffers(&td->comp_rd_buffer); |
476 | 0 | for (int j = 0; j < 2; ++j) { |
477 | 0 | aom_free(td->tmp_pred_bufs[j]); |
478 | 0 | } |
479 | 0 | aom_free(td->pixel_gradient_info); |
480 | 0 | aom_free(td->src_var_info_of_4x4_sub_blocks); |
481 | 0 | release_obmc_buffers(&td->obmc_buffer); |
482 | 0 | aom_free(td->vt64x64); |
483 | |
|
484 | 0 | for (int x = 0; x < 2; x++) { |
485 | 0 | for (int y = 0; y < 2; y++) { |
486 | 0 | aom_free(td->hash_value_buffer[x][y]); |
487 | 0 | td->hash_value_buffer[x][y] = NULL; |
488 | 0 | } |
489 | 0 | } |
490 | 0 | aom_free(td->mv_costs_alloc); |
491 | 0 | td->mv_costs_alloc = NULL; |
492 | 0 | aom_free(td->dv_costs_alloc); |
493 | 0 | td->dv_costs_alloc = NULL; |
494 | 0 | aom_free(td->counts); |
495 | 0 | av1_free_pmc(td->firstpass_ctx, num_planes); |
496 | 0 | td->firstpass_ctx = NULL; |
497 | 0 | av1_free_shared_coeff_buffer(&td->shared_coeff_buf); |
498 | 0 | av1_free_sms_tree(td); |
499 | | // This call ensures that the buffers allocated by tf_alloc_and_reset_data() |
500 | | // in prepare_tf_workers() for MT encode are freed in case an error is |
501 | | // encountered during temporal filtering (due to early termination |
502 | | // tf_dealloc_thread_data() in av1_tf_do_filtering_mt() would not be |
503 | | // invoked). |
504 | 0 | if (t < num_tf_workers) tf_dealloc_data(&td->tf_data, is_highbitdepth); |
505 | | // This call ensures that tpl_tmp_buffers for MT encode are freed in case of |
506 | | // an error during tpl. |
507 | 0 | if (t < num_tpl_workers) tpl_dealloc_temp_buffers(&td->tpl_tmp_buffers); |
508 | | // This call ensures that the buffers in gm_data for MT encode are freed in |
509 | | // case of an error during gm. |
510 | 0 | gm_dealloc_data(&td->gm_data); |
511 | 0 | av1_dealloc_mb_data(&td->mb, num_planes); |
512 | 0 | aom_free(td->mb.sb_stats_cache); |
513 | 0 | td->mb.sb_stats_cache = NULL; |
514 | 0 | aom_free(td->mb.sb_fp_stats); |
515 | 0 | td->mb.sb_fp_stats = NULL; |
516 | | #if CONFIG_PARTITION_SEARCH_ORDER |
517 | | aom_free(td->mb.rdcost); |
518 | | td->mb.rdcost = NULL; |
519 | | #endif |
520 | 0 | av1_free_pc_tree_recursive(td->pc_root, num_planes, 0, 0, SEARCH_PARTITION); |
521 | 0 | td->pc_root = NULL; |
522 | 0 | av1_dealloc_mb_wiener_var_pred_buf(td); |
523 | 0 | aom_free(td); |
524 | 0 | thread_data->td = NULL; |
525 | 0 | thread_data->original_td = NULL; |
526 | 0 | } |
527 | 0 | } Unexecuted instantiation: av1_cx_iface.c:free_thread_data Unexecuted instantiation: encoder.c:free_thread_data Unexecuted instantiation: encoder_utils.c:free_thread_data Unexecuted instantiation: ethread.c:free_thread_data Unexecuted instantiation: superres_scale.c:free_thread_data Unexecuted instantiation: svc_layercontext.c:free_thread_data Unexecuted instantiation: compound_type.c:free_thread_data Unexecuted instantiation: encode_strategy.c:free_thread_data |
528 | | |
529 | | #ifdef __cplusplus |
530 | | } // extern "C" |
531 | | #endif |
532 | | |
533 | | #endif // AOM_AV1_ENCODER_ENCODER_ALLOC_H_ |