/src/ffmpeg/libavcodec/pthread_frame.c
Line | Count | Source |
1 | | /* |
2 | | * This file is part of FFmpeg. |
3 | | * |
4 | | * FFmpeg is free software; you can redistribute it and/or |
5 | | * modify it under the terms of the GNU Lesser General Public |
6 | | * License as published by the Free Software Foundation; either |
7 | | * version 2.1 of the License, or (at your option) any later version. |
8 | | * |
9 | | * FFmpeg is distributed in the hope that it will be useful, |
10 | | * but WITHOUT ANY WARRANTY; without even the implied warranty of |
11 | | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
12 | | * Lesser General Public License for more details. |
13 | | * |
14 | | * You should have received a copy of the GNU Lesser General Public |
15 | | * License along with FFmpeg; if not, write to the Free Software |
16 | | * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
17 | | */ |
18 | | |
19 | | /** |
20 | | * @file |
21 | | * Frame multithreading support functions |
22 | | * @see doc/multithreading.txt |
23 | | */ |
24 | | |
25 | | #include <stdatomic.h> |
26 | | |
27 | | #include "avcodec.h" |
28 | | #include "avcodec_internal.h" |
29 | | #include "codec_desc.h" |
30 | | #include "codec_internal.h" |
31 | | #include "decode.h" |
32 | | #include "hwaccel_internal.h" |
33 | | #include "hwconfig.h" |
34 | | #include "internal.h" |
35 | | #include "packet_internal.h" |
36 | | #include "pthread_internal.h" |
37 | | #include "libavutil/refstruct.h" |
38 | | #include "thread.h" |
39 | | #include "threadframe.h" |
40 | | #include "version_major.h" |
41 | | |
42 | | #include "libavutil/avassert.h" |
43 | | #include "libavutil/buffer.h" |
44 | | #include "libavutil/common.h" |
45 | | #include "libavutil/cpu.h" |
46 | | #include "libavutil/frame.h" |
47 | | #include "libavutil/internal.h" |
48 | | #include "libavutil/log.h" |
49 | | #include "libavutil/mem.h" |
50 | | #include "libavutil/opt.h" |
51 | | #include "libavutil/thread.h" |
52 | | |
53 | | enum { |
54 | | /// Set when the thread is awaiting a packet. |
55 | | STATE_INPUT_READY, |
56 | | /// Set before the codec has called ff_thread_finish_setup(). |
57 | | STATE_SETTING_UP, |
58 | | /// Set after the codec has called ff_thread_finish_setup(). |
59 | | STATE_SETUP_FINISHED, |
60 | | }; |
61 | | |
62 | | enum { |
63 | | UNINITIALIZED, ///< Thread has not been created, AVCodec->close mustn't be called |
64 | | NEEDS_CLOSE, ///< FFCodec->close needs to be called |
65 | | INITIALIZED, ///< Thread has been properly set up |
66 | | }; |
67 | | |
68 | | typedef struct DecodedFrames { |
69 | | AVFrame **f; |
70 | | size_t nb_f; |
71 | | size_t nb_f_allocated; |
72 | | } DecodedFrames; |
73 | | |
74 | | typedef struct ThreadFrameProgress { |
75 | | atomic_int progress[2]; |
76 | | } ThreadFrameProgress; |
77 | | |
78 | | /** |
79 | | * Context used by codec threads and stored in their AVCodecInternal thread_ctx. |
80 | | */ |
81 | | typedef struct PerThreadContext { |
82 | | struct FrameThreadContext *parent; |
83 | | |
84 | | pthread_t thread; |
85 | | int thread_init; |
86 | | unsigned pthread_init_cnt;///< Number of successfully initialized mutexes/conditions |
87 | | pthread_cond_t input_cond; ///< Used to wait for a new packet from the main thread. |
88 | | pthread_cond_t progress_cond; ///< Used by child threads to wait for progress to change. |
89 | | pthread_cond_t output_cond; ///< Used by the main thread to wait for frames to finish. |
90 | | |
91 | | pthread_mutex_t mutex; ///< Mutex used to protect the contents of the PerThreadContext. |
92 | | pthread_mutex_t progress_mutex; ///< Mutex used to protect frame progress values and progress_cond. |
93 | | |
94 | | AVCodecContext *avctx; ///< Context used to decode packets passed to this thread. |
95 | | |
96 | | AVPacket *avpkt; ///< Input packet (for decoding) or output (for encoding). |
97 | | |
98 | | /** |
99 | | * Decoded frames from a single decode iteration. |
100 | | */ |
101 | | DecodedFrames df; |
102 | | int result; ///< The result of the last codec decode/encode() call. |
103 | | |
104 | | atomic_int state; |
105 | | |
106 | | int die; ///< Set when the thread should exit. |
107 | | |
108 | | int hwaccel_serializing; |
109 | | int async_serializing; |
110 | | |
111 | | // set to 1 in ff_thread_finish_setup() when a threadsafe hwaccel is used; |
112 | | // cannot check hwaccel caps directly, because |
113 | | // worked threads clear hwaccel state for thread-unsafe hwaccels |
114 | | // after each decode call |
115 | | int hwaccel_threadsafe; |
116 | | |
117 | | atomic_int debug_threads; ///< Set if the FF_DEBUG_THREADS option is set. |
118 | | } PerThreadContext; |
119 | | |
120 | | /** |
121 | | * Context stored in the client AVCodecInternal thread_ctx. |
122 | | */ |
123 | | typedef struct FrameThreadContext { |
124 | | PerThreadContext *threads; ///< The contexts for each thread. |
125 | | PerThreadContext *prev_thread; ///< The last thread submit_packet() was called on. |
126 | | |
127 | | unsigned pthread_init_cnt; ///< Number of successfully initialized mutexes/conditions |
128 | | pthread_mutex_t buffer_mutex; ///< Mutex used to protect get/release_buffer(). |
129 | | /** |
130 | | * This lock is used for ensuring threads run in serial when thread-unsafe |
131 | | * hwaccel is used. |
132 | | */ |
133 | | pthread_mutex_t hwaccel_mutex; |
134 | | pthread_mutex_t async_mutex; |
135 | | pthread_cond_t async_cond; |
136 | | int async_lock; |
137 | | |
138 | | DecodedFrames df; |
139 | | int result; |
140 | | |
141 | | /** |
142 | | * Packet to be submitted to the next thread for decoding. |
143 | | */ |
144 | | AVPacket *next_pkt; |
145 | | |
146 | | int next_decoding; ///< The next context to submit a packet to. |
147 | | int next_finished; ///< The next context to return output from. |
148 | | |
149 | | /* hwaccel state for thread-unsafe hwaccels is temporarily stored here in |
150 | | * order to transfer its ownership to the next decoding thread without the |
151 | | * need for extra synchronization */ |
152 | | const AVHWAccel *stash_hwaccel; |
153 | | void *stash_hwaccel_context; |
154 | | void *stash_hwaccel_priv; |
155 | | } FrameThreadContext; |
156 | | |
157 | | static int hwaccel_serial(const AVCodecContext *avctx) |
158 | 0 | { |
159 | 0 | return avctx->hwaccel && !(ffhwaccel(avctx->hwaccel)->caps_internal & HWACCEL_CAP_THREAD_SAFE); |
160 | 0 | } |
161 | | |
162 | | static void async_lock(FrameThreadContext *fctx) |
163 | 0 | { |
164 | 0 | pthread_mutex_lock(&fctx->async_mutex); |
165 | 0 | while (fctx->async_lock) |
166 | 0 | pthread_cond_wait(&fctx->async_cond, &fctx->async_mutex); |
167 | 0 | fctx->async_lock = 1; |
168 | 0 | pthread_mutex_unlock(&fctx->async_mutex); |
169 | 0 | } |
170 | | |
171 | | static void async_unlock(FrameThreadContext *fctx) |
172 | 0 | { |
173 | 0 | pthread_mutex_lock(&fctx->async_mutex); |
174 | 0 | av_assert0(fctx->async_lock); |
175 | 0 | fctx->async_lock = 0; |
176 | 0 | pthread_cond_broadcast(&fctx->async_cond); |
177 | 0 | pthread_mutex_unlock(&fctx->async_mutex); |
178 | 0 | } |
179 | | |
180 | | static void thread_set_name(PerThreadContext *p) |
181 | 0 | { |
182 | 0 | AVCodecContext *avctx = p->avctx; |
183 | 0 | int idx = p - p->parent->threads; |
184 | 0 | char name[16]; |
185 | |
|
186 | 0 | snprintf(name, sizeof(name), "av:%.7s:df%d", avctx->codec->name, idx); |
187 | |
|
188 | 0 | ff_thread_setname(name); |
189 | 0 | } |
190 | | |
191 | | // get a free frame to decode into |
192 | | static AVFrame *decoded_frames_get_free(DecodedFrames *df) |
193 | 0 | { |
194 | 0 | if (df->nb_f == df->nb_f_allocated) { |
195 | 0 | AVFrame **tmp = av_realloc_array(df->f, df->nb_f + 1, |
196 | 0 | sizeof(*df->f)); |
197 | 0 | if (!tmp) |
198 | 0 | return NULL; |
199 | 0 | df->f = tmp; |
200 | |
|
201 | 0 | df->f[df->nb_f] = av_frame_alloc(); |
202 | 0 | if (!df->f[df->nb_f]) |
203 | 0 | return NULL; |
204 | | |
205 | 0 | df->nb_f_allocated++; |
206 | 0 | } |
207 | | |
208 | 0 | av_assert0(!df->f[df->nb_f]->buf[0]); |
209 | | |
210 | 0 | return df->f[df->nb_f]; |
211 | 0 | } |
212 | | |
213 | | static void decoded_frames_pop(DecodedFrames *df, AVFrame *dst) |
214 | 0 | { |
215 | 0 | AVFrame *tmp_frame = df->f[0]; |
216 | 0 | av_frame_move_ref(dst, tmp_frame); |
217 | 0 | memmove(df->f, df->f + 1, (df->nb_f - 1) * sizeof(*df->f)); |
218 | 0 | df->f[--df->nb_f] = tmp_frame; |
219 | 0 | } |
220 | | |
221 | | static void decoded_frames_flush(DecodedFrames *df) |
222 | 0 | { |
223 | 0 | for (size_t i = 0; i < df->nb_f; i++) |
224 | 0 | av_frame_unref(df->f[i]); |
225 | 0 | df->nb_f = 0; |
226 | 0 | } |
227 | | |
228 | | static void decoded_frames_free(DecodedFrames *df) |
229 | 0 | { |
230 | 0 | for (size_t i = 0; i < df->nb_f_allocated; i++) |
231 | 0 | av_frame_free(&df->f[i]); |
232 | 0 | av_freep(&df->f); |
233 | 0 | df->nb_f = 0; |
234 | 0 | df->nb_f_allocated = 0; |
235 | 0 | } |
236 | | |
237 | | /** |
238 | | * Codec worker thread. |
239 | | * |
240 | | * Automatically calls ff_thread_finish_setup() if the codec does |
241 | | * not provide an update_thread_context method, or if the codec returns |
242 | | * before calling it. |
243 | | */ |
244 | | static attribute_align_arg void *frame_worker_thread(void *arg) |
245 | 0 | { |
246 | 0 | PerThreadContext *p = arg; |
247 | 0 | AVCodecContext *avctx = p->avctx; |
248 | 0 | const FFCodec *codec = ffcodec(avctx->codec); |
249 | |
|
250 | 0 | thread_set_name(p); |
251 | |
|
252 | 0 | pthread_mutex_lock(&p->mutex); |
253 | 0 | while (1) { |
254 | 0 | int ret; |
255 | |
|
256 | 0 | while (atomic_load(&p->state) == STATE_INPUT_READY && !p->die) |
257 | 0 | pthread_cond_wait(&p->input_cond, &p->mutex); |
258 | |
|
259 | 0 | if (p->die) break; |
260 | | |
261 | 0 | if (!codec->update_thread_context) |
262 | 0 | ff_thread_finish_setup(avctx); |
263 | | |
264 | | /* If a decoder supports hwaccel, then it must call ff_get_format(). |
265 | | * Since that call must happen before ff_thread_finish_setup(), the |
266 | | * decoder is required to implement update_thread_context() and call |
267 | | * ff_thread_finish_setup() manually. Therefore the above |
268 | | * ff_thread_finish_setup() call did not happen and hwaccel_serializing |
269 | | * cannot be true here. */ |
270 | 0 | av_assert0(!p->hwaccel_serializing); |
271 | | |
272 | | /* if the previous thread uses thread-unsafe hwaccel then we take the |
273 | | * lock to ensure the threads don't run concurrently */ |
274 | 0 | if (hwaccel_serial(avctx)) { |
275 | 0 | pthread_mutex_lock(&p->parent->hwaccel_mutex); |
276 | 0 | p->hwaccel_serializing = 1; |
277 | 0 | } |
278 | |
|
279 | 0 | ret = 0; |
280 | 0 | while (ret >= 0) { |
281 | 0 | AVFrame *frame; |
282 | | |
283 | | /* get the frame which will store the output */ |
284 | 0 | frame = decoded_frames_get_free(&p->df); |
285 | 0 | if (!frame) { |
286 | 0 | p->result = AVERROR(ENOMEM); |
287 | 0 | goto alloc_fail; |
288 | 0 | } |
289 | | |
290 | | /* do the actual decoding */ |
291 | 0 | ret = ff_decode_receive_frame_internal(avctx, frame); |
292 | 0 | if (ret == 0) |
293 | 0 | p->df.nb_f++; |
294 | 0 | else if (ret < 0 && frame->buf[0]) |
295 | 0 | av_frame_unref(frame); |
296 | |
|
297 | 0 | p->result = (ret == AVERROR(EAGAIN)) ? 0 : ret; |
298 | 0 | } |
299 | | |
300 | 0 | if (atomic_load(&p->state) == STATE_SETTING_UP) |
301 | 0 | ff_thread_finish_setup(avctx); |
302 | |
|
303 | 0 | alloc_fail: |
304 | 0 | if (p->hwaccel_serializing) { |
305 | | /* wipe hwaccel state for thread-unsafe hwaccels to avoid stale |
306 | | * pointers lying around; |
307 | | * the state was transferred to FrameThreadContext in |
308 | | * ff_thread_finish_setup(), so nothing is leaked */ |
309 | 0 | avctx->hwaccel = NULL; |
310 | 0 | avctx->hwaccel_context = NULL; |
311 | 0 | avctx->internal->hwaccel_priv_data = NULL; |
312 | |
|
313 | 0 | p->hwaccel_serializing = 0; |
314 | 0 | pthread_mutex_unlock(&p->parent->hwaccel_mutex); |
315 | 0 | } |
316 | 0 | av_assert0(!avctx->hwaccel || |
317 | 0 | (ffhwaccel(avctx->hwaccel)->caps_internal & HWACCEL_CAP_THREAD_SAFE)); |
318 | | |
319 | 0 | if (p->async_serializing) { |
320 | 0 | p->async_serializing = 0; |
321 | |
|
322 | 0 | async_unlock(p->parent); |
323 | 0 | } |
324 | |
|
325 | 0 | pthread_mutex_lock(&p->progress_mutex); |
326 | |
|
327 | 0 | atomic_store(&p->state, STATE_INPUT_READY); |
328 | |
|
329 | 0 | pthread_cond_broadcast(&p->progress_cond); |
330 | 0 | pthread_cond_signal(&p->output_cond); |
331 | 0 | pthread_mutex_unlock(&p->progress_mutex); |
332 | 0 | } |
333 | 0 | pthread_mutex_unlock(&p->mutex); |
334 | |
|
335 | 0 | return NULL; |
336 | 0 | } |
337 | | |
338 | | /** |
339 | | * Update the next thread's AVCodecContext with values from the reference thread's context. |
340 | | * |
341 | | * @param dst The destination context. |
342 | | * @param src The source context. |
343 | | * @param for_user 0 if the destination is a codec thread, 1 if the destination is the user's thread |
344 | | * @return 0 on success, negative error code on failure |
345 | | */ |
346 | | static int update_context_from_thread(AVCodecContext *dst, const AVCodecContext *src, int for_user) |
347 | 0 | { |
348 | 0 | const FFCodec *const codec = ffcodec(dst->codec); |
349 | 0 | int err = 0; |
350 | |
|
351 | 0 | if (dst != src && (for_user || codec->update_thread_context)) { |
352 | 0 | dst->time_base = src->time_base; |
353 | 0 | dst->framerate = src->framerate; |
354 | 0 | dst->width = src->width; |
355 | 0 | dst->height = src->height; |
356 | 0 | dst->pix_fmt = src->pix_fmt; |
357 | 0 | dst->sw_pix_fmt = src->sw_pix_fmt; |
358 | |
|
359 | 0 | dst->coded_width = src->coded_width; |
360 | 0 | dst->coded_height = src->coded_height; |
361 | |
|
362 | 0 | dst->has_b_frames = src->has_b_frames; |
363 | 0 | dst->idct_algo = src->idct_algo; |
364 | 0 | #if FF_API_CODEC_PROPS |
365 | 0 | FF_DISABLE_DEPRECATION_WARNINGS |
366 | 0 | dst->properties = src->properties; |
367 | 0 | FF_ENABLE_DEPRECATION_WARNINGS |
368 | 0 | #endif |
369 | |
|
370 | 0 | dst->bits_per_coded_sample = src->bits_per_coded_sample; |
371 | 0 | dst->sample_aspect_ratio = src->sample_aspect_ratio; |
372 | |
|
373 | 0 | dst->profile = src->profile; |
374 | 0 | dst->level = src->level; |
375 | |
|
376 | 0 | dst->bits_per_raw_sample = src->bits_per_raw_sample; |
377 | 0 | dst->color_primaries = src->color_primaries; |
378 | |
|
379 | 0 | dst->alpha_mode = src->alpha_mode; |
380 | |
|
381 | 0 | dst->color_trc = src->color_trc; |
382 | 0 | dst->colorspace = src->colorspace; |
383 | 0 | dst->color_range = src->color_range; |
384 | 0 | dst->chroma_sample_location = src->chroma_sample_location; |
385 | |
|
386 | 0 | dst->sample_rate = src->sample_rate; |
387 | 0 | dst->sample_fmt = src->sample_fmt; |
388 | 0 | err = av_channel_layout_copy(&dst->ch_layout, &src->ch_layout); |
389 | 0 | if (err < 0) |
390 | 0 | return err; |
391 | | |
392 | 0 | if (!!dst->hw_frames_ctx != !!src->hw_frames_ctx || |
393 | 0 | (dst->hw_frames_ctx && dst->hw_frames_ctx->data != src->hw_frames_ctx->data)) { |
394 | 0 | av_buffer_unref(&dst->hw_frames_ctx); |
395 | |
|
396 | 0 | if (src->hw_frames_ctx) { |
397 | 0 | dst->hw_frames_ctx = av_buffer_ref(src->hw_frames_ctx); |
398 | 0 | if (!dst->hw_frames_ctx) |
399 | 0 | return AVERROR(ENOMEM); |
400 | 0 | } |
401 | 0 | } |
402 | | |
403 | 0 | dst->hwaccel_flags = src->hwaccel_flags; |
404 | |
|
405 | 0 | av_refstruct_replace(&dst->internal->pool, src->internal->pool); |
406 | 0 | ff_decode_internal_sync(dst, src); |
407 | 0 | } |
408 | | |
409 | 0 | if (for_user) { |
410 | 0 | if (codec->update_thread_context_for_user) |
411 | 0 | err = codec->update_thread_context_for_user(dst, src); |
412 | 0 | } else { |
413 | 0 | const PerThreadContext *p_src = src->internal->thread_ctx; |
414 | 0 | PerThreadContext *p_dst = dst->internal->thread_ctx; |
415 | |
|
416 | 0 | if (codec->update_thread_context) { |
417 | 0 | err = codec->update_thread_context(dst, src); |
418 | 0 | if (err < 0) |
419 | 0 | return err; |
420 | 0 | } |
421 | | |
422 | | // reset dst hwaccel state if needed |
423 | 0 | av_assert0(p_dst->hwaccel_threadsafe || |
424 | 0 | (!dst->hwaccel && !dst->internal->hwaccel_priv_data)); |
425 | 0 | if (p_dst->hwaccel_threadsafe && |
426 | 0 | (!p_src->hwaccel_threadsafe || dst->hwaccel != src->hwaccel)) { |
427 | 0 | ff_hwaccel_uninit(dst); |
428 | 0 | p_dst->hwaccel_threadsafe = 0; |
429 | 0 | } |
430 | | |
431 | | // propagate hwaccel state for threadsafe hwaccels |
432 | 0 | if (p_src->hwaccel_threadsafe) { |
433 | 0 | const FFHWAccel *hwaccel = ffhwaccel(src->hwaccel); |
434 | 0 | if (!dst->hwaccel) { |
435 | 0 | if (hwaccel->priv_data_size) { |
436 | 0 | av_assert0(hwaccel->update_thread_context); |
437 | | |
438 | 0 | dst->internal->hwaccel_priv_data = |
439 | 0 | av_mallocz(hwaccel->priv_data_size); |
440 | 0 | if (!dst->internal->hwaccel_priv_data) |
441 | 0 | return AVERROR(ENOMEM); |
442 | 0 | } |
443 | 0 | dst->hwaccel = src->hwaccel; |
444 | 0 | } |
445 | 0 | av_assert0(dst->hwaccel == src->hwaccel); |
446 | | |
447 | 0 | if (hwaccel->update_thread_context) { |
448 | 0 | err = hwaccel->update_thread_context(dst, src); |
449 | 0 | if (err < 0) { |
450 | 0 | av_log(dst, AV_LOG_ERROR, "Error propagating hwaccel state\n"); |
451 | 0 | ff_hwaccel_uninit(dst); |
452 | 0 | return err; |
453 | 0 | } |
454 | 0 | } |
455 | 0 | p_dst->hwaccel_threadsafe = 1; |
456 | 0 | } |
457 | 0 | } |
458 | | |
459 | 0 | return err; |
460 | 0 | } |
461 | | |
462 | | /** |
463 | | * Update the next thread's AVCodecContext with values set by the user. |
464 | | * |
465 | | * @param dst The destination context. |
466 | | * @param src The source context. |
467 | | * @return 0 on success, negative error code on failure |
468 | | */ |
469 | | static int update_context_from_user(AVCodecContext *dst, const AVCodecContext *src) |
470 | 0 | { |
471 | 0 | int err; |
472 | |
|
473 | 0 | dst->flags = src->flags; |
474 | |
|
475 | 0 | dst->draw_horiz_band= src->draw_horiz_band; |
476 | 0 | dst->get_buffer2 = src->get_buffer2; |
477 | |
|
478 | 0 | dst->opaque = src->opaque; |
479 | 0 | dst->debug = src->debug; |
480 | |
|
481 | 0 | dst->slice_flags = src->slice_flags; |
482 | 0 | dst->flags2 = src->flags2; |
483 | 0 | dst->export_side_data = src->export_side_data; |
484 | |
|
485 | 0 | dst->skip_loop_filter = src->skip_loop_filter; |
486 | 0 | dst->skip_idct = src->skip_idct; |
487 | 0 | dst->skip_frame = src->skip_frame; |
488 | |
|
489 | 0 | dst->frame_num = src->frame_num; |
490 | |
|
491 | 0 | av_packet_unref(dst->internal->last_pkt_props); |
492 | 0 | err = av_packet_copy_props(dst->internal->last_pkt_props, src->internal->last_pkt_props); |
493 | 0 | if (err < 0) |
494 | 0 | return err; |
495 | | |
496 | 0 | return 0; |
497 | 0 | } |
498 | | |
499 | | static int submit_packet(PerThreadContext *p, AVCodecContext *user_avctx, |
500 | | AVPacket *in_pkt) |
501 | 0 | { |
502 | 0 | FrameThreadContext *fctx = p->parent; |
503 | 0 | PerThreadContext *prev_thread = fctx->prev_thread; |
504 | 0 | const AVCodec *codec = p->avctx->codec; |
505 | 0 | int ret; |
506 | |
|
507 | 0 | pthread_mutex_lock(&p->mutex); |
508 | |
|
509 | 0 | av_packet_unref(p->avpkt); |
510 | 0 | av_packet_move_ref(p->avpkt, in_pkt); |
511 | |
|
512 | 0 | if (AVPACKET_IS_EMPTY(p->avpkt)) |
513 | 0 | p->avctx->internal->draining = 1; |
514 | |
|
515 | 0 | ret = update_context_from_user(p->avctx, user_avctx); |
516 | 0 | if (ret) { |
517 | 0 | pthread_mutex_unlock(&p->mutex); |
518 | 0 | return ret; |
519 | 0 | } |
520 | 0 | atomic_store_explicit(&p->debug_threads, |
521 | 0 | (p->avctx->debug & FF_DEBUG_THREADS) != 0, |
522 | 0 | memory_order_relaxed); |
523 | |
|
524 | 0 | if (prev_thread) { |
525 | 0 | if (atomic_load(&prev_thread->state) == STATE_SETTING_UP) { |
526 | 0 | pthread_mutex_lock(&prev_thread->progress_mutex); |
527 | 0 | while (atomic_load(&prev_thread->state) == STATE_SETTING_UP) |
528 | 0 | pthread_cond_wait(&prev_thread->progress_cond, &prev_thread->progress_mutex); |
529 | 0 | pthread_mutex_unlock(&prev_thread->progress_mutex); |
530 | 0 | } |
531 | | |
532 | | /* codecs without delay might not be prepared to be called repeatedly here during |
533 | | * flushing (vp3/theora), and also don't need to be, since from this point on, they |
534 | | * will always return EOF anyway */ |
535 | 0 | if (!p->avctx->internal->draining || |
536 | 0 | (codec->capabilities & AV_CODEC_CAP_DELAY)) { |
537 | 0 | ret = update_context_from_thread(p->avctx, prev_thread->avctx, 0); |
538 | 0 | if (ret) { |
539 | 0 | pthread_mutex_unlock(&p->mutex); |
540 | 0 | return ret; |
541 | 0 | } |
542 | 0 | } |
543 | 0 | } |
544 | | |
545 | | /* transfer the stashed hwaccel state, if any */ |
546 | 0 | av_assert0(!p->avctx->hwaccel || p->hwaccel_threadsafe); |
547 | 0 | if (!p->hwaccel_threadsafe) { |
548 | 0 | FFSWAP(const AVHWAccel*, p->avctx->hwaccel, fctx->stash_hwaccel); |
549 | 0 | FFSWAP(void*, p->avctx->hwaccel_context, fctx->stash_hwaccel_context); |
550 | 0 | FFSWAP(void*, p->avctx->internal->hwaccel_priv_data, fctx->stash_hwaccel_priv); |
551 | 0 | } |
552 | |
|
553 | 0 | atomic_store(&p->state, STATE_SETTING_UP); |
554 | 0 | pthread_cond_signal(&p->input_cond); |
555 | 0 | pthread_mutex_unlock(&p->mutex); |
556 | |
|
557 | 0 | fctx->prev_thread = p; |
558 | 0 | fctx->next_decoding = (fctx->next_decoding + 1) % p->avctx->thread_count; |
559 | |
|
560 | 0 | return 0; |
561 | 0 | } |
562 | | |
563 | | int ff_thread_receive_frame(AVCodecContext *avctx, AVFrame *frame, unsigned flags) |
564 | 0 | { |
565 | 0 | FrameThreadContext *fctx = avctx->internal->thread_ctx; |
566 | 0 | int ret = 0; |
567 | | |
568 | | /* release the async lock, permitting blocked hwaccel threads to |
569 | | * go forward while we are in this function */ |
570 | 0 | async_unlock(fctx); |
571 | | |
572 | | /* submit packets to threads while there are no buffered results to return */ |
573 | 0 | while (!fctx->df.nb_f && !fctx->result) { |
574 | 0 | PerThreadContext *p; |
575 | |
|
576 | 0 | if (fctx->next_decoding != fctx->next_finished && |
577 | 0 | (flags & AV_CODEC_RECEIVE_FRAME_FLAG_SYNCHRONOUS)) |
578 | 0 | goto wait_for_result; |
579 | | |
580 | | /* get a packet to be submitted to the next thread */ |
581 | 0 | av_packet_unref(fctx->next_pkt); |
582 | 0 | ret = ff_decode_get_packet(avctx, fctx->next_pkt); |
583 | 0 | if (ret < 0 && ret != AVERROR_EOF) |
584 | 0 | goto finish; |
585 | | |
586 | 0 | ret = submit_packet(&fctx->threads[fctx->next_decoding], avctx, |
587 | 0 | fctx->next_pkt); |
588 | 0 | if (ret < 0) |
589 | 0 | goto finish; |
590 | | |
591 | | /* do not return any frames until all threads have something to do */ |
592 | 0 | if (fctx->next_decoding != fctx->next_finished && |
593 | 0 | !avctx->internal->draining) |
594 | 0 | continue; |
595 | | |
596 | 0 | wait_for_result: |
597 | 0 | p = &fctx->threads[fctx->next_finished]; |
598 | 0 | fctx->next_finished = (fctx->next_finished + 1) % avctx->thread_count; |
599 | |
|
600 | 0 | if (atomic_load(&p->state) != STATE_INPUT_READY) { |
601 | 0 | pthread_mutex_lock(&p->progress_mutex); |
602 | 0 | while (atomic_load_explicit(&p->state, memory_order_relaxed) != STATE_INPUT_READY) |
603 | 0 | pthread_cond_wait(&p->output_cond, &p->progress_mutex); |
604 | 0 | pthread_mutex_unlock(&p->progress_mutex); |
605 | 0 | } |
606 | |
|
607 | 0 | update_context_from_thread(avctx, p->avctx, 1); |
608 | 0 | fctx->result = p->result; |
609 | 0 | p->result = 0; |
610 | 0 | if (p->df.nb_f) |
611 | 0 | FFSWAP(DecodedFrames, fctx->df, p->df); |
612 | 0 | } |
613 | | |
614 | | /* a thread may return multiple frames AND an error |
615 | | * we first return all the frames, then the error */ |
616 | 0 | if (fctx->df.nb_f) { |
617 | 0 | decoded_frames_pop(&fctx->df, frame); |
618 | 0 | ret = 0; |
619 | 0 | } else { |
620 | 0 | ret = fctx->result; |
621 | 0 | fctx->result = 0; |
622 | 0 | } |
623 | |
|
624 | 0 | finish: |
625 | 0 | async_lock(fctx); |
626 | 0 | return ret; |
627 | 0 | } |
628 | | |
629 | | void ff_thread_report_progress(ThreadFrame *f, int n, int field) |
630 | 0 | { |
631 | 0 | PerThreadContext *p; |
632 | 0 | atomic_int *progress = f->progress ? f->progress->progress : NULL; |
633 | |
|
634 | 0 | if (!progress || |
635 | 0 | atomic_load_explicit(&progress[field], memory_order_relaxed) >= n) |
636 | 0 | return; |
637 | | |
638 | 0 | p = f->owner[field]->internal->thread_ctx; |
639 | |
|
640 | 0 | if (atomic_load_explicit(&p->debug_threads, memory_order_relaxed)) |
641 | 0 | av_log(f->owner[field], AV_LOG_DEBUG, |
642 | 0 | "%p finished %d field %d\n", progress, n, field); |
643 | |
|
644 | 0 | pthread_mutex_lock(&p->progress_mutex); |
645 | |
|
646 | 0 | atomic_store_explicit(&progress[field], n, memory_order_release); |
647 | |
|
648 | 0 | pthread_cond_broadcast(&p->progress_cond); |
649 | 0 | pthread_mutex_unlock(&p->progress_mutex); |
650 | 0 | } |
651 | | |
652 | | void ff_thread_await_progress(const ThreadFrame *f, int n, int field) |
653 | 0 | { |
654 | 0 | PerThreadContext *p; |
655 | 0 | atomic_int *progress = f->progress ? f->progress->progress : NULL; |
656 | |
|
657 | 0 | if (!progress || |
658 | 0 | atomic_load_explicit(&progress[field], memory_order_acquire) >= n) |
659 | 0 | return; |
660 | | |
661 | 0 | p = f->owner[field]->internal->thread_ctx; |
662 | |
|
663 | 0 | if (atomic_load_explicit(&p->debug_threads, memory_order_relaxed)) |
664 | 0 | av_log(f->owner[field], AV_LOG_DEBUG, |
665 | 0 | "thread awaiting %d field %d from %p\n", n, field, progress); |
666 | |
|
667 | 0 | pthread_mutex_lock(&p->progress_mutex); |
668 | 0 | while (atomic_load_explicit(&progress[field], memory_order_relaxed) < n) |
669 | 0 | pthread_cond_wait(&p->progress_cond, &p->progress_mutex); |
670 | 0 | pthread_mutex_unlock(&p->progress_mutex); |
671 | 0 | } |
672 | | |
673 | 0 | void ff_thread_finish_setup(AVCodecContext *avctx) { |
674 | 0 | PerThreadContext *p; |
675 | |
|
676 | 0 | if (!(avctx->active_thread_type&FF_THREAD_FRAME)) return; |
677 | | |
678 | 0 | p = avctx->internal->thread_ctx; |
679 | |
|
680 | 0 | p->hwaccel_threadsafe = avctx->hwaccel && |
681 | 0 | (ffhwaccel(avctx->hwaccel)->caps_internal & HWACCEL_CAP_THREAD_SAFE); |
682 | |
|
683 | 0 | if (hwaccel_serial(avctx) && !p->hwaccel_serializing) { |
684 | 0 | pthread_mutex_lock(&p->parent->hwaccel_mutex); |
685 | 0 | p->hwaccel_serializing = 1; |
686 | 0 | } |
687 | | |
688 | | /* this assumes that no hwaccel calls happen before ff_thread_finish_setup() */ |
689 | 0 | if (avctx->hwaccel && |
690 | 0 | !(ffhwaccel(avctx->hwaccel)->caps_internal & HWACCEL_CAP_ASYNC_SAFE)) { |
691 | 0 | p->async_serializing = 1; |
692 | |
|
693 | 0 | async_lock(p->parent); |
694 | 0 | } |
695 | | |
696 | | /* thread-unsafe hwaccels share a single private data instance, so we |
697 | | * save hwaccel state for passing to the next thread; |
698 | | * this is done here so that this worker thread can wipe its own hwaccel |
699 | | * state after decoding, without requiring synchronization */ |
700 | 0 | av_assert0(!p->parent->stash_hwaccel); |
701 | 0 | if (hwaccel_serial(avctx)) { |
702 | 0 | p->parent->stash_hwaccel = avctx->hwaccel; |
703 | 0 | p->parent->stash_hwaccel_context = avctx->hwaccel_context; |
704 | 0 | p->parent->stash_hwaccel_priv = avctx->internal->hwaccel_priv_data; |
705 | 0 | } |
706 | |
|
707 | 0 | pthread_mutex_lock(&p->progress_mutex); |
708 | 0 | if(atomic_load(&p->state) == STATE_SETUP_FINISHED){ |
709 | 0 | av_log(avctx, AV_LOG_WARNING, "Multiple ff_thread_finish_setup() calls\n"); |
710 | 0 | } |
711 | |
|
712 | 0 | atomic_store(&p->state, STATE_SETUP_FINISHED); |
713 | |
|
714 | 0 | pthread_cond_broadcast(&p->progress_cond); |
715 | 0 | pthread_mutex_unlock(&p->progress_mutex); |
716 | 0 | } |
717 | | |
718 | | /// Waits for all threads to finish. |
719 | | static av_cold void park_frame_worker_threads(FrameThreadContext *fctx, int thread_count) |
720 | 0 | { |
721 | 0 | int i; |
722 | |
|
723 | 0 | async_unlock(fctx); |
724 | |
|
725 | 0 | for (i = 0; i < thread_count; i++) { |
726 | 0 | PerThreadContext *p = &fctx->threads[i]; |
727 | |
|
728 | 0 | if (atomic_load(&p->state) != STATE_INPUT_READY) { |
729 | 0 | pthread_mutex_lock(&p->progress_mutex); |
730 | 0 | while (atomic_load(&p->state) != STATE_INPUT_READY) |
731 | 0 | pthread_cond_wait(&p->output_cond, &p->progress_mutex); |
732 | 0 | pthread_mutex_unlock(&p->progress_mutex); |
733 | 0 | } |
734 | 0 | } |
735 | |
|
736 | 0 | async_lock(fctx); |
737 | 0 | } |
738 | | |
739 | | #define OFF(member) offsetof(FrameThreadContext, member) |
740 | | DEFINE_OFFSET_ARRAY(FrameThreadContext, thread_ctx, pthread_init_cnt, |
741 | | (OFF(buffer_mutex), OFF(hwaccel_mutex), OFF(async_mutex)), |
742 | | (OFF(async_cond))); |
743 | | #undef OFF |
744 | | |
745 | | #define OFF(member) offsetof(PerThreadContext, member) |
746 | | DEFINE_OFFSET_ARRAY(PerThreadContext, per_thread, pthread_init_cnt, |
747 | | (OFF(progress_mutex), OFF(mutex)), |
748 | | (OFF(input_cond), OFF(progress_cond), OFF(output_cond))); |
749 | | #undef OFF |
750 | | |
751 | | av_cold void ff_frame_thread_free(AVCodecContext *avctx, int thread_count) |
752 | 0 | { |
753 | 0 | FrameThreadContext *fctx = avctx->internal->thread_ctx; |
754 | 0 | const FFCodec *codec = ffcodec(avctx->codec); |
755 | 0 | int i; |
756 | |
|
757 | 0 | park_frame_worker_threads(fctx, thread_count); |
758 | |
|
759 | 0 | for (i = 0; i < thread_count; i++) { |
760 | 0 | PerThreadContext *p = &fctx->threads[i]; |
761 | 0 | AVCodecContext *ctx = p->avctx; |
762 | |
|
763 | 0 | if (ctx->internal) { |
764 | 0 | if (p->thread_init == INITIALIZED) { |
765 | 0 | pthread_mutex_lock(&p->mutex); |
766 | 0 | p->die = 1; |
767 | 0 | pthread_cond_signal(&p->input_cond); |
768 | 0 | pthread_mutex_unlock(&p->mutex); |
769 | |
|
770 | 0 | pthread_join(p->thread, NULL); |
771 | 0 | } |
772 | 0 | if (codec->close && p->thread_init != UNINITIALIZED) |
773 | 0 | codec->close(ctx); |
774 | | |
775 | | /* When using a threadsafe hwaccel, this is where |
776 | | * each thread's context is uninit'd and freed. */ |
777 | 0 | ff_hwaccel_uninit(ctx); |
778 | |
|
779 | 0 | if (ctx->priv_data) { |
780 | 0 | if (codec->p.priv_class) |
781 | 0 | av_opt_free(ctx->priv_data); |
782 | 0 | av_freep(&ctx->priv_data); |
783 | 0 | } |
784 | |
|
785 | 0 | av_refstruct_unref(&ctx->internal->pool); |
786 | 0 | av_packet_free(&ctx->internal->in_pkt); |
787 | 0 | av_packet_free(&ctx->internal->last_pkt_props); |
788 | 0 | ff_decode_internal_uninit(ctx); |
789 | 0 | av_freep(&ctx->internal); |
790 | 0 | av_buffer_unref(&ctx->hw_frames_ctx); |
791 | 0 | av_frame_side_data_free(&ctx->decoded_side_data, |
792 | 0 | &ctx->nb_decoded_side_data); |
793 | 0 | } |
794 | |
|
795 | 0 | decoded_frames_free(&p->df); |
796 | |
|
797 | 0 | ff_pthread_free(p, per_thread_offsets); |
798 | 0 | av_packet_free(&p->avpkt); |
799 | |
|
800 | 0 | av_freep(&p->avctx); |
801 | 0 | } |
802 | |
|
803 | 0 | decoded_frames_free(&fctx->df); |
804 | 0 | av_packet_free(&fctx->next_pkt); |
805 | |
|
806 | 0 | av_freep(&fctx->threads); |
807 | 0 | ff_pthread_free(fctx, thread_ctx_offsets); |
808 | | |
809 | | /* if we have stashed hwaccel state, move it to the user-facing context, |
810 | | * so it will be freed in ff_codec_close() */ |
811 | 0 | av_assert0(!avctx->hwaccel); |
812 | 0 | FFSWAP(const AVHWAccel*, avctx->hwaccel, fctx->stash_hwaccel); |
813 | 0 | FFSWAP(void*, avctx->hwaccel_context, fctx->stash_hwaccel_context); |
814 | 0 | FFSWAP(void*, avctx->internal->hwaccel_priv_data, fctx->stash_hwaccel_priv); |
815 | |
|
816 | 0 | av_freep(&avctx->internal->thread_ctx); |
817 | 0 | } |
818 | | |
819 | | static av_cold int init_thread(PerThreadContext *p, int *threads_to_free, |
820 | | FrameThreadContext *fctx, AVCodecContext *avctx, |
821 | | const FFCodec *codec, int first) |
822 | 0 | { |
823 | 0 | AVCodecContext *copy; |
824 | 0 | int err; |
825 | |
|
826 | 0 | atomic_init(&p->state, STATE_INPUT_READY); |
827 | |
|
828 | 0 | copy = av_memdup(avctx, sizeof(*avctx)); |
829 | 0 | if (!copy) |
830 | 0 | return AVERROR(ENOMEM); |
831 | 0 | copy->priv_data = NULL; |
832 | 0 | copy->decoded_side_data = NULL; |
833 | 0 | copy->nb_decoded_side_data = 0; |
834 | | |
835 | | /* From now on, this PerThreadContext will be cleaned up by |
836 | | * ff_frame_thread_free in case of errors. */ |
837 | 0 | (*threads_to_free)++; |
838 | |
|
839 | 0 | p->parent = fctx; |
840 | 0 | p->avctx = copy; |
841 | |
|
842 | 0 | copy->internal = ff_decode_internal_alloc(); |
843 | 0 | if (!copy->internal) |
844 | 0 | return AVERROR(ENOMEM); |
845 | 0 | ff_decode_internal_sync(copy, avctx); |
846 | 0 | copy->internal->thread_ctx = p; |
847 | 0 | copy->internal->progress_frame_pool = avctx->internal->progress_frame_pool; |
848 | |
|
849 | 0 | copy->delay = avctx->delay; |
850 | |
|
851 | 0 | if (codec->priv_data_size) { |
852 | 0 | copy->priv_data = av_mallocz(codec->priv_data_size); |
853 | 0 | if (!copy->priv_data) |
854 | 0 | return AVERROR(ENOMEM); |
855 | | |
856 | 0 | if (codec->p.priv_class) { |
857 | 0 | *(const AVClass **)copy->priv_data = codec->p.priv_class; |
858 | 0 | err = av_opt_copy(copy->priv_data, avctx->priv_data); |
859 | 0 | if (err < 0) |
860 | 0 | return err; |
861 | 0 | } |
862 | 0 | } |
863 | | |
864 | 0 | err = ff_pthread_init(p, per_thread_offsets); |
865 | 0 | if (err < 0) |
866 | 0 | return err; |
867 | | |
868 | 0 | if (!(p->avpkt = av_packet_alloc())) |
869 | 0 | return AVERROR(ENOMEM); |
870 | | |
871 | 0 | copy->internal->is_frame_mt = 1; |
872 | 0 | if (!first) |
873 | 0 | copy->internal->is_copy = 1; |
874 | |
|
875 | 0 | copy->internal->in_pkt = av_packet_alloc(); |
876 | 0 | if (!copy->internal->in_pkt) |
877 | 0 | return AVERROR(ENOMEM); |
878 | | |
879 | 0 | copy->internal->last_pkt_props = av_packet_alloc(); |
880 | 0 | if (!copy->internal->last_pkt_props) |
881 | 0 | return AVERROR(ENOMEM); |
882 | | |
883 | 0 | if (codec->init) { |
884 | 0 | err = codec->init(copy); |
885 | 0 | if (err < 0) { |
886 | 0 | if (codec->caps_internal & FF_CODEC_CAP_INIT_CLEANUP) |
887 | 0 | p->thread_init = NEEDS_CLOSE; |
888 | 0 | return err; |
889 | 0 | } |
890 | 0 | } |
891 | 0 | p->thread_init = NEEDS_CLOSE; |
892 | |
|
893 | 0 | if (first) { |
894 | 0 | update_context_from_thread(avctx, copy, 1); |
895 | |
|
896 | 0 | av_frame_side_data_free(&avctx->decoded_side_data, &avctx->nb_decoded_side_data); |
897 | 0 | for (int i = 0; i < copy->nb_decoded_side_data; i++) { |
898 | 0 | err = av_frame_side_data_clone(&avctx->decoded_side_data, |
899 | 0 | &avctx->nb_decoded_side_data, |
900 | 0 | copy->decoded_side_data[i], 0); |
901 | 0 | if (err < 0) |
902 | 0 | return err; |
903 | 0 | } |
904 | 0 | } |
905 | | |
906 | 0 | atomic_init(&p->debug_threads, (copy->debug & FF_DEBUG_THREADS) != 0); |
907 | |
|
908 | 0 | err = AVERROR(pthread_create(&p->thread, NULL, frame_worker_thread, p)); |
909 | 0 | if (err < 0) |
910 | 0 | return err; |
911 | 0 | p->thread_init = INITIALIZED; |
912 | |
|
913 | 0 | return 0; |
914 | 0 | } |
915 | | |
916 | | av_cold int ff_frame_thread_init(AVCodecContext *avctx) |
917 | 0 | { |
918 | 0 | int thread_count = avctx->thread_count; |
919 | 0 | const FFCodec *codec = ffcodec(avctx->codec); |
920 | 0 | FrameThreadContext *fctx; |
921 | 0 | int err, i = 0; |
922 | |
|
923 | 0 | if (!thread_count) { |
924 | 0 | int nb_cpus = av_cpu_count(); |
925 | | // use number of cores + 1 as thread count if there is more than one |
926 | 0 | if (nb_cpus > 1) |
927 | 0 | thread_count = avctx->thread_count = FFMIN(nb_cpus + 1, MAX_AUTO_THREADS); |
928 | 0 | else |
929 | 0 | thread_count = avctx->thread_count = 1; |
930 | 0 | } |
931 | |
|
932 | 0 | if (thread_count <= 1) { |
933 | 0 | avctx->active_thread_type = 0; |
934 | 0 | return 0; |
935 | 0 | } |
936 | | |
937 | 0 | avctx->internal->thread_ctx = fctx = av_mallocz(sizeof(FrameThreadContext)); |
938 | 0 | if (!fctx) |
939 | 0 | return AVERROR(ENOMEM); |
940 | | |
941 | 0 | err = ff_pthread_init(fctx, thread_ctx_offsets); |
942 | 0 | if (err < 0) { |
943 | 0 | ff_pthread_free(fctx, thread_ctx_offsets); |
944 | 0 | av_freep(&avctx->internal->thread_ctx); |
945 | 0 | return err; |
946 | 0 | } |
947 | | |
948 | 0 | fctx->next_pkt = av_packet_alloc(); |
949 | 0 | if (!fctx->next_pkt) |
950 | 0 | return AVERROR(ENOMEM); |
951 | | |
952 | 0 | fctx->async_lock = 1; |
953 | |
|
954 | 0 | if (codec->p.type == AVMEDIA_TYPE_VIDEO) |
955 | 0 | avctx->delay = avctx->thread_count - 1; |
956 | |
|
957 | 0 | fctx->threads = av_calloc(thread_count, sizeof(*fctx->threads)); |
958 | 0 | if (!fctx->threads) { |
959 | 0 | err = AVERROR(ENOMEM); |
960 | 0 | goto error; |
961 | 0 | } |
962 | | |
963 | 0 | for (; i < thread_count; ) { |
964 | 0 | PerThreadContext *p = &fctx->threads[i]; |
965 | 0 | int first = !i; |
966 | |
|
967 | 0 | err = init_thread(p, &i, fctx, avctx, codec, first); |
968 | 0 | if (err < 0) |
969 | 0 | goto error; |
970 | 0 | } |
971 | | |
972 | 0 | return 0; |
973 | | |
974 | 0 | error: |
975 | 0 | ff_frame_thread_free(avctx, i); |
976 | 0 | return err; |
977 | 0 | } |
978 | | |
979 | | av_cold void ff_thread_flush(AVCodecContext *avctx) |
980 | 0 | { |
981 | 0 | int i; |
982 | 0 | FrameThreadContext *fctx = avctx->internal->thread_ctx; |
983 | |
|
984 | 0 | if (!fctx) return; |
985 | | |
986 | 0 | park_frame_worker_threads(fctx, avctx->thread_count); |
987 | 0 | if (fctx->prev_thread) { |
988 | 0 | if (fctx->prev_thread != &fctx->threads[0]) |
989 | 0 | update_context_from_thread(fctx->threads[0].avctx, fctx->prev_thread->avctx, 0); |
990 | 0 | } |
991 | |
|
992 | 0 | fctx->next_decoding = fctx->next_finished = 0; |
993 | 0 | fctx->prev_thread = NULL; |
994 | |
|
995 | 0 | decoded_frames_flush(&fctx->df); |
996 | 0 | fctx->result = 0; |
997 | |
|
998 | 0 | for (i = 0; i < avctx->thread_count; i++) { |
999 | 0 | PerThreadContext *p = &fctx->threads[i]; |
1000 | |
|
1001 | 0 | decoded_frames_flush(&p->df); |
1002 | 0 | p->result = 0; |
1003 | |
|
1004 | 0 | avcodec_flush_buffers(p->avctx); |
1005 | 0 | } |
1006 | 0 | } |
1007 | | |
1008 | | int ff_thread_can_start_frame(AVCodecContext *avctx) |
1009 | 0 | { |
1010 | 0 | if ((avctx->active_thread_type & FF_THREAD_FRAME) && |
1011 | 0 | ffcodec(avctx->codec)->update_thread_context) { |
1012 | 0 | PerThreadContext *p = avctx->internal->thread_ctx; |
1013 | |
|
1014 | 0 | if (atomic_load(&p->state) != STATE_SETTING_UP) |
1015 | 0 | return 0; |
1016 | 0 | } |
1017 | | |
1018 | 0 | return 1; |
1019 | 0 | } |
1020 | | |
1021 | | static int thread_get_buffer_internal(AVCodecContext *avctx, AVFrame *f, int flags) |
1022 | 0 | { |
1023 | 0 | PerThreadContext *p; |
1024 | 0 | int err; |
1025 | |
|
1026 | 0 | if (!(avctx->active_thread_type & FF_THREAD_FRAME)) |
1027 | 0 | return ff_get_buffer(avctx, f, flags); |
1028 | | |
1029 | 0 | p = avctx->internal->thread_ctx; |
1030 | 0 | if (atomic_load(&p->state) != STATE_SETTING_UP && |
1031 | 0 | ffcodec(avctx->codec)->update_thread_context) { |
1032 | 0 | av_log(avctx, AV_LOG_ERROR, "get_buffer() cannot be called after ff_thread_finish_setup()\n"); |
1033 | 0 | return -1; |
1034 | 0 | } |
1035 | | |
1036 | 0 | pthread_mutex_lock(&p->parent->buffer_mutex); |
1037 | 0 | err = ff_get_buffer(avctx, f, flags); |
1038 | |
|
1039 | 0 | pthread_mutex_unlock(&p->parent->buffer_mutex); |
1040 | |
|
1041 | 0 | return err; |
1042 | 0 | } |
1043 | | |
1044 | | int ff_thread_get_buffer(AVCodecContext *avctx, AVFrame *f, int flags) |
1045 | 0 | { |
1046 | 0 | int ret = thread_get_buffer_internal(avctx, f, flags); |
1047 | 0 | if (ret < 0) |
1048 | 0 | av_log(avctx, AV_LOG_ERROR, "thread_get_buffer() failed\n"); |
1049 | 0 | return ret; |
1050 | 0 | } |
1051 | | |
1052 | | int ff_thread_get_ext_buffer(AVCodecContext *avctx, ThreadFrame *f, int flags) |
1053 | 0 | { |
1054 | 0 | int ret; |
1055 | |
|
1056 | 0 | f->owner[0] = f->owner[1] = avctx; |
1057 | 0 | if (!(avctx->active_thread_type & FF_THREAD_FRAME)) |
1058 | 0 | return ff_get_buffer(avctx, f->f, flags); |
1059 | | |
1060 | 0 | f->progress = av_refstruct_allocz(sizeof(*f->progress)); |
1061 | 0 | if (!f->progress) |
1062 | 0 | return AVERROR(ENOMEM); |
1063 | | |
1064 | 0 | atomic_init(&f->progress->progress[0], -1); |
1065 | 0 | atomic_init(&f->progress->progress[1], -1); |
1066 | |
|
1067 | 0 | ret = ff_thread_get_buffer(avctx, f->f, flags); |
1068 | 0 | if (ret) |
1069 | 0 | av_refstruct_unref(&f->progress); |
1070 | 0 | return ret; |
1071 | 0 | } |
1072 | | |
1073 | | void ff_thread_release_ext_buffer(ThreadFrame *f) |
1074 | 0 | { |
1075 | 0 | av_refstruct_unref(&f->progress); |
1076 | 0 | f->owner[0] = f->owner[1] = NULL; |
1077 | 0 | if (f->f) |
1078 | 0 | av_frame_unref(f->f); |
1079 | 0 | } |
1080 | | |
1081 | | av_cold enum ThreadingStatus ff_thread_sync_ref(AVCodecContext *avctx, size_t offset) |
1082 | 0 | { |
1083 | 0 | PerThreadContext *p; |
1084 | 0 | const void *ref; |
1085 | |
|
1086 | 0 | if (!avctx->internal->is_copy) |
1087 | 0 | return avctx->active_thread_type & FF_THREAD_FRAME ? |
1088 | 0 | FF_THREAD_IS_FIRST_THREAD : FF_THREAD_NO_FRAME_THREADING; |
1089 | | |
1090 | 0 | p = avctx->internal->thread_ctx; |
1091 | |
|
1092 | 0 | av_assert1(memcpy(&ref, (char*)avctx->priv_data + offset, sizeof(ref)) && ref == NULL); |
1093 | |
|
1094 | 0 | memcpy(&ref, (const char*)p->parent->threads[0].avctx->priv_data + offset, sizeof(ref)); |
1095 | 0 | av_assert1(ref); |
1096 | 0 | av_refstruct_replace((char*)avctx->priv_data + offset, ref); |
1097 | |
|
1098 | 0 | return FF_THREAD_IS_COPY; |
1099 | 0 | } |
1100 | | |
1101 | | int ff_thread_get_packet(AVCodecContext *avctx, AVPacket *pkt) |
1102 | 0 | { |
1103 | 0 | PerThreadContext *p = avctx->internal->thread_ctx; |
1104 | |
|
1105 | 0 | if (!AVPACKET_IS_EMPTY(p->avpkt)) { |
1106 | 0 | av_packet_move_ref(pkt, p->avpkt); |
1107 | 0 | return 0; |
1108 | 0 | } |
1109 | | |
1110 | 0 | return avctx->internal->draining ? AVERROR_EOF : AVERROR(EAGAIN); |
1111 | 0 | } |