1 /*
2 * Videotoolbox hardware acceleration
3 *
4 * copyright (c) 2012 Sebastien Zwickert
5 *
6 * This file is part of FFmpeg.
7 *
8 * FFmpeg is free software; you can redistribute it and/or
9 * modify it under the terms of the GNU Lesser General Public
10 * License as published by the Free Software Foundation; either
11 * version 2.1 of the License, or (at your option) any later version.
12 *
13 * FFmpeg is distributed in the hope that it will be useful,
14 * but WITHOUT ANY WARRANTY; without even the implied warranty of
15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
16 * Lesser General Public License for more details.
17 *
18 * You should have received a copy of the GNU Lesser General Public
19 * License along with FFmpeg; if not, write to the Free Software
20 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
21 */
22
23 #include "config.h"
24 #include "videotoolbox.h"
25 #include "libavutil/hwcontext_videotoolbox.h"
26 #include "vt_internal.h"
27 #include "libavutil/avutil.h"
28 #include "libavutil/hwcontext.h"
29 #include "libavutil/pixdesc.h"
30 #include "bytestream.h"
31 #include "decode.h"
32 #include "h264dec.h"
33 #include "hevcdec.h"
34 #include "mpegvideo.h"
35 #include <TargetConditionals.h>
36
37 #ifndef kVTVideoDecoderSpecification_RequireHardwareAcceleratedVideoDecoder
38 # define kVTVideoDecoderSpecification_RequireHardwareAcceleratedVideoDecoder CFSTR("RequireHardwareAcceleratedVideoDecoder")
39 #endif
40 #ifndef kVTVideoDecoderSpecification_EnableHardwareAcceleratedVideoDecoder
41 # define kVTVideoDecoderSpecification_EnableHardwareAcceleratedVideoDecoder CFSTR("EnableHardwareAcceleratedVideoDecoder")
42 #endif
43
44 #if !HAVE_KCMVIDEOCODECTYPE_HEVC
45 enum { kCMVideoCodecType_HEVC = 'hvc1' };
46 #endif
47
48 #define VIDEOTOOLBOX_ESDS_EXTRADATA_PADDING 12
49
50 typedef struct VTHWFrame {
51 CVPixelBufferRef pixbuf;
52 AVBufferRef *hw_frames_ctx;
53 } VTHWFrame;
54
videotoolbox_buffer_release(void * opaque,uint8_t * data)55 static void videotoolbox_buffer_release(void *opaque, uint8_t *data)
56 {
57 VTHWFrame *ref = (VTHWFrame *)data;
58 av_buffer_unref(&ref->hw_frames_ctx);
59 CVPixelBufferRelease(ref->pixbuf);
60
61 av_free(data);
62 }
63
videotoolbox_buffer_copy(VTContext * vtctx,const uint8_t * buffer,uint32_t size)64 static int videotoolbox_buffer_copy(VTContext *vtctx,
65 const uint8_t *buffer,
66 uint32_t size)
67 {
68 void *tmp;
69
70 tmp = av_fast_realloc(vtctx->bitstream,
71 &vtctx->allocated_size,
72 size);
73
74 if (!tmp)
75 return AVERROR(ENOMEM);
76
77 vtctx->bitstream = tmp;
78 memcpy(vtctx->bitstream, buffer, size);
79 vtctx->bitstream_size = size;
80
81 return 0;
82 }
83
videotoolbox_postproc_frame(void * avctx,AVFrame * frame)84 static int videotoolbox_postproc_frame(void *avctx, AVFrame *frame)
85 {
86 VTHWFrame *ref = (VTHWFrame *)frame->buf[0]->data;
87
88 if (!ref->pixbuf) {
89 av_log(avctx, AV_LOG_ERROR, "No frame decoded?\n");
90 av_frame_unref(frame);
91 return AVERROR_EXTERNAL;
92 }
93
94 frame->crop_right = 0;
95 frame->crop_left = 0;
96 frame->crop_top = 0;
97 frame->crop_bottom = 0;
98
99 frame->data[3] = (uint8_t*)ref->pixbuf;
100
101 if (ref->hw_frames_ctx) {
102 av_buffer_unref(&frame->hw_frames_ctx);
103 frame->hw_frames_ctx = av_buffer_ref(ref->hw_frames_ctx);
104 if (!frame->hw_frames_ctx)
105 return AVERROR(ENOMEM);
106 }
107
108 return 0;
109 }
110
ff_videotoolbox_alloc_frame(AVCodecContext * avctx,AVFrame * frame)111 int ff_videotoolbox_alloc_frame(AVCodecContext *avctx, AVFrame *frame)
112 {
113 size_t size = sizeof(VTHWFrame);
114 uint8_t *data = NULL;
115 AVBufferRef *buf = NULL;
116 int ret = ff_attach_decode_data(frame);
117 FrameDecodeData *fdd;
118 if (ret < 0)
119 return ret;
120
121 data = av_mallocz(size);
122 if (!data)
123 return AVERROR(ENOMEM);
124 buf = av_buffer_create(data, size, videotoolbox_buffer_release, NULL, 0);
125 if (!buf) {
126 av_freep(&data);
127 return AVERROR(ENOMEM);
128 }
129 frame->buf[0] = buf;
130
131 fdd = (FrameDecodeData*)frame->private_ref->data;
132 fdd->post_process = videotoolbox_postproc_frame;
133
134 frame->width = avctx->width;
135 frame->height = avctx->height;
136 frame->format = avctx->pix_fmt;
137
138 return 0;
139 }
140
141 #define AV_W8(p, v) *(p) = (v)
142
ff_videotoolbox_avcc_extradata_create(AVCodecContext * avctx)143 CFDataRef ff_videotoolbox_avcc_extradata_create(AVCodecContext *avctx)
144 {
145 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
146 H264Context *h = avctx->priv_data;
147 CFDataRef data = NULL;
148 uint8_t *p;
149 int vt_extradata_size = 6 + 2 + h->ps.sps->data_size + 3 + h->ps.pps->data_size;
150 uint8_t *vt_extradata = av_malloc(vt_extradata_size);
151 if (!vt_extradata)
152 return NULL;
153
154 p = vt_extradata;
155
156 AV_W8(p + 0, 1); /* version */
157 AV_W8(p + 1, h->ps.sps->data[1]); /* profile */
158 AV_W8(p + 2, h->ps.sps->data[2]); /* profile compat */
159 AV_W8(p + 3, h->ps.sps->data[3]); /* level */
160 AV_W8(p + 4, 0xff); /* 6 bits reserved (111111) + 2 bits nal size length - 3 (11) */
161 AV_W8(p + 5, 0xe1); /* 3 bits reserved (111) + 5 bits number of sps (00001) */
162 AV_WB16(p + 6, h->ps.sps->data_size);
163 memcpy(p + 8, h->ps.sps->data, h->ps.sps->data_size);
164 p += 8 + h->ps.sps->data_size;
165 AV_W8(p + 0, 1); /* number of pps */
166 AV_WB16(p + 1, h->ps.pps->data_size);
167 memcpy(p + 3, h->ps.pps->data, h->ps.pps->data_size);
168
169 p += 3 + h->ps.pps->data_size;
170 av_assert0(p - vt_extradata == vt_extradata_size);
171
172 // save sps header (profile/level) used to create decoder session,
173 // so we can detect changes and recreate it.
174 if (vtctx)
175 memcpy(vtctx->sps, h->ps.sps->data + 1, 3);
176
177 data = CFDataCreate(kCFAllocatorDefault, vt_extradata, vt_extradata_size);
178 av_free(vt_extradata);
179 return data;
180 }
181
ff_videotoolbox_hvcc_extradata_create(AVCodecContext * avctx)182 CFDataRef ff_videotoolbox_hvcc_extradata_create(AVCodecContext *avctx)
183 {
184 HEVCContext *h = avctx->priv_data;
185 int i, num_vps = 0, num_sps = 0, num_pps = 0;
186 const HEVCVPS *vps = h->ps.vps;
187 const HEVCSPS *sps = h->ps.sps;
188 const HEVCPPS *pps = h->ps.pps;
189 PTLCommon ptlc = vps->ptl.general_ptl;
190 VUI vui = sps->vui;
191 uint8_t parallelismType;
192 CFDataRef data = NULL;
193 uint8_t *p;
194 int vt_extradata_size = 23 + 3 + 3 + 3;
195 uint8_t *vt_extradata;
196
197 #define COUNT_SIZE_PS(T, t) \
198 for (i = 0; i < HEVC_MAX_##T##PS_COUNT; i++) { \
199 if (h->ps.t##ps_list[i]) { \
200 const HEVC##T##PS *lps = (const HEVC##T##PS *)h->ps.t##ps_list[i]->data; \
201 vt_extradata_size += 2 + lps->data_size; \
202 num_##t##ps++; \
203 } \
204 }
205
206 COUNT_SIZE_PS(V, v)
207 COUNT_SIZE_PS(S, s)
208 COUNT_SIZE_PS(P, p)
209
210 vt_extradata = av_malloc(vt_extradata_size);
211 if (!vt_extradata)
212 return NULL;
213 p = vt_extradata;
214
215 /* unsigned int(8) configurationVersion = 1; */
216 AV_W8(p + 0, 1);
217
218 /*
219 * unsigned int(2) general_profile_space;
220 * unsigned int(1) general_tier_flag;
221 * unsigned int(5) general_profile_idc;
222 */
223 AV_W8(p + 1, ptlc.profile_space << 6 |
224 ptlc.tier_flag << 5 |
225 ptlc.profile_idc);
226
227 /* unsigned int(32) general_profile_compatibility_flags; */
228 memcpy(p + 2, ptlc.profile_compatibility_flag, 4);
229
230 /* unsigned int(48) general_constraint_indicator_flags; */
231 AV_W8(p + 6, ptlc.progressive_source_flag << 7 |
232 ptlc.interlaced_source_flag << 6 |
233 ptlc.non_packed_constraint_flag << 5 |
234 ptlc.frame_only_constraint_flag << 4);
235 AV_W8(p + 7, 0);
236 AV_WN32(p + 8, 0);
237
238 /* unsigned int(8) general_level_idc; */
239 AV_W8(p + 12, ptlc.level_idc);
240
241 /*
242 * bit(4) reserved = ‘1111’b;
243 * unsigned int(12) min_spatial_segmentation_idc;
244 */
245 AV_W8(p + 13, 0xf0 | (vui.min_spatial_segmentation_idc >> 4));
246 AV_W8(p + 14, vui.min_spatial_segmentation_idc & 0xff);
247
248 /*
249 * bit(6) reserved = ‘111111’b;
250 * unsigned int(2) parallelismType;
251 */
252 if (!vui.min_spatial_segmentation_idc)
253 parallelismType = 0;
254 else if (pps->entropy_coding_sync_enabled_flag && pps->tiles_enabled_flag)
255 parallelismType = 0;
256 else if (pps->entropy_coding_sync_enabled_flag)
257 parallelismType = 3;
258 else if (pps->tiles_enabled_flag)
259 parallelismType = 2;
260 else
261 parallelismType = 1;
262 AV_W8(p + 15, 0xfc | parallelismType);
263
264 /*
265 * bit(6) reserved = ‘111111’b;
266 * unsigned int(2) chromaFormat;
267 */
268 AV_W8(p + 16, sps->chroma_format_idc | 0xfc);
269
270 /*
271 * bit(5) reserved = ‘11111’b;
272 * unsigned int(3) bitDepthLumaMinus8;
273 */
274 AV_W8(p + 17, (sps->bit_depth - 8) | 0xfc);
275
276 /*
277 * bit(5) reserved = ‘11111’b;
278 * unsigned int(3) bitDepthChromaMinus8;
279 */
280 AV_W8(p + 18, (sps->bit_depth_chroma - 8) | 0xfc);
281
282 /* bit(16) avgFrameRate; */
283 AV_WB16(p + 19, 0);
284
285 /*
286 * bit(2) constantFrameRate;
287 * bit(3) numTemporalLayers;
288 * bit(1) temporalIdNested;
289 * unsigned int(2) lengthSizeMinusOne;
290 */
291 AV_W8(p + 21, 0 << 6 |
292 sps->max_sub_layers << 3 |
293 sps->temporal_id_nesting_flag << 2 |
294 3);
295
296 /* unsigned int(8) numOfArrays; */
297 AV_W8(p + 22, 3);
298
299 p += 23;
300
301 #define APPEND_PS(T, t) \
302 /* \
303 * bit(1) array_completeness; \
304 * unsigned int(1) reserved = 0; \
305 * unsigned int(6) NAL_unit_type; \
306 */ \
307 AV_W8(p, 1 << 7 | \
308 HEVC_NAL_##T##PS & 0x3f); \
309 /* unsigned int(16) numNalus; */ \
310 AV_WB16(p + 1, num_##t##ps); \
311 p += 3; \
312 for (i = 0; i < HEVC_MAX_##T##PS_COUNT; i++) { \
313 if (h->ps.t##ps_list[i]) { \
314 const HEVC##T##PS *lps = (const HEVC##T##PS *)h->ps.t##ps_list[i]->data; \
315 /* unsigned int(16) nalUnitLength; */ \
316 AV_WB16(p, lps->data_size); \
317 /* bit(8*nalUnitLength) nalUnit; */ \
318 memcpy(p + 2, lps->data, lps->data_size); \
319 p += 2 + lps->data_size; \
320 } \
321 }
322
323 APPEND_PS(V, v)
324 APPEND_PS(S, s)
325 APPEND_PS(P, p)
326
327 av_assert0(p - vt_extradata == vt_extradata_size);
328
329 data = CFDataCreate(kCFAllocatorDefault, vt_extradata, vt_extradata_size);
330 av_free(vt_extradata);
331 return data;
332 }
333
ff_videotoolbox_h264_start_frame(AVCodecContext * avctx,const uint8_t * buffer,uint32_t size)334 int ff_videotoolbox_h264_start_frame(AVCodecContext *avctx,
335 const uint8_t *buffer,
336 uint32_t size)
337 {
338 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
339 H264Context *h = avctx->priv_data;
340
341 if (h->is_avc == 1) {
342 return videotoolbox_buffer_copy(vtctx, buffer, size);
343 }
344
345 return 0;
346 }
347
videotoolbox_h264_decode_params(AVCodecContext * avctx,int type,const uint8_t * buffer,uint32_t size)348 static int videotoolbox_h264_decode_params(AVCodecContext *avctx,
349 int type,
350 const uint8_t *buffer,
351 uint32_t size)
352 {
353 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
354 H264Context *h = avctx->priv_data;
355
356 // save sps header (profile/level) used to create decoder session
357 if (!vtctx->sps[0])
358 memcpy(vtctx->sps, h->ps.sps->data + 1, 3);
359
360 if (type == H264_NAL_SPS) {
361 if (size > 4 && memcmp(vtctx->sps, buffer + 1, 3) != 0) {
362 vtctx->reconfig_needed = true;
363 memcpy(vtctx->sps, buffer + 1, 3);
364 }
365 }
366
367 // pass-through SPS/PPS changes to the decoder
368 return ff_videotoolbox_h264_decode_slice(avctx, buffer, size);
369 }
370
videotoolbox_common_decode_slice(AVCodecContext * avctx,const uint8_t * buffer,uint32_t size)371 static int videotoolbox_common_decode_slice(AVCodecContext *avctx,
372 const uint8_t *buffer,
373 uint32_t size)
374 {
375 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
376 void *tmp;
377
378 tmp = av_fast_realloc(vtctx->bitstream,
379 &vtctx->allocated_size,
380 vtctx->bitstream_size+size+4);
381 if (!tmp)
382 return AVERROR(ENOMEM);
383
384 vtctx->bitstream = tmp;
385
386 AV_WB32(vtctx->bitstream + vtctx->bitstream_size, size);
387 memcpy(vtctx->bitstream + vtctx->bitstream_size + 4, buffer, size);
388
389 vtctx->bitstream_size += size + 4;
390
391 return 0;
392 }
393
ff_videotoolbox_h264_decode_slice(AVCodecContext * avctx,const uint8_t * buffer,uint32_t size)394 int ff_videotoolbox_h264_decode_slice(AVCodecContext *avctx,
395 const uint8_t *buffer,
396 uint32_t size)
397 {
398 H264Context *h = avctx->priv_data;
399
400 if (h->is_avc == 1)
401 return 0;
402
403 return videotoolbox_common_decode_slice(avctx, buffer, size);
404 }
405
ff_videotoolbox_uninit(AVCodecContext * avctx)406 int ff_videotoolbox_uninit(AVCodecContext *avctx)
407 {
408 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
409 if (vtctx) {
410 av_freep(&vtctx->bitstream);
411 if (vtctx->frame)
412 CVPixelBufferRelease(vtctx->frame);
413 }
414
415 return 0;
416 }
417
418 #if CONFIG_VIDEOTOOLBOX
419 // Return the AVVideotoolboxContext that matters currently. Where it comes from
420 // depends on the API used.
videotoolbox_get_context(AVCodecContext * avctx)421 static AVVideotoolboxContext *videotoolbox_get_context(AVCodecContext *avctx)
422 {
423 // Somewhat tricky because the user can call av_videotoolbox_default_free()
424 // at any time, even when the codec is closed.
425 if (avctx->internal && avctx->internal->hwaccel_priv_data) {
426 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
427 if (vtctx->vt_ctx)
428 return vtctx->vt_ctx;
429 }
430 return avctx->hwaccel_context;
431 }
432
videotoolbox_buffer_create(AVCodecContext * avctx,AVFrame * frame)433 static int videotoolbox_buffer_create(AVCodecContext *avctx, AVFrame *frame)
434 {
435 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
436 CVPixelBufferRef pixbuf = (CVPixelBufferRef)vtctx->frame;
437 OSType pixel_format = CVPixelBufferGetPixelFormatType(pixbuf);
438 enum AVPixelFormat sw_format = av_map_videotoolbox_format_to_pixfmt(pixel_format);
439 int width = CVPixelBufferGetWidth(pixbuf);
440 int height = CVPixelBufferGetHeight(pixbuf);
441 AVHWFramesContext *cached_frames;
442 VTHWFrame *ref;
443 int ret;
444
445 if (!frame->buf[0] || frame->data[3]) {
446 av_log(avctx, AV_LOG_ERROR, "videotoolbox: invalid state\n");
447 av_frame_unref(frame);
448 return AVERROR_EXTERNAL;
449 }
450
451 ref = (VTHWFrame *)frame->buf[0]->data;
452
453 if (ref->pixbuf)
454 CVPixelBufferRelease(ref->pixbuf);
455 ref->pixbuf = vtctx->frame;
456 vtctx->frame = NULL;
457
458 // Old API code path.
459 if (!vtctx->cached_hw_frames_ctx)
460 return 0;
461
462 cached_frames = (AVHWFramesContext*)vtctx->cached_hw_frames_ctx->data;
463
464 if (cached_frames->sw_format != sw_format ||
465 cached_frames->width != width ||
466 cached_frames->height != height) {
467 AVBufferRef *hw_frames_ctx = av_hwframe_ctx_alloc(cached_frames->device_ref);
468 AVHWFramesContext *hw_frames;
469 if (!hw_frames_ctx)
470 return AVERROR(ENOMEM);
471
472 hw_frames = (AVHWFramesContext*)hw_frames_ctx->data;
473 hw_frames->format = cached_frames->format;
474 hw_frames->sw_format = sw_format;
475 hw_frames->width = width;
476 hw_frames->height = height;
477
478 ret = av_hwframe_ctx_init(hw_frames_ctx);
479 if (ret < 0) {
480 av_buffer_unref(&hw_frames_ctx);
481 return ret;
482 }
483
484 av_buffer_unref(&vtctx->cached_hw_frames_ctx);
485 vtctx->cached_hw_frames_ctx = hw_frames_ctx;
486 }
487
488 av_buffer_unref(&ref->hw_frames_ctx);
489 ref->hw_frames_ctx = av_buffer_ref(vtctx->cached_hw_frames_ctx);
490 if (!ref->hw_frames_ctx)
491 return AVERROR(ENOMEM);
492
493 return 0;
494 }
495
videotoolbox_write_mp4_descr_length(PutByteContext * pb,int length)496 static void videotoolbox_write_mp4_descr_length(PutByteContext *pb, int length)
497 {
498 int i;
499 uint8_t b;
500
501 for (i = 3; i >= 0; i--) {
502 b = (length >> (i * 7)) & 0x7F;
503 if (i != 0)
504 b |= 0x80;
505
506 bytestream2_put_byteu(pb, b);
507 }
508 }
509
videotoolbox_esds_extradata_create(AVCodecContext * avctx)510 static CFDataRef videotoolbox_esds_extradata_create(AVCodecContext *avctx)
511 {
512 CFDataRef data;
513 uint8_t *rw_extradata;
514 PutByteContext pb;
515 int full_size = 3 + 5 + 13 + 5 + avctx->extradata_size + 3;
516 // ES_DescrTag data + DecoderConfigDescrTag + data + DecSpecificInfoTag + size + SLConfigDescriptor
517 int config_size = 13 + 5 + avctx->extradata_size;
518 int s;
519
520 if (!(rw_extradata = av_mallocz(full_size + VIDEOTOOLBOX_ESDS_EXTRADATA_PADDING)))
521 return NULL;
522
523 bytestream2_init_writer(&pb, rw_extradata, full_size + VIDEOTOOLBOX_ESDS_EXTRADATA_PADDING);
524 bytestream2_put_byteu(&pb, 0); // version
525 bytestream2_put_ne24(&pb, 0); // flags
526
527 // elementary stream descriptor
528 bytestream2_put_byteu(&pb, 0x03); // ES_DescrTag
529 videotoolbox_write_mp4_descr_length(&pb, full_size);
530 bytestream2_put_ne16(&pb, 0); // esid
531 bytestream2_put_byteu(&pb, 0); // stream priority (0-32)
532
533 // decoder configuration descriptor
534 bytestream2_put_byteu(&pb, 0x04); // DecoderConfigDescrTag
535 videotoolbox_write_mp4_descr_length(&pb, config_size);
536 bytestream2_put_byteu(&pb, 32); // object type indication. 32 = AV_CODEC_ID_MPEG4
537 bytestream2_put_byteu(&pb, 0x11); // stream type
538 bytestream2_put_ne24(&pb, 0); // buffer size
539 bytestream2_put_ne32(&pb, 0); // max bitrate
540 bytestream2_put_ne32(&pb, 0); // avg bitrate
541
542 // decoder specific descriptor
543 bytestream2_put_byteu(&pb, 0x05); ///< DecSpecificInfoTag
544 videotoolbox_write_mp4_descr_length(&pb, avctx->extradata_size);
545
546 bytestream2_put_buffer(&pb, avctx->extradata, avctx->extradata_size);
547
548 // SLConfigDescriptor
549 bytestream2_put_byteu(&pb, 0x06); // SLConfigDescrTag
550 bytestream2_put_byteu(&pb, 0x01); // length
551 bytestream2_put_byteu(&pb, 0x02); //
552
553 s = bytestream2_size_p(&pb);
554
555 data = CFDataCreate(kCFAllocatorDefault, rw_extradata, s);
556
557 av_freep(&rw_extradata);
558 return data;
559 }
560
videotoolbox_sample_buffer_create(CMFormatDescriptionRef fmt_desc,void * buffer,int size)561 static CMSampleBufferRef videotoolbox_sample_buffer_create(CMFormatDescriptionRef fmt_desc,
562 void *buffer,
563 int size)
564 {
565 OSStatus status;
566 CMBlockBufferRef block_buf;
567 CMSampleBufferRef sample_buf;
568
569 block_buf = NULL;
570 sample_buf = NULL;
571
572 status = CMBlockBufferCreateWithMemoryBlock(kCFAllocatorDefault,// structureAllocator
573 buffer, // memoryBlock
574 size, // blockLength
575 kCFAllocatorNull, // blockAllocator
576 NULL, // customBlockSource
577 0, // offsetToData
578 size, // dataLength
579 0, // flags
580 &block_buf);
581
582 if (!status) {
583 status = CMSampleBufferCreate(kCFAllocatorDefault, // allocator
584 block_buf, // dataBuffer
585 TRUE, // dataReady
586 0, // makeDataReadyCallback
587 0, // makeDataReadyRefcon
588 fmt_desc, // formatDescription
589 1, // numSamples
590 0, // numSampleTimingEntries
591 NULL, // sampleTimingArray
592 0, // numSampleSizeEntries
593 NULL, // sampleSizeArray
594 &sample_buf);
595 }
596
597 if (block_buf)
598 CFRelease(block_buf);
599
600 return sample_buf;
601 }
602
videotoolbox_decoder_callback(void * opaque,void * sourceFrameRefCon,OSStatus status,VTDecodeInfoFlags flags,CVImageBufferRef image_buffer,CMTime pts,CMTime duration)603 static void videotoolbox_decoder_callback(void *opaque,
604 void *sourceFrameRefCon,
605 OSStatus status,
606 VTDecodeInfoFlags flags,
607 CVImageBufferRef image_buffer,
608 CMTime pts,
609 CMTime duration)
610 {
611 AVCodecContext *avctx = opaque;
612 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
613
614 if (vtctx->frame) {
615 CVPixelBufferRelease(vtctx->frame);
616 vtctx->frame = NULL;
617 }
618
619 if (!image_buffer) {
620 av_log(avctx, AV_LOG_DEBUG, "vt decoder cb: output image buffer is null\n");
621 return;
622 }
623
624 vtctx->frame = CVPixelBufferRetain(image_buffer);
625 }
626
videotoolbox_session_decode_frame(AVCodecContext * avctx)627 static OSStatus videotoolbox_session_decode_frame(AVCodecContext *avctx)
628 {
629 OSStatus status;
630 CMSampleBufferRef sample_buf;
631 AVVideotoolboxContext *videotoolbox = videotoolbox_get_context(avctx);
632 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
633
634 sample_buf = videotoolbox_sample_buffer_create(videotoolbox->cm_fmt_desc,
635 vtctx->bitstream,
636 vtctx->bitstream_size);
637
638 if (!sample_buf)
639 return -1;
640
641 status = VTDecompressionSessionDecodeFrame(videotoolbox->session,
642 sample_buf,
643 0, // decodeFlags
644 NULL, // sourceFrameRefCon
645 0); // infoFlagsOut
646 if (status == noErr)
647 status = VTDecompressionSessionWaitForAsynchronousFrames(videotoolbox->session);
648
649 CFRelease(sample_buf);
650
651 return status;
652 }
653
videotoolbox_format_desc_create(CMVideoCodecType codec_type,CFDictionaryRef decoder_spec,int width,int height)654 static CMVideoFormatDescriptionRef videotoolbox_format_desc_create(CMVideoCodecType codec_type,
655 CFDictionaryRef decoder_spec,
656 int width,
657 int height)
658 {
659 CMFormatDescriptionRef cm_fmt_desc;
660 OSStatus status;
661
662 status = CMVideoFormatDescriptionCreate(kCFAllocatorDefault,
663 codec_type,
664 width,
665 height,
666 decoder_spec, // Dictionary of extension
667 &cm_fmt_desc);
668
669 if (status)
670 return NULL;
671
672 return cm_fmt_desc;
673 }
674
videotoolbox_buffer_attributes_create(int width,int height,OSType pix_fmt)675 static CFDictionaryRef videotoolbox_buffer_attributes_create(int width,
676 int height,
677 OSType pix_fmt)
678 {
679 CFMutableDictionaryRef buffer_attributes;
680 CFMutableDictionaryRef io_surface_properties;
681 CFNumberRef cv_pix_fmt;
682 CFNumberRef w;
683 CFNumberRef h;
684
685 w = CFNumberCreate(kCFAllocatorDefault, kCFNumberSInt32Type, &width);
686 h = CFNumberCreate(kCFAllocatorDefault, kCFNumberSInt32Type, &height);
687 cv_pix_fmt = CFNumberCreate(kCFAllocatorDefault, kCFNumberSInt32Type, &pix_fmt);
688
689 buffer_attributes = CFDictionaryCreateMutable(kCFAllocatorDefault,
690 4,
691 &kCFTypeDictionaryKeyCallBacks,
692 &kCFTypeDictionaryValueCallBacks);
693 io_surface_properties = CFDictionaryCreateMutable(kCFAllocatorDefault,
694 0,
695 &kCFTypeDictionaryKeyCallBacks,
696 &kCFTypeDictionaryValueCallBacks);
697
698 if (pix_fmt)
699 CFDictionarySetValue(buffer_attributes, kCVPixelBufferPixelFormatTypeKey, cv_pix_fmt);
700 CFDictionarySetValue(buffer_attributes, kCVPixelBufferIOSurfacePropertiesKey, io_surface_properties);
701 CFDictionarySetValue(buffer_attributes, kCVPixelBufferWidthKey, w);
702 CFDictionarySetValue(buffer_attributes, kCVPixelBufferHeightKey, h);
703 #if TARGET_OS_IPHONE
704 CFDictionarySetValue(buffer_attributes, kCVPixelBufferOpenGLESCompatibilityKey, kCFBooleanTrue);
705 #else
706 CFDictionarySetValue(buffer_attributes, kCVPixelBufferIOSurfaceOpenGLTextureCompatibilityKey, kCFBooleanTrue);
707 #endif
708
709 CFRelease(io_surface_properties);
710 CFRelease(cv_pix_fmt);
711 CFRelease(w);
712 CFRelease(h);
713
714 return buffer_attributes;
715 }
716
videotoolbox_decoder_config_create(CMVideoCodecType codec_type,AVCodecContext * avctx)717 static CFDictionaryRef videotoolbox_decoder_config_create(CMVideoCodecType codec_type,
718 AVCodecContext *avctx)
719 {
720 CFMutableDictionaryRef config_info = CFDictionaryCreateMutable(kCFAllocatorDefault,
721 0,
722 &kCFTypeDictionaryKeyCallBacks,
723 &kCFTypeDictionaryValueCallBacks);
724
725 CFDictionarySetValue(config_info,
726 codec_type == kCMVideoCodecType_HEVC ?
727 kVTVideoDecoderSpecification_EnableHardwareAcceleratedVideoDecoder :
728 kVTVideoDecoderSpecification_RequireHardwareAcceleratedVideoDecoder,
729 kCFBooleanTrue);
730
731 CFMutableDictionaryRef avc_info;
732 CFDataRef data = NULL;
733
734 avc_info = CFDictionaryCreateMutable(kCFAllocatorDefault,
735 1,
736 &kCFTypeDictionaryKeyCallBacks,
737 &kCFTypeDictionaryValueCallBacks);
738
739 switch (codec_type) {
740 case kCMVideoCodecType_MPEG4Video :
741 if (avctx->extradata_size)
742 data = videotoolbox_esds_extradata_create(avctx);
743 if (data)
744 CFDictionarySetValue(avc_info, CFSTR("esds"), data);
745 break;
746 case kCMVideoCodecType_H264 :
747 data = ff_videotoolbox_avcc_extradata_create(avctx);
748 if (data)
749 CFDictionarySetValue(avc_info, CFSTR("avcC"), data);
750 break;
751 case kCMVideoCodecType_HEVC :
752 data = ff_videotoolbox_hvcc_extradata_create(avctx);
753 if (data)
754 CFDictionarySetValue(avc_info, CFSTR("hvcC"), data);
755 break;
756 default:
757 break;
758 }
759
760 CFDictionarySetValue(config_info,
761 kCMFormatDescriptionExtension_SampleDescriptionExtensionAtoms,
762 avc_info);
763
764 if (data)
765 CFRelease(data);
766
767 CFRelease(avc_info);
768 return config_info;
769 }
770
videotoolbox_start(AVCodecContext * avctx)771 static int videotoolbox_start(AVCodecContext *avctx)
772 {
773 AVVideotoolboxContext *videotoolbox = videotoolbox_get_context(avctx);
774 OSStatus status;
775 VTDecompressionOutputCallbackRecord decoder_cb;
776 CFDictionaryRef decoder_spec;
777 CFDictionaryRef buf_attr;
778
779 if (!videotoolbox) {
780 av_log(avctx, AV_LOG_ERROR, "hwaccel context is not set\n");
781 return -1;
782 }
783
784 switch( avctx->codec_id ) {
785 case AV_CODEC_ID_H263 :
786 videotoolbox->cm_codec_type = kCMVideoCodecType_H263;
787 break;
788 case AV_CODEC_ID_H264 :
789 videotoolbox->cm_codec_type = kCMVideoCodecType_H264;
790 break;
791 case AV_CODEC_ID_HEVC :
792 videotoolbox->cm_codec_type = kCMVideoCodecType_HEVC;
793 break;
794 case AV_CODEC_ID_MPEG1VIDEO :
795 videotoolbox->cm_codec_type = kCMVideoCodecType_MPEG1Video;
796 break;
797 case AV_CODEC_ID_MPEG2VIDEO :
798 videotoolbox->cm_codec_type = kCMVideoCodecType_MPEG2Video;
799 break;
800 case AV_CODEC_ID_MPEG4 :
801 videotoolbox->cm_codec_type = kCMVideoCodecType_MPEG4Video;
802 break;
803 default :
804 break;
805 }
806
807 decoder_spec = videotoolbox_decoder_config_create(videotoolbox->cm_codec_type, avctx);
808
809 if (!decoder_spec) {
810 av_log(avctx, AV_LOG_ERROR, "decoder specification creation failed\n");
811 return -1;
812 }
813
814 videotoolbox->cm_fmt_desc = videotoolbox_format_desc_create(videotoolbox->cm_codec_type,
815 decoder_spec,
816 avctx->width,
817 avctx->height);
818 if (!videotoolbox->cm_fmt_desc) {
819 if (decoder_spec)
820 CFRelease(decoder_spec);
821
822 av_log(avctx, AV_LOG_ERROR, "format description creation failed\n");
823 return -1;
824 }
825
826 buf_attr = videotoolbox_buffer_attributes_create(avctx->width,
827 avctx->height,
828 videotoolbox->cv_pix_fmt_type);
829
830 decoder_cb.decompressionOutputCallback = videotoolbox_decoder_callback;
831 decoder_cb.decompressionOutputRefCon = avctx;
832
833 status = VTDecompressionSessionCreate(NULL, // allocator
834 videotoolbox->cm_fmt_desc, // videoFormatDescription
835 decoder_spec, // videoDecoderSpecification
836 buf_attr, // destinationImageBufferAttributes
837 &decoder_cb, // outputCallback
838 &videotoolbox->session); // decompressionSessionOut
839
840 if (decoder_spec)
841 CFRelease(decoder_spec);
842 if (buf_attr)
843 CFRelease(buf_attr);
844
845 switch (status) {
846 case kVTVideoDecoderNotAvailableNowErr:
847 av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox session not available.\n");
848 return AVERROR(ENOSYS);
849 case kVTVideoDecoderUnsupportedDataFormatErr:
850 av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox does not support this format.\n");
851 return AVERROR(ENOSYS);
852 case kVTCouldNotFindVideoDecoderErr:
853 av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox decoder for this format not found.\n");
854 return AVERROR(ENOSYS);
855 case kVTVideoDecoderMalfunctionErr:
856 av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox malfunction.\n");
857 return AVERROR(EINVAL);
858 case kVTVideoDecoderBadDataErr:
859 av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox reported invalid data.\n");
860 return AVERROR_INVALIDDATA;
861 case 0:
862 return 0;
863 default:
864 av_log(avctx, AV_LOG_VERBOSE, "Unknown VideoToolbox session creation error %d\n", (int)status);
865 return AVERROR_UNKNOWN;
866 }
867 }
868
videotoolbox_stop(AVCodecContext * avctx)869 static void videotoolbox_stop(AVCodecContext *avctx)
870 {
871 AVVideotoolboxContext *videotoolbox = videotoolbox_get_context(avctx);
872 if (!videotoolbox)
873 return;
874
875 if (videotoolbox->cm_fmt_desc) {
876 CFRelease(videotoolbox->cm_fmt_desc);
877 videotoolbox->cm_fmt_desc = NULL;
878 }
879
880 if (videotoolbox->session) {
881 VTDecompressionSessionInvalidate(videotoolbox->session);
882 CFRelease(videotoolbox->session);
883 videotoolbox->session = NULL;
884 }
885 }
886
videotoolbox_error_string(OSStatus status)887 static const char *videotoolbox_error_string(OSStatus status)
888 {
889 switch (status) {
890 case kVTVideoDecoderBadDataErr:
891 return "bad data";
892 case kVTVideoDecoderMalfunctionErr:
893 return "decoder malfunction";
894 case kVTInvalidSessionErr:
895 return "invalid session";
896 }
897 return "unknown";
898 }
899
videotoolbox_common_end_frame(AVCodecContext * avctx,AVFrame * frame)900 static int videotoolbox_common_end_frame(AVCodecContext *avctx, AVFrame *frame)
901 {
902 OSStatus status;
903 AVVideotoolboxContext *videotoolbox = videotoolbox_get_context(avctx);
904 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
905
906 if (vtctx->reconfig_needed == true) {
907 vtctx->reconfig_needed = false;
908 av_log(avctx, AV_LOG_VERBOSE, "VideoToolbox decoder needs reconfig, restarting..\n");
909 videotoolbox_stop(avctx);
910 if (videotoolbox_start(avctx) != 0) {
911 return AVERROR_EXTERNAL;
912 }
913 }
914
915 if (!videotoolbox->session || !vtctx->bitstream || !vtctx->bitstream_size)
916 return AVERROR_INVALIDDATA;
917
918 status = videotoolbox_session_decode_frame(avctx);
919 if (status != noErr) {
920 if (status == kVTVideoDecoderMalfunctionErr || status == kVTInvalidSessionErr)
921 vtctx->reconfig_needed = true;
922 av_log(avctx, AV_LOG_ERROR, "Failed to decode frame (%s, %d)\n", videotoolbox_error_string(status), (int)status);
923 return AVERROR_UNKNOWN;
924 }
925
926 if (!vtctx->frame) {
927 vtctx->reconfig_needed = true;
928 return AVERROR_UNKNOWN;
929 }
930
931 return videotoolbox_buffer_create(avctx, frame);
932 }
933
videotoolbox_h264_end_frame(AVCodecContext * avctx)934 static int videotoolbox_h264_end_frame(AVCodecContext *avctx)
935 {
936 H264Context *h = avctx->priv_data;
937 AVFrame *frame = h->cur_pic_ptr->f;
938 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
939 int ret = videotoolbox_common_end_frame(avctx, frame);
940 vtctx->bitstream_size = 0;
941 return ret;
942 }
943
videotoolbox_hevc_start_frame(AVCodecContext * avctx,const uint8_t * buffer,uint32_t size)944 static int videotoolbox_hevc_start_frame(AVCodecContext *avctx,
945 const uint8_t *buffer,
946 uint32_t size)
947 {
948 return 0;
949 }
950
videotoolbox_hevc_decode_slice(AVCodecContext * avctx,const uint8_t * buffer,uint32_t size)951 static int videotoolbox_hevc_decode_slice(AVCodecContext *avctx,
952 const uint8_t *buffer,
953 uint32_t size)
954 {
955 return videotoolbox_common_decode_slice(avctx, buffer, size);
956 }
957
958
videotoolbox_hevc_decode_params(AVCodecContext * avctx,int type,const uint8_t * buffer,uint32_t size)959 static int videotoolbox_hevc_decode_params(AVCodecContext *avctx,
960 int type,
961 const uint8_t *buffer,
962 uint32_t size)
963 {
964 return videotoolbox_common_decode_slice(avctx, buffer, size);
965 }
966
videotoolbox_hevc_end_frame(AVCodecContext * avctx)967 static int videotoolbox_hevc_end_frame(AVCodecContext *avctx)
968 {
969 HEVCContext *h = avctx->priv_data;
970 AVFrame *frame = h->ref->frame;
971 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
972
973 h->output_frame->crop_right = 0;
974 h->output_frame->crop_left = 0;
975 h->output_frame->crop_top = 0;
976 h->output_frame->crop_bottom = 0;
977
978 int ret = videotoolbox_common_end_frame(avctx, frame);
979 vtctx->bitstream_size = 0;
980 return ret;
981 }
982
videotoolbox_mpeg_start_frame(AVCodecContext * avctx,const uint8_t * buffer,uint32_t size)983 static int videotoolbox_mpeg_start_frame(AVCodecContext *avctx,
984 const uint8_t *buffer,
985 uint32_t size)
986 {
987 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
988
989 return videotoolbox_buffer_copy(vtctx, buffer, size);
990 }
991
videotoolbox_mpeg_decode_slice(AVCodecContext * avctx,const uint8_t * buffer,uint32_t size)992 static int videotoolbox_mpeg_decode_slice(AVCodecContext *avctx,
993 const uint8_t *buffer,
994 uint32_t size)
995 {
996 return 0;
997 }
998
videotoolbox_mpeg_end_frame(AVCodecContext * avctx)999 static int videotoolbox_mpeg_end_frame(AVCodecContext *avctx)
1000 {
1001 MpegEncContext *s = avctx->priv_data;
1002 AVFrame *frame = s->current_picture_ptr->f;
1003
1004 return videotoolbox_common_end_frame(avctx, frame);
1005 }
1006
videotoolbox_uninit(AVCodecContext * avctx)1007 static int videotoolbox_uninit(AVCodecContext *avctx)
1008 {
1009 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
1010 if (!vtctx)
1011 return 0;
1012
1013 ff_videotoolbox_uninit(avctx);
1014
1015 if (vtctx->vt_ctx)
1016 videotoolbox_stop(avctx);
1017
1018 av_buffer_unref(&vtctx->cached_hw_frames_ctx);
1019 av_freep(&vtctx->vt_ctx);
1020
1021 return 0;
1022 }
1023
videotoolbox_best_pixel_format(AVCodecContext * avctx)1024 static enum AVPixelFormat videotoolbox_best_pixel_format(AVCodecContext *avctx) {
1025 const AVPixFmtDescriptor *descriptor = av_pix_fmt_desc_get(avctx->pix_fmt);
1026 if (!descriptor)
1027 return AV_PIX_FMT_NV12; // same as av_videotoolbox_alloc_context()
1028
1029 int depth = descriptor->comp[0].depth;
1030 if (depth > 8) {
1031 return AV_PIX_FMT_P010;
1032 }
1033
1034 return AV_PIX_FMT_NV12;
1035 }
1036
videotoolbox_common_init(AVCodecContext * avctx)1037 static int videotoolbox_common_init(AVCodecContext *avctx)
1038 {
1039 VTContext *vtctx = avctx->internal->hwaccel_priv_data;
1040 AVHWFramesContext *hw_frames;
1041 int err;
1042
1043 // Old API - do nothing.
1044 if (avctx->hwaccel_context)
1045 return 0;
1046
1047 if (!avctx->hw_frames_ctx && !avctx->hw_device_ctx) {
1048 av_log(avctx, AV_LOG_ERROR,
1049 "Either hw_frames_ctx or hw_device_ctx must be set.\n");
1050 return AVERROR(EINVAL);
1051 }
1052
1053 vtctx->vt_ctx = av_videotoolbox_alloc_context();
1054 if (!vtctx->vt_ctx) {
1055 err = AVERROR(ENOMEM);
1056 goto fail;
1057 }
1058
1059 if (avctx->hw_frames_ctx) {
1060 hw_frames = (AVHWFramesContext*)avctx->hw_frames_ctx->data;
1061 } else {
1062 avctx->hw_frames_ctx = av_hwframe_ctx_alloc(avctx->hw_device_ctx);
1063 if (!avctx->hw_frames_ctx) {
1064 err = AVERROR(ENOMEM);
1065 goto fail;
1066 }
1067
1068 hw_frames = (AVHWFramesContext*)avctx->hw_frames_ctx->data;
1069 hw_frames->format = AV_PIX_FMT_VIDEOTOOLBOX;
1070 hw_frames->sw_format = videotoolbox_best_pixel_format(avctx);
1071 hw_frames->width = avctx->width;
1072 hw_frames->height = avctx->height;
1073
1074 err = av_hwframe_ctx_init(avctx->hw_frames_ctx);
1075 if (err < 0) {
1076 av_buffer_unref(&avctx->hw_frames_ctx);
1077 goto fail;
1078 }
1079 }
1080
1081 vtctx->cached_hw_frames_ctx = av_buffer_ref(avctx->hw_frames_ctx);
1082 if (!vtctx->cached_hw_frames_ctx) {
1083 err = AVERROR(ENOMEM);
1084 goto fail;
1085 }
1086
1087 bool full_range = avctx->color_range == AVCOL_RANGE_JPEG;
1088 vtctx->vt_ctx->cv_pix_fmt_type =
1089 av_map_videotoolbox_format_from_pixfmt2(hw_frames->sw_format, full_range);
1090 if (!vtctx->vt_ctx->cv_pix_fmt_type) {
1091 const AVPixFmtDescriptor *attempted_format =
1092 av_pix_fmt_desc_get(hw_frames->sw_format);
1093 av_log(avctx, AV_LOG_ERROR,
1094 "Failed to map underlying FFmpeg pixel format %s (%s range) to "
1095 "a VideoToolbox format!\n",
1096 attempted_format ? attempted_format->name : "<unknown>",
1097 av_color_range_name(avctx->color_range));
1098 err = AVERROR(EINVAL);
1099 goto fail;
1100 }
1101
1102 err = videotoolbox_start(avctx);
1103 if (err < 0)
1104 goto fail;
1105
1106 return 0;
1107
1108 fail:
1109 videotoolbox_uninit(avctx);
1110 return err;
1111 }
1112
videotoolbox_frame_params(AVCodecContext * avctx,AVBufferRef * hw_frames_ctx)1113 static int videotoolbox_frame_params(AVCodecContext *avctx,
1114 AVBufferRef *hw_frames_ctx)
1115 {
1116 AVHWFramesContext *frames_ctx = (AVHWFramesContext*)hw_frames_ctx->data;
1117
1118 frames_ctx->format = AV_PIX_FMT_VIDEOTOOLBOX;
1119 frames_ctx->width = avctx->coded_width;
1120 frames_ctx->height = avctx->coded_height;
1121 frames_ctx->sw_format = videotoolbox_best_pixel_format(avctx);
1122
1123 return 0;
1124 }
1125
1126 const AVHWAccel ff_h263_videotoolbox_hwaccel = {
1127 .name = "h263_videotoolbox",
1128 .type = AVMEDIA_TYPE_VIDEO,
1129 .id = AV_CODEC_ID_H263,
1130 .pix_fmt = AV_PIX_FMT_VIDEOTOOLBOX,
1131 .alloc_frame = ff_videotoolbox_alloc_frame,
1132 .start_frame = videotoolbox_mpeg_start_frame,
1133 .decode_slice = videotoolbox_mpeg_decode_slice,
1134 .end_frame = videotoolbox_mpeg_end_frame,
1135 .frame_params = videotoolbox_frame_params,
1136 .init = videotoolbox_common_init,
1137 .uninit = videotoolbox_uninit,
1138 .priv_data_size = sizeof(VTContext),
1139 };
1140
1141 const AVHWAccel ff_hevc_videotoolbox_hwaccel = {
1142 .name = "hevc_videotoolbox",
1143 .type = AVMEDIA_TYPE_VIDEO,
1144 .id = AV_CODEC_ID_HEVC,
1145 .pix_fmt = AV_PIX_FMT_VIDEOTOOLBOX,
1146 .alloc_frame = ff_videotoolbox_alloc_frame,
1147 .start_frame = videotoolbox_hevc_start_frame,
1148 .decode_slice = videotoolbox_hevc_decode_slice,
1149 .decode_params = videotoolbox_hevc_decode_params,
1150 .end_frame = videotoolbox_hevc_end_frame,
1151 .frame_params = videotoolbox_frame_params,
1152 .init = videotoolbox_common_init,
1153 .uninit = videotoolbox_uninit,
1154 .priv_data_size = sizeof(VTContext),
1155 };
1156
1157 const AVHWAccel ff_h264_videotoolbox_hwaccel = {
1158 .name = "h264_videotoolbox",
1159 .type = AVMEDIA_TYPE_VIDEO,
1160 .id = AV_CODEC_ID_H264,
1161 .pix_fmt = AV_PIX_FMT_VIDEOTOOLBOX,
1162 .alloc_frame = ff_videotoolbox_alloc_frame,
1163 .start_frame = ff_videotoolbox_h264_start_frame,
1164 .decode_slice = ff_videotoolbox_h264_decode_slice,
1165 .decode_params = videotoolbox_h264_decode_params,
1166 .end_frame = videotoolbox_h264_end_frame,
1167 .frame_params = videotoolbox_frame_params,
1168 .init = videotoolbox_common_init,
1169 .uninit = videotoolbox_uninit,
1170 .priv_data_size = sizeof(VTContext),
1171 };
1172
1173 const AVHWAccel ff_mpeg1_videotoolbox_hwaccel = {
1174 .name = "mpeg1_videotoolbox",
1175 .type = AVMEDIA_TYPE_VIDEO,
1176 .id = AV_CODEC_ID_MPEG1VIDEO,
1177 .pix_fmt = AV_PIX_FMT_VIDEOTOOLBOX,
1178 .alloc_frame = ff_videotoolbox_alloc_frame,
1179 .start_frame = videotoolbox_mpeg_start_frame,
1180 .decode_slice = videotoolbox_mpeg_decode_slice,
1181 .end_frame = videotoolbox_mpeg_end_frame,
1182 .frame_params = videotoolbox_frame_params,
1183 .init = videotoolbox_common_init,
1184 .uninit = videotoolbox_uninit,
1185 .priv_data_size = sizeof(VTContext),
1186 };
1187
1188 const AVHWAccel ff_mpeg2_videotoolbox_hwaccel = {
1189 .name = "mpeg2_videotoolbox",
1190 .type = AVMEDIA_TYPE_VIDEO,
1191 .id = AV_CODEC_ID_MPEG2VIDEO,
1192 .pix_fmt = AV_PIX_FMT_VIDEOTOOLBOX,
1193 .alloc_frame = ff_videotoolbox_alloc_frame,
1194 .start_frame = videotoolbox_mpeg_start_frame,
1195 .decode_slice = videotoolbox_mpeg_decode_slice,
1196 .end_frame = videotoolbox_mpeg_end_frame,
1197 .frame_params = videotoolbox_frame_params,
1198 .init = videotoolbox_common_init,
1199 .uninit = videotoolbox_uninit,
1200 .priv_data_size = sizeof(VTContext),
1201 };
1202
1203 const AVHWAccel ff_mpeg4_videotoolbox_hwaccel = {
1204 .name = "mpeg4_videotoolbox",
1205 .type = AVMEDIA_TYPE_VIDEO,
1206 .id = AV_CODEC_ID_MPEG4,
1207 .pix_fmt = AV_PIX_FMT_VIDEOTOOLBOX,
1208 .alloc_frame = ff_videotoolbox_alloc_frame,
1209 .start_frame = videotoolbox_mpeg_start_frame,
1210 .decode_slice = videotoolbox_mpeg_decode_slice,
1211 .end_frame = videotoolbox_mpeg_end_frame,
1212 .frame_params = videotoolbox_frame_params,
1213 .init = videotoolbox_common_init,
1214 .uninit = videotoolbox_uninit,
1215 .priv_data_size = sizeof(VTContext),
1216 };
1217
av_videotoolbox_alloc_context_with_pix_fmt(enum AVPixelFormat pix_fmt,bool full_range)1218 static AVVideotoolboxContext *av_videotoolbox_alloc_context_with_pix_fmt(enum AVPixelFormat pix_fmt,
1219 bool full_range)
1220 {
1221 AVVideotoolboxContext *ret = av_mallocz(sizeof(*ret));
1222
1223 if (ret) {
1224 ret->output_callback = videotoolbox_decoder_callback;
1225
1226 OSType cv_pix_fmt_type = av_map_videotoolbox_format_from_pixfmt2(pix_fmt, full_range);
1227 if (cv_pix_fmt_type == 0) {
1228 cv_pix_fmt_type = kCVPixelFormatType_420YpCbCr8BiPlanarVideoRange;
1229 }
1230 ret->cv_pix_fmt_type = cv_pix_fmt_type;
1231 }
1232
1233 return ret;
1234 }
1235
av_videotoolbox_alloc_context(void)1236 AVVideotoolboxContext *av_videotoolbox_alloc_context(void)
1237 {
1238 return av_videotoolbox_alloc_context_with_pix_fmt(AV_PIX_FMT_NONE, false);
1239 }
1240
av_videotoolbox_default_init(AVCodecContext * avctx)1241 int av_videotoolbox_default_init(AVCodecContext *avctx)
1242 {
1243 return av_videotoolbox_default_init2(avctx, NULL);
1244 }
1245
av_videotoolbox_default_init2(AVCodecContext * avctx,AVVideotoolboxContext * vtctx)1246 int av_videotoolbox_default_init2(AVCodecContext *avctx, AVVideotoolboxContext *vtctx)
1247 {
1248 enum AVPixelFormat pix_fmt = videotoolbox_best_pixel_format(avctx);
1249 bool full_range = avctx->color_range == AVCOL_RANGE_JPEG;
1250 avctx->hwaccel_context = vtctx ?: av_videotoolbox_alloc_context_with_pix_fmt(pix_fmt, full_range);
1251 if (!avctx->hwaccel_context)
1252 return AVERROR(ENOMEM);
1253 return videotoolbox_start(avctx);
1254 }
1255
av_videotoolbox_default_free(AVCodecContext * avctx)1256 void av_videotoolbox_default_free(AVCodecContext *avctx)
1257 {
1258
1259 videotoolbox_stop(avctx);
1260 av_freep(&avctx->hwaccel_context);
1261 }
1262 #endif /* CONFIG_VIDEOTOOLBOX */
1263