1 /*
2 * copyright (c) 2015 Rick Kern <kernrj@gmail.com>
3 *
4 * This file is part of FFmpeg.
5 *
6 * FFmpeg is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2.1 of the License, or (at your option) any later version.
10 *
11 * FFmpeg is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
15 *
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with FFmpeg; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
19 */
20
21 #include <VideoToolbox/VideoToolbox.h>
22 #include <CoreVideo/CoreVideo.h>
23 #include <CoreMedia/CoreMedia.h>
24 #include <TargetConditionals.h>
25 #include <Availability.h>
26 #include "avcodec.h"
27 #include "libavutil/opt.h"
28 #include "libavutil/avassert.h"
29 #include "libavutil/avstring.h"
30 #include "libavcodec/avcodec.h"
31 #include "libavutil/pixdesc.h"
32 #include "internal.h"
33 #include <pthread.h>
34 #include "atsc_a53.h"
35 #include "h264.h"
36 #include "h264_sei.h"
37 #include <dlfcn.h>
38
39 #if !HAVE_KCMVIDEOCODECTYPE_HEVC
40 enum { kCMVideoCodecType_HEVC = 'hvc1' };
41 #endif
42
43 #if !HAVE_KCMVIDEOCODECTYPE_HEVCWITHALPHA
44 enum { kCMVideoCodecType_HEVCWithAlpha = 'muxa' };
45 #endif
46
47 #if !HAVE_KCVPIXELFORMATTYPE_420YPCBCR10BIPLANARVIDEORANGE
48 enum { kCVPixelFormatType_420YpCbCr10BiPlanarFullRange = 'xf20' };
49 enum { kCVPixelFormatType_420YpCbCr10BiPlanarVideoRange = 'x420' };
50 #endif
51
52 #ifndef TARGET_CPU_ARM64
53 # define TARGET_CPU_ARM64 0
54 #endif
55
56 typedef OSStatus (*getParameterSetAtIndex)(CMFormatDescriptionRef videoDesc,
57 size_t parameterSetIndex,
58 const uint8_t **parameterSetPointerOut,
59 size_t *parameterSetSizeOut,
60 size_t *parameterSetCountOut,
61 int *NALUnitHeaderLengthOut);
62
63 //These symbols may not be present
64 static struct{
65 CFStringRef kCVImageBufferColorPrimaries_ITU_R_2020;
66 CFStringRef kCVImageBufferTransferFunction_ITU_R_2020;
67 CFStringRef kCVImageBufferYCbCrMatrix_ITU_R_2020;
68
69 CFStringRef kVTCompressionPropertyKey_H264EntropyMode;
70 CFStringRef kVTH264EntropyMode_CAVLC;
71 CFStringRef kVTH264EntropyMode_CABAC;
72
73 CFStringRef kVTProfileLevel_H264_Baseline_4_0;
74 CFStringRef kVTProfileLevel_H264_Baseline_4_2;
75 CFStringRef kVTProfileLevel_H264_Baseline_5_0;
76 CFStringRef kVTProfileLevel_H264_Baseline_5_1;
77 CFStringRef kVTProfileLevel_H264_Baseline_5_2;
78 CFStringRef kVTProfileLevel_H264_Baseline_AutoLevel;
79 CFStringRef kVTProfileLevel_H264_Main_4_2;
80 CFStringRef kVTProfileLevel_H264_Main_5_1;
81 CFStringRef kVTProfileLevel_H264_Main_5_2;
82 CFStringRef kVTProfileLevel_H264_Main_AutoLevel;
83 CFStringRef kVTProfileLevel_H264_High_3_0;
84 CFStringRef kVTProfileLevel_H264_High_3_1;
85 CFStringRef kVTProfileLevel_H264_High_3_2;
86 CFStringRef kVTProfileLevel_H264_High_4_0;
87 CFStringRef kVTProfileLevel_H264_High_4_1;
88 CFStringRef kVTProfileLevel_H264_High_4_2;
89 CFStringRef kVTProfileLevel_H264_High_5_1;
90 CFStringRef kVTProfileLevel_H264_High_5_2;
91 CFStringRef kVTProfileLevel_H264_High_AutoLevel;
92 CFStringRef kVTProfileLevel_H264_Extended_5_0;
93 CFStringRef kVTProfileLevel_H264_Extended_AutoLevel;
94
95 CFStringRef kVTProfileLevel_HEVC_Main_AutoLevel;
96 CFStringRef kVTProfileLevel_HEVC_Main10_AutoLevel;
97
98 CFStringRef kVTCompressionPropertyKey_RealTime;
99 CFStringRef kVTCompressionPropertyKey_TargetQualityForAlpha;
100
101 CFStringRef kVTVideoEncoderSpecification_EnableHardwareAcceleratedVideoEncoder;
102 CFStringRef kVTVideoEncoderSpecification_RequireHardwareAcceleratedVideoEncoder;
103
104 getParameterSetAtIndex CMVideoFormatDescriptionGetHEVCParameterSetAtIndex;
105 } compat_keys;
106
107 #define GET_SYM(symbol, defaultVal) \
108 do{ \
109 CFStringRef* handle = (CFStringRef*)dlsym(RTLD_DEFAULT, #symbol); \
110 if(!handle) \
111 compat_keys.symbol = CFSTR(defaultVal); \
112 else \
113 compat_keys.symbol = *handle; \
114 }while(0)
115
116 static pthread_once_t once_ctrl = PTHREAD_ONCE_INIT;
117
loadVTEncSymbols()118 static void loadVTEncSymbols(){
119 compat_keys.CMVideoFormatDescriptionGetHEVCParameterSetAtIndex =
120 (getParameterSetAtIndex)dlsym(
121 RTLD_DEFAULT,
122 "CMVideoFormatDescriptionGetHEVCParameterSetAtIndex"
123 );
124
125 GET_SYM(kCVImageBufferColorPrimaries_ITU_R_2020, "ITU_R_2020");
126 GET_SYM(kCVImageBufferTransferFunction_ITU_R_2020, "ITU_R_2020");
127 GET_SYM(kCVImageBufferYCbCrMatrix_ITU_R_2020, "ITU_R_2020");
128
129 GET_SYM(kVTCompressionPropertyKey_H264EntropyMode, "H264EntropyMode");
130 GET_SYM(kVTH264EntropyMode_CAVLC, "CAVLC");
131 GET_SYM(kVTH264EntropyMode_CABAC, "CABAC");
132
133 GET_SYM(kVTProfileLevel_H264_Baseline_4_0, "H264_Baseline_4_0");
134 GET_SYM(kVTProfileLevel_H264_Baseline_4_2, "H264_Baseline_4_2");
135 GET_SYM(kVTProfileLevel_H264_Baseline_5_0, "H264_Baseline_5_0");
136 GET_SYM(kVTProfileLevel_H264_Baseline_5_1, "H264_Baseline_5_1");
137 GET_SYM(kVTProfileLevel_H264_Baseline_5_2, "H264_Baseline_5_2");
138 GET_SYM(kVTProfileLevel_H264_Baseline_AutoLevel, "H264_Baseline_AutoLevel");
139 GET_SYM(kVTProfileLevel_H264_Main_4_2, "H264_Main_4_2");
140 GET_SYM(kVTProfileLevel_H264_Main_5_1, "H264_Main_5_1");
141 GET_SYM(kVTProfileLevel_H264_Main_5_2, "H264_Main_5_2");
142 GET_SYM(kVTProfileLevel_H264_Main_AutoLevel, "H264_Main_AutoLevel");
143 GET_SYM(kVTProfileLevel_H264_High_3_0, "H264_High_3_0");
144 GET_SYM(kVTProfileLevel_H264_High_3_1, "H264_High_3_1");
145 GET_SYM(kVTProfileLevel_H264_High_3_2, "H264_High_3_2");
146 GET_SYM(kVTProfileLevel_H264_High_4_0, "H264_High_4_0");
147 GET_SYM(kVTProfileLevel_H264_High_4_1, "H264_High_4_1");
148 GET_SYM(kVTProfileLevel_H264_High_4_2, "H264_High_4_2");
149 GET_SYM(kVTProfileLevel_H264_High_5_1, "H264_High_5_1");
150 GET_SYM(kVTProfileLevel_H264_High_5_2, "H264_High_5_2");
151 GET_SYM(kVTProfileLevel_H264_High_AutoLevel, "H264_High_AutoLevel");
152 GET_SYM(kVTProfileLevel_H264_Extended_5_0, "H264_Extended_5_0");
153 GET_SYM(kVTProfileLevel_H264_Extended_AutoLevel, "H264_Extended_AutoLevel");
154
155 GET_SYM(kVTProfileLevel_HEVC_Main_AutoLevel, "HEVC_Main_AutoLevel");
156 GET_SYM(kVTProfileLevel_HEVC_Main10_AutoLevel, "HEVC_Main10_AutoLevel");
157
158 GET_SYM(kVTCompressionPropertyKey_RealTime, "RealTime");
159 GET_SYM(kVTCompressionPropertyKey_TargetQualityForAlpha,
160 "TargetQualityForAlpha");
161
162 GET_SYM(kVTVideoEncoderSpecification_EnableHardwareAcceleratedVideoEncoder,
163 "EnableHardwareAcceleratedVideoEncoder");
164 GET_SYM(kVTVideoEncoderSpecification_RequireHardwareAcceleratedVideoEncoder,
165 "RequireHardwareAcceleratedVideoEncoder");
166 }
167
168 typedef enum VT_H264Profile {
169 H264_PROF_AUTO,
170 H264_PROF_BASELINE,
171 H264_PROF_MAIN,
172 H264_PROF_HIGH,
173 H264_PROF_EXTENDED,
174 H264_PROF_COUNT
175 } VT_H264Profile;
176
177 typedef enum VTH264Entropy{
178 VT_ENTROPY_NOT_SET,
179 VT_CAVLC,
180 VT_CABAC
181 } VTH264Entropy;
182
183 typedef enum VT_HEVCProfile {
184 HEVC_PROF_AUTO,
185 HEVC_PROF_MAIN,
186 HEVC_PROF_MAIN10,
187 HEVC_PROF_COUNT
188 } VT_HEVCProfile;
189
190 static const uint8_t start_code[] = { 0, 0, 0, 1 };
191
192 typedef struct ExtraSEI {
193 void *data;
194 size_t size;
195 } ExtraSEI;
196
197 typedef struct BufNode {
198 CMSampleBufferRef cm_buffer;
199 ExtraSEI *sei;
200 struct BufNode* next;
201 int error;
202 } BufNode;
203
204 typedef struct VTEncContext {
205 AVClass *class;
206 enum AVCodecID codec_id;
207 VTCompressionSessionRef session;
208 CFStringRef ycbcr_matrix;
209 CFStringRef color_primaries;
210 CFStringRef transfer_function;
211 getParameterSetAtIndex get_param_set_func;
212
213 pthread_mutex_t lock;
214 pthread_cond_t cv_sample_sent;
215
216 int async_error;
217
218 BufNode *q_head;
219 BufNode *q_tail;
220
221 int64_t frame_ct_out;
222 int64_t frame_ct_in;
223
224 int64_t first_pts;
225 int64_t dts_delta;
226
227 int64_t profile;
228 int64_t level;
229 int64_t entropy;
230 int64_t realtime;
231 int64_t frames_before;
232 int64_t frames_after;
233
234 int64_t allow_sw;
235 int64_t require_sw;
236 double alpha_quality;
237
238 bool flushing;
239 int has_b_frames;
240 bool warned_color_range;
241
242 /* can't be bool type since AVOption will access it as int */
243 int a53_cc;
244 } VTEncContext;
245
246 static int vtenc_populate_extradata(AVCodecContext *avctx,
247 CMVideoCodecType codec_type,
248 CFStringRef profile_level,
249 CFNumberRef gamma_level,
250 CFDictionaryRef enc_info,
251 CFDictionaryRef pixel_buffer_info);
252
253 /**
254 * NULL-safe release of *refPtr, and sets value to NULL.
255 */
vt_release_num(CFNumberRef * refPtr)256 static void vt_release_num(CFNumberRef* refPtr){
257 if (!*refPtr) {
258 return;
259 }
260
261 CFRelease(*refPtr);
262 *refPtr = NULL;
263 }
264
set_async_error(VTEncContext * vtctx,int err)265 static void set_async_error(VTEncContext *vtctx, int err)
266 {
267 BufNode *info;
268
269 pthread_mutex_lock(&vtctx->lock);
270
271 vtctx->async_error = err;
272
273 info = vtctx->q_head;
274 vtctx->q_head = vtctx->q_tail = NULL;
275
276 while (info) {
277 BufNode *next = info->next;
278 CFRelease(info->cm_buffer);
279 av_free(info);
280 info = next;
281 }
282
283 pthread_mutex_unlock(&vtctx->lock);
284 }
285
clear_frame_queue(VTEncContext * vtctx)286 static void clear_frame_queue(VTEncContext *vtctx)
287 {
288 set_async_error(vtctx, 0);
289 }
290
vtenc_q_pop(VTEncContext * vtctx,bool wait,CMSampleBufferRef * buf,ExtraSEI ** sei)291 static int vtenc_q_pop(VTEncContext *vtctx, bool wait, CMSampleBufferRef *buf, ExtraSEI **sei)
292 {
293 BufNode *info;
294
295 pthread_mutex_lock(&vtctx->lock);
296
297 if (vtctx->async_error) {
298 pthread_mutex_unlock(&vtctx->lock);
299 return vtctx->async_error;
300 }
301
302 if (vtctx->flushing && vtctx->frame_ct_in == vtctx->frame_ct_out) {
303 *buf = NULL;
304
305 pthread_mutex_unlock(&vtctx->lock);
306 return 0;
307 }
308
309 while (!vtctx->q_head && !vtctx->async_error && wait && !vtctx->flushing) {
310 pthread_cond_wait(&vtctx->cv_sample_sent, &vtctx->lock);
311 }
312
313 if (!vtctx->q_head) {
314 pthread_mutex_unlock(&vtctx->lock);
315 *buf = NULL;
316 return 0;
317 }
318
319 info = vtctx->q_head;
320 vtctx->q_head = vtctx->q_head->next;
321 if (!vtctx->q_head) {
322 vtctx->q_tail = NULL;
323 }
324
325 vtctx->frame_ct_out++;
326 pthread_mutex_unlock(&vtctx->lock);
327
328 *buf = info->cm_buffer;
329 if (sei && *buf) {
330 *sei = info->sei;
331 } else if (info->sei) {
332 if (info->sei->data) av_free(info->sei->data);
333 av_free(info->sei);
334 }
335 av_free(info);
336
337
338 return 0;
339 }
340
vtenc_q_push(VTEncContext * vtctx,CMSampleBufferRef buffer,ExtraSEI * sei)341 static void vtenc_q_push(VTEncContext *vtctx, CMSampleBufferRef buffer, ExtraSEI *sei)
342 {
343 BufNode *info = av_malloc(sizeof(BufNode));
344 if (!info) {
345 set_async_error(vtctx, AVERROR(ENOMEM));
346 return;
347 }
348
349 CFRetain(buffer);
350 info->cm_buffer = buffer;
351 info->sei = sei;
352 info->next = NULL;
353
354 pthread_mutex_lock(&vtctx->lock);
355
356 if (!vtctx->q_head) {
357 vtctx->q_head = info;
358 } else {
359 vtctx->q_tail->next = info;
360 }
361
362 vtctx->q_tail = info;
363
364 pthread_cond_signal(&vtctx->cv_sample_sent);
365 pthread_mutex_unlock(&vtctx->lock);
366 }
367
count_nalus(size_t length_code_size,CMSampleBufferRef sample_buffer,int * count)368 static int count_nalus(size_t length_code_size,
369 CMSampleBufferRef sample_buffer,
370 int *count)
371 {
372 size_t offset = 0;
373 int status;
374 int nalu_ct = 0;
375 uint8_t size_buf[4];
376 size_t src_size = CMSampleBufferGetTotalSampleSize(sample_buffer);
377 CMBlockBufferRef block = CMSampleBufferGetDataBuffer(sample_buffer);
378
379 if (length_code_size > 4)
380 return AVERROR_INVALIDDATA;
381
382 while (offset < src_size) {
383 size_t curr_src_len;
384 size_t box_len = 0;
385 size_t i;
386
387 status = CMBlockBufferCopyDataBytes(block,
388 offset,
389 length_code_size,
390 size_buf);
391
392 for (i = 0; i < length_code_size; i++) {
393 box_len <<= 8;
394 box_len |= size_buf[i];
395 }
396
397 curr_src_len = box_len + length_code_size;
398 offset += curr_src_len;
399
400 nalu_ct++;
401 }
402
403 *count = nalu_ct;
404 return 0;
405 }
406
get_cm_codec_type(enum AVCodecID id,enum AVPixelFormat fmt,double alpha_quality)407 static CMVideoCodecType get_cm_codec_type(enum AVCodecID id,
408 enum AVPixelFormat fmt,
409 double alpha_quality)
410 {
411 switch (id) {
412 case AV_CODEC_ID_H264: return kCMVideoCodecType_H264;
413 case AV_CODEC_ID_HEVC:
414 if (fmt == AV_PIX_FMT_BGRA && alpha_quality > 0.0) {
415 return kCMVideoCodecType_HEVCWithAlpha;
416 }
417 return kCMVideoCodecType_HEVC;
418 default: return 0;
419 }
420 }
421
422 /**
423 * Get the parameter sets from a CMSampleBufferRef.
424 * @param dst If *dst isn't NULL, the parameters are copied into existing
425 * memory. *dst_size must be set accordingly when *dst != NULL.
426 * If *dst is NULL, it will be allocated.
427 * In all cases, *dst_size is set to the number of bytes used starting
428 * at *dst.
429 */
get_params_size(AVCodecContext * avctx,CMVideoFormatDescriptionRef vid_fmt,size_t * size)430 static int get_params_size(
431 AVCodecContext *avctx,
432 CMVideoFormatDescriptionRef vid_fmt,
433 size_t *size)
434 {
435 VTEncContext *vtctx = avctx->priv_data;
436 size_t total_size = 0;
437 size_t ps_count;
438 int is_count_bad = 0;
439 size_t i;
440 int status;
441 status = vtctx->get_param_set_func(vid_fmt,
442 0,
443 NULL,
444 NULL,
445 &ps_count,
446 NULL);
447 if (status) {
448 is_count_bad = 1;
449 ps_count = 0;
450 status = 0;
451 }
452
453 for (i = 0; i < ps_count || is_count_bad; i++) {
454 const uint8_t *ps;
455 size_t ps_size;
456 status = vtctx->get_param_set_func(vid_fmt,
457 i,
458 &ps,
459 &ps_size,
460 NULL,
461 NULL);
462 if (status) {
463 /*
464 * When ps_count is invalid, status != 0 ends the loop normally
465 * unless we didn't get any parameter sets.
466 */
467 if (i > 0 && is_count_bad) status = 0;
468
469 break;
470 }
471
472 total_size += ps_size + sizeof(start_code);
473 }
474
475 if (status) {
476 av_log(avctx, AV_LOG_ERROR, "Error getting parameter set sizes: %d\n", status);
477 return AVERROR_EXTERNAL;
478 }
479
480 *size = total_size;
481 return 0;
482 }
483
copy_param_sets(AVCodecContext * avctx,CMVideoFormatDescriptionRef vid_fmt,uint8_t * dst,size_t dst_size)484 static int copy_param_sets(
485 AVCodecContext *avctx,
486 CMVideoFormatDescriptionRef vid_fmt,
487 uint8_t *dst,
488 size_t dst_size)
489 {
490 VTEncContext *vtctx = avctx->priv_data;
491 size_t ps_count;
492 int is_count_bad = 0;
493 int status;
494 size_t offset = 0;
495 size_t i;
496
497 status = vtctx->get_param_set_func(vid_fmt,
498 0,
499 NULL,
500 NULL,
501 &ps_count,
502 NULL);
503 if (status) {
504 is_count_bad = 1;
505 ps_count = 0;
506 status = 0;
507 }
508
509
510 for (i = 0; i < ps_count || is_count_bad; i++) {
511 const uint8_t *ps;
512 size_t ps_size;
513 size_t next_offset;
514
515 status = vtctx->get_param_set_func(vid_fmt,
516 i,
517 &ps,
518 &ps_size,
519 NULL,
520 NULL);
521 if (status) {
522 if (i > 0 && is_count_bad) status = 0;
523
524 break;
525 }
526
527 next_offset = offset + sizeof(start_code) + ps_size;
528 if (dst_size < next_offset) {
529 av_log(avctx, AV_LOG_ERROR, "Error: buffer too small for parameter sets.\n");
530 return AVERROR_BUFFER_TOO_SMALL;
531 }
532
533 memcpy(dst + offset, start_code, sizeof(start_code));
534 offset += sizeof(start_code);
535
536 memcpy(dst + offset, ps, ps_size);
537 offset = next_offset;
538 }
539
540 if (status) {
541 av_log(avctx, AV_LOG_ERROR, "Error getting parameter set data: %d\n", status);
542 return AVERROR_EXTERNAL;
543 }
544
545 return 0;
546 }
547
set_extradata(AVCodecContext * avctx,CMSampleBufferRef sample_buffer)548 static int set_extradata(AVCodecContext *avctx, CMSampleBufferRef sample_buffer)
549 {
550 CMVideoFormatDescriptionRef vid_fmt;
551 size_t total_size;
552 int status;
553
554 vid_fmt = CMSampleBufferGetFormatDescription(sample_buffer);
555 if (!vid_fmt) {
556 av_log(avctx, AV_LOG_ERROR, "No video format.\n");
557 return AVERROR_EXTERNAL;
558 }
559
560 status = get_params_size(avctx, vid_fmt, &total_size);
561 if (status) {
562 av_log(avctx, AV_LOG_ERROR, "Could not get parameter sets.\n");
563 return status;
564 }
565
566 avctx->extradata = av_mallocz(total_size + AV_INPUT_BUFFER_PADDING_SIZE);
567 if (!avctx->extradata) {
568 return AVERROR(ENOMEM);
569 }
570 avctx->extradata_size = total_size;
571
572 status = copy_param_sets(avctx, vid_fmt, avctx->extradata, total_size);
573
574 if (status) {
575 av_log(avctx, AV_LOG_ERROR, "Could not copy param sets.\n");
576 return status;
577 }
578
579 return 0;
580 }
581
vtenc_output_callback(void * ctx,void * sourceFrameCtx,OSStatus status,VTEncodeInfoFlags flags,CMSampleBufferRef sample_buffer)582 static void vtenc_output_callback(
583 void *ctx,
584 void *sourceFrameCtx,
585 OSStatus status,
586 VTEncodeInfoFlags flags,
587 CMSampleBufferRef sample_buffer)
588 {
589 AVCodecContext *avctx = ctx;
590 VTEncContext *vtctx = avctx->priv_data;
591 ExtraSEI *sei = sourceFrameCtx;
592
593 if (vtctx->async_error) {
594 return;
595 }
596
597 if (status) {
598 av_log(avctx, AV_LOG_ERROR, "Error encoding frame: %d\n", (int)status);
599 set_async_error(vtctx, AVERROR_EXTERNAL);
600 return;
601 }
602
603 if (!sample_buffer) {
604 return;
605 }
606
607 if (!avctx->extradata && (avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER)) {
608 int set_status = set_extradata(avctx, sample_buffer);
609 if (set_status) {
610 set_async_error(vtctx, set_status);
611 return;
612 }
613 }
614
615 vtenc_q_push(vtctx, sample_buffer, sei);
616 }
617
get_length_code_size(AVCodecContext * avctx,CMSampleBufferRef sample_buffer,size_t * size)618 static int get_length_code_size(
619 AVCodecContext *avctx,
620 CMSampleBufferRef sample_buffer,
621 size_t *size)
622 {
623 VTEncContext *vtctx = avctx->priv_data;
624 CMVideoFormatDescriptionRef vid_fmt;
625 int isize;
626 int status;
627
628 vid_fmt = CMSampleBufferGetFormatDescription(sample_buffer);
629 if (!vid_fmt) {
630 av_log(avctx, AV_LOG_ERROR, "Error getting buffer format description.\n");
631 return AVERROR_EXTERNAL;
632 }
633
634 status = vtctx->get_param_set_func(vid_fmt,
635 0,
636 NULL,
637 NULL,
638 NULL,
639 &isize);
640 if (status) {
641 av_log(avctx, AV_LOG_ERROR, "Error getting length code size: %d\n", status);
642 return AVERROR_EXTERNAL;
643 }
644
645 *size = isize;
646 return 0;
647 }
648
649 /*
650 * Returns true on success.
651 *
652 * If profile_level_val is NULL and this method returns true, don't specify the
653 * profile/level to the encoder.
654 */
get_vt_h264_profile_level(AVCodecContext * avctx,CFStringRef * profile_level_val)655 static bool get_vt_h264_profile_level(AVCodecContext *avctx,
656 CFStringRef *profile_level_val)
657 {
658 VTEncContext *vtctx = avctx->priv_data;
659 int64_t profile = vtctx->profile;
660
661 if (profile == H264_PROF_AUTO && vtctx->level) {
662 //Need to pick a profile if level is not auto-selected.
663 profile = vtctx->has_b_frames ? H264_PROF_MAIN : H264_PROF_BASELINE;
664 }
665
666 *profile_level_val = NULL;
667
668 switch (profile) {
669 case H264_PROF_AUTO:
670 return true;
671
672 case H264_PROF_BASELINE:
673 switch (vtctx->level) {
674 case 0: *profile_level_val =
675 compat_keys.kVTProfileLevel_H264_Baseline_AutoLevel; break;
676 case 13: *profile_level_val = kVTProfileLevel_H264_Baseline_1_3; break;
677 case 30: *profile_level_val = kVTProfileLevel_H264_Baseline_3_0; break;
678 case 31: *profile_level_val = kVTProfileLevel_H264_Baseline_3_1; break;
679 case 32: *profile_level_val = kVTProfileLevel_H264_Baseline_3_2; break;
680 case 40: *profile_level_val =
681 compat_keys.kVTProfileLevel_H264_Baseline_4_0; break;
682 case 41: *profile_level_val = kVTProfileLevel_H264_Baseline_4_1; break;
683 case 42: *profile_level_val =
684 compat_keys.kVTProfileLevel_H264_Baseline_4_2; break;
685 case 50: *profile_level_val =
686 compat_keys.kVTProfileLevel_H264_Baseline_5_0; break;
687 case 51: *profile_level_val =
688 compat_keys.kVTProfileLevel_H264_Baseline_5_1; break;
689 case 52: *profile_level_val =
690 compat_keys.kVTProfileLevel_H264_Baseline_5_2; break;
691 }
692 break;
693
694 case H264_PROF_MAIN:
695 switch (vtctx->level) {
696 case 0: *profile_level_val =
697 compat_keys.kVTProfileLevel_H264_Main_AutoLevel; break;
698 case 30: *profile_level_val = kVTProfileLevel_H264_Main_3_0; break;
699 case 31: *profile_level_val = kVTProfileLevel_H264_Main_3_1; break;
700 case 32: *profile_level_val = kVTProfileLevel_H264_Main_3_2; break;
701 case 40: *profile_level_val = kVTProfileLevel_H264_Main_4_0; break;
702 case 41: *profile_level_val = kVTProfileLevel_H264_Main_4_1; break;
703 case 42: *profile_level_val =
704 compat_keys.kVTProfileLevel_H264_Main_4_2; break;
705 case 50: *profile_level_val = kVTProfileLevel_H264_Main_5_0; break;
706 case 51: *profile_level_val =
707 compat_keys.kVTProfileLevel_H264_Main_5_1; break;
708 case 52: *profile_level_val =
709 compat_keys.kVTProfileLevel_H264_Main_5_2; break;
710 }
711 break;
712
713 case H264_PROF_HIGH:
714 switch (vtctx->level) {
715 case 0: *profile_level_val =
716 compat_keys.kVTProfileLevel_H264_High_AutoLevel; break;
717 case 30: *profile_level_val =
718 compat_keys.kVTProfileLevel_H264_High_3_0; break;
719 case 31: *profile_level_val =
720 compat_keys.kVTProfileLevel_H264_High_3_1; break;
721 case 32: *profile_level_val =
722 compat_keys.kVTProfileLevel_H264_High_3_2; break;
723 case 40: *profile_level_val =
724 compat_keys.kVTProfileLevel_H264_High_4_0; break;
725 case 41: *profile_level_val =
726 compat_keys.kVTProfileLevel_H264_High_4_1; break;
727 case 42: *profile_level_val =
728 compat_keys.kVTProfileLevel_H264_High_4_2; break;
729 case 50: *profile_level_val = kVTProfileLevel_H264_High_5_0; break;
730 case 51: *profile_level_val =
731 compat_keys.kVTProfileLevel_H264_High_5_1; break;
732 case 52: *profile_level_val =
733 compat_keys.kVTProfileLevel_H264_High_5_2; break;
734 }
735 break;
736 case H264_PROF_EXTENDED:
737 switch (vtctx->level) {
738 case 0: *profile_level_val =
739 compat_keys.kVTProfileLevel_H264_Extended_AutoLevel; break;
740 case 50: *profile_level_val =
741 compat_keys.kVTProfileLevel_H264_Extended_5_0; break;
742 }
743 break;
744 }
745
746 if (!*profile_level_val) {
747 av_log(avctx, AV_LOG_ERROR, "Invalid Profile/Level.\n");
748 return false;
749 }
750
751 return true;
752 }
753
754 /*
755 * Returns true on success.
756 *
757 * If profile_level_val is NULL and this method returns true, don't specify the
758 * profile/level to the encoder.
759 */
get_vt_hevc_profile_level(AVCodecContext * avctx,CFStringRef * profile_level_val)760 static bool get_vt_hevc_profile_level(AVCodecContext *avctx,
761 CFStringRef *profile_level_val)
762 {
763 VTEncContext *vtctx = avctx->priv_data;
764 int64_t profile = vtctx->profile;
765
766 *profile_level_val = NULL;
767
768 switch (profile) {
769 case HEVC_PROF_AUTO:
770 return true;
771 case HEVC_PROF_MAIN:
772 *profile_level_val =
773 compat_keys.kVTProfileLevel_HEVC_Main_AutoLevel;
774 break;
775 case HEVC_PROF_MAIN10:
776 *profile_level_val =
777 compat_keys.kVTProfileLevel_HEVC_Main10_AutoLevel;
778 break;
779 }
780
781 if (!*profile_level_val) {
782 av_log(avctx, AV_LOG_ERROR, "Invalid Profile/Level.\n");
783 return false;
784 }
785
786 return true;
787 }
788
get_cv_pixel_format(AVCodecContext * avctx,enum AVPixelFormat fmt,enum AVColorRange range,int * av_pixel_format,int * range_guessed)789 static int get_cv_pixel_format(AVCodecContext* avctx,
790 enum AVPixelFormat fmt,
791 enum AVColorRange range,
792 int* av_pixel_format,
793 int* range_guessed)
794 {
795 if (range_guessed) *range_guessed = range != AVCOL_RANGE_MPEG &&
796 range != AVCOL_RANGE_JPEG;
797
798 //MPEG range is used when no range is set
799 if (fmt == AV_PIX_FMT_NV12) {
800 *av_pixel_format = range == AVCOL_RANGE_JPEG ?
801 kCVPixelFormatType_420YpCbCr8BiPlanarFullRange :
802 kCVPixelFormatType_420YpCbCr8BiPlanarVideoRange;
803 } else if (fmt == AV_PIX_FMT_YUV420P) {
804 *av_pixel_format = range == AVCOL_RANGE_JPEG ?
805 kCVPixelFormatType_420YpCbCr8PlanarFullRange :
806 kCVPixelFormatType_420YpCbCr8Planar;
807 } else if (fmt == AV_PIX_FMT_BGRA) {
808 *av_pixel_format = kCVPixelFormatType_32BGRA;
809 } else if (fmt == AV_PIX_FMT_P010LE) {
810 *av_pixel_format = range == AVCOL_RANGE_JPEG ?
811 kCVPixelFormatType_420YpCbCr10BiPlanarFullRange :
812 kCVPixelFormatType_420YpCbCr10BiPlanarVideoRange;
813 } else {
814 return AVERROR(EINVAL);
815 }
816
817 return 0;
818 }
819
add_color_attr(AVCodecContext * avctx,CFMutableDictionaryRef dict)820 static void add_color_attr(AVCodecContext *avctx, CFMutableDictionaryRef dict) {
821 VTEncContext *vtctx = avctx->priv_data;
822
823 if (vtctx->color_primaries) {
824 CFDictionarySetValue(dict,
825 kCVImageBufferColorPrimariesKey,
826 vtctx->color_primaries);
827 }
828
829 if (vtctx->transfer_function) {
830 CFDictionarySetValue(dict,
831 kCVImageBufferTransferFunctionKey,
832 vtctx->transfer_function);
833 }
834
835 if (vtctx->ycbcr_matrix) {
836 CFDictionarySetValue(dict,
837 kCVImageBufferYCbCrMatrixKey,
838 vtctx->ycbcr_matrix);
839 }
840 }
841
create_cv_pixel_buffer_info(AVCodecContext * avctx,CFMutableDictionaryRef * dict)842 static int create_cv_pixel_buffer_info(AVCodecContext* avctx,
843 CFMutableDictionaryRef* dict)
844 {
845 CFNumberRef cv_color_format_num = NULL;
846 CFNumberRef width_num = NULL;
847 CFNumberRef height_num = NULL;
848 CFMutableDictionaryRef pixel_buffer_info = NULL;
849 int cv_color_format;
850 int status = get_cv_pixel_format(avctx,
851 avctx->pix_fmt,
852 avctx->color_range,
853 &cv_color_format,
854 NULL);
855 if (status) return status;
856
857 pixel_buffer_info = CFDictionaryCreateMutable(
858 kCFAllocatorDefault,
859 20,
860 &kCFCopyStringDictionaryKeyCallBacks,
861 &kCFTypeDictionaryValueCallBacks);
862
863 if (!pixel_buffer_info) goto pbinfo_nomem;
864
865 cv_color_format_num = CFNumberCreate(kCFAllocatorDefault,
866 kCFNumberSInt32Type,
867 &cv_color_format);
868 if (!cv_color_format_num) goto pbinfo_nomem;
869
870 CFDictionarySetValue(pixel_buffer_info,
871 kCVPixelBufferPixelFormatTypeKey,
872 cv_color_format_num);
873 vt_release_num(&cv_color_format_num);
874
875 width_num = CFNumberCreate(kCFAllocatorDefault,
876 kCFNumberSInt32Type,
877 &avctx->width);
878 if (!width_num) return AVERROR(ENOMEM);
879
880 CFDictionarySetValue(pixel_buffer_info,
881 kCVPixelBufferWidthKey,
882 width_num);
883 vt_release_num(&width_num);
884
885 height_num = CFNumberCreate(kCFAllocatorDefault,
886 kCFNumberSInt32Type,
887 &avctx->height);
888 if (!height_num) goto pbinfo_nomem;
889
890 CFDictionarySetValue(pixel_buffer_info,
891 kCVPixelBufferHeightKey,
892 height_num);
893 vt_release_num(&height_num);
894
895 add_color_attr(avctx, pixel_buffer_info);
896
897 *dict = pixel_buffer_info;
898 return 0;
899
900 pbinfo_nomem:
901 vt_release_num(&cv_color_format_num);
902 vt_release_num(&width_num);
903 vt_release_num(&height_num);
904 if (pixel_buffer_info) CFRelease(pixel_buffer_info);
905
906 return AVERROR(ENOMEM);
907 }
908
get_cv_color_primaries(AVCodecContext * avctx,CFStringRef * primaries)909 static int get_cv_color_primaries(AVCodecContext *avctx,
910 CFStringRef *primaries)
911 {
912 enum AVColorPrimaries pri = avctx->color_primaries;
913 switch (pri) {
914 case AVCOL_PRI_UNSPECIFIED:
915 *primaries = NULL;
916 break;
917
918 case AVCOL_PRI_BT470BG:
919 *primaries = kCVImageBufferColorPrimaries_EBU_3213;
920 break;
921
922 case AVCOL_PRI_SMPTE170M:
923 *primaries = kCVImageBufferColorPrimaries_SMPTE_C;
924 break;
925
926 case AVCOL_PRI_BT709:
927 *primaries = kCVImageBufferColorPrimaries_ITU_R_709_2;
928 break;
929
930 case AVCOL_PRI_BT2020:
931 *primaries = compat_keys.kCVImageBufferColorPrimaries_ITU_R_2020;
932 break;
933
934 default:
935 av_log(avctx, AV_LOG_ERROR, "Color primaries %s is not supported.\n", av_color_primaries_name(pri));
936 *primaries = NULL;
937 return -1;
938 }
939
940 return 0;
941 }
942
get_cv_transfer_function(AVCodecContext * avctx,CFStringRef * transfer_fnc,CFNumberRef * gamma_level)943 static int get_cv_transfer_function(AVCodecContext *avctx,
944 CFStringRef *transfer_fnc,
945 CFNumberRef *gamma_level)
946 {
947 enum AVColorTransferCharacteristic trc = avctx->color_trc;
948 Float32 gamma;
949 *gamma_level = NULL;
950
951 switch (trc) {
952 case AVCOL_TRC_UNSPECIFIED:
953 *transfer_fnc = NULL;
954 break;
955
956 case AVCOL_TRC_BT709:
957 *transfer_fnc = kCVImageBufferTransferFunction_ITU_R_709_2;
958 break;
959
960 case AVCOL_TRC_SMPTE240M:
961 *transfer_fnc = kCVImageBufferTransferFunction_SMPTE_240M_1995;
962 break;
963
964 #if HAVE_KCVIMAGEBUFFERTRANSFERFUNCTION_SMPTE_ST_2084_PQ
965 case AVCOL_TRC_SMPTE2084:
966 *transfer_fnc = kCVImageBufferTransferFunction_SMPTE_ST_2084_PQ;
967 break;
968 #endif
969 #if HAVE_KCVIMAGEBUFFERTRANSFERFUNCTION_LINEAR
970 case AVCOL_TRC_LINEAR:
971 *transfer_fnc = kCVImageBufferTransferFunction_Linear;
972 break;
973 #endif
974 #if HAVE_KCVIMAGEBUFFERTRANSFERFUNCTION_ITU_R_2100_HLG
975 case AVCOL_TRC_ARIB_STD_B67:
976 *transfer_fnc = kCVImageBufferTransferFunction_ITU_R_2100_HLG;
977 break;
978 #endif
979
980 case AVCOL_TRC_GAMMA22:
981 gamma = 2.2;
982 *transfer_fnc = kCVImageBufferTransferFunction_UseGamma;
983 *gamma_level = CFNumberCreate(NULL, kCFNumberFloat32Type, &gamma);
984 break;
985
986 case AVCOL_TRC_GAMMA28:
987 gamma = 2.8;
988 *transfer_fnc = kCVImageBufferTransferFunction_UseGamma;
989 *gamma_level = CFNumberCreate(NULL, kCFNumberFloat32Type, &gamma);
990 break;
991
992 case AVCOL_TRC_BT2020_10:
993 case AVCOL_TRC_BT2020_12:
994 *transfer_fnc = compat_keys.kCVImageBufferTransferFunction_ITU_R_2020;
995 break;
996
997 default:
998 *transfer_fnc = NULL;
999 av_log(avctx, AV_LOG_ERROR, "Transfer function %s is not supported.\n", av_color_transfer_name(trc));
1000 return -1;
1001 }
1002
1003 return 0;
1004 }
1005
get_cv_ycbcr_matrix(AVCodecContext * avctx,CFStringRef * matrix)1006 static int get_cv_ycbcr_matrix(AVCodecContext *avctx, CFStringRef *matrix) {
1007 switch(avctx->colorspace) {
1008 case AVCOL_SPC_BT709:
1009 *matrix = kCVImageBufferYCbCrMatrix_ITU_R_709_2;
1010 break;
1011
1012 case AVCOL_SPC_UNSPECIFIED:
1013 *matrix = NULL;
1014 break;
1015
1016 case AVCOL_SPC_BT470BG:
1017 case AVCOL_SPC_SMPTE170M:
1018 *matrix = kCVImageBufferYCbCrMatrix_ITU_R_601_4;
1019 break;
1020
1021 case AVCOL_SPC_SMPTE240M:
1022 *matrix = kCVImageBufferYCbCrMatrix_SMPTE_240M_1995;
1023 break;
1024
1025 case AVCOL_SPC_BT2020_NCL:
1026 *matrix = compat_keys.kCVImageBufferYCbCrMatrix_ITU_R_2020;
1027 break;
1028
1029 default:
1030 av_log(avctx, AV_LOG_ERROR, "Color space %s is not supported.\n", av_color_space_name(avctx->colorspace));
1031 return -1;
1032 }
1033
1034 return 0;
1035 }
1036
1037 // constant quality only on Macs with Apple Silicon
vtenc_qscale_enabled(void)1038 static bool vtenc_qscale_enabled(void)
1039 {
1040 return TARGET_OS_OSX && TARGET_CPU_ARM64;
1041 }
1042
vtenc_create_encoder(AVCodecContext * avctx,CMVideoCodecType codec_type,CFStringRef profile_level,CFNumberRef gamma_level,CFDictionaryRef enc_info,CFDictionaryRef pixel_buffer_info,VTCompressionSessionRef * session)1043 static int vtenc_create_encoder(AVCodecContext *avctx,
1044 CMVideoCodecType codec_type,
1045 CFStringRef profile_level,
1046 CFNumberRef gamma_level,
1047 CFDictionaryRef enc_info,
1048 CFDictionaryRef pixel_buffer_info,
1049 VTCompressionSessionRef *session)
1050 {
1051 VTEncContext *vtctx = avctx->priv_data;
1052 SInt32 bit_rate = avctx->bit_rate;
1053 SInt32 max_rate = avctx->rc_max_rate;
1054 Float32 quality = avctx->global_quality / FF_QP2LAMBDA;
1055 CFNumberRef bit_rate_num;
1056 CFNumberRef quality_num;
1057 CFNumberRef bytes_per_second;
1058 CFNumberRef one_second;
1059 CFArrayRef data_rate_limits;
1060 int64_t bytes_per_second_value = 0;
1061 int64_t one_second_value = 0;
1062 void *nums[2];
1063
1064 int status = VTCompressionSessionCreate(kCFAllocatorDefault,
1065 avctx->width,
1066 avctx->height,
1067 codec_type,
1068 enc_info,
1069 pixel_buffer_info,
1070 kCFAllocatorDefault,
1071 vtenc_output_callback,
1072 avctx,
1073 session);
1074
1075 if (status || !vtctx->session) {
1076 av_log(avctx, AV_LOG_ERROR, "Error: cannot create compression session: %d\n", status);
1077
1078 #if !TARGET_OS_IPHONE
1079 if (!vtctx->allow_sw) {
1080 av_log(avctx, AV_LOG_ERROR, "Try -allow_sw 1. The hardware encoder may be busy, or not supported.\n");
1081 }
1082 #endif
1083
1084 return AVERROR_EXTERNAL;
1085 }
1086
1087 if (avctx->flags & AV_CODEC_FLAG_QSCALE && !vtenc_qscale_enabled()) {
1088 av_log(avctx, AV_LOG_ERROR, "Error: -q:v qscale not available for encoder. Use -b:v bitrate instead.\n");
1089 return AVERROR_EXTERNAL;
1090 }
1091
1092 if (avctx->flags & AV_CODEC_FLAG_QSCALE) {
1093 quality = quality >= 100 ? 1.0 : quality / 100;
1094 quality_num = CFNumberCreate(kCFAllocatorDefault,
1095 kCFNumberFloat32Type,
1096 &quality);
1097 if (!quality_num) return AVERROR(ENOMEM);
1098
1099 status = VTSessionSetProperty(vtctx->session,
1100 kVTCompressionPropertyKey_Quality,
1101 quality_num);
1102 CFRelease(quality_num);
1103 } else {
1104 bit_rate_num = CFNumberCreate(kCFAllocatorDefault,
1105 kCFNumberSInt32Type,
1106 &bit_rate);
1107 if (!bit_rate_num) return AVERROR(ENOMEM);
1108
1109 status = VTSessionSetProperty(vtctx->session,
1110 kVTCompressionPropertyKey_AverageBitRate,
1111 bit_rate_num);
1112 CFRelease(bit_rate_num);
1113 }
1114
1115 if (status) {
1116 av_log(avctx, AV_LOG_ERROR, "Error setting bitrate property: %d\n", status);
1117 return AVERROR_EXTERNAL;
1118 }
1119
1120 if (vtctx->codec_id == AV_CODEC_ID_H264 && max_rate > 0) {
1121 // kVTCompressionPropertyKey_DataRateLimits is not available for HEVC
1122 bytes_per_second_value = max_rate >> 3;
1123 bytes_per_second = CFNumberCreate(kCFAllocatorDefault,
1124 kCFNumberSInt64Type,
1125 &bytes_per_second_value);
1126 if (!bytes_per_second) {
1127 return AVERROR(ENOMEM);
1128 }
1129 one_second_value = 1;
1130 one_second = CFNumberCreate(kCFAllocatorDefault,
1131 kCFNumberSInt64Type,
1132 &one_second_value);
1133 if (!one_second) {
1134 CFRelease(bytes_per_second);
1135 return AVERROR(ENOMEM);
1136 }
1137 nums[0] = (void *)bytes_per_second;
1138 nums[1] = (void *)one_second;
1139 data_rate_limits = CFArrayCreate(kCFAllocatorDefault,
1140 (const void **)nums,
1141 2,
1142 &kCFTypeArrayCallBacks);
1143
1144 if (!data_rate_limits) {
1145 CFRelease(bytes_per_second);
1146 CFRelease(one_second);
1147 return AVERROR(ENOMEM);
1148 }
1149 status = VTSessionSetProperty(vtctx->session,
1150 kVTCompressionPropertyKey_DataRateLimits,
1151 data_rate_limits);
1152
1153 CFRelease(bytes_per_second);
1154 CFRelease(one_second);
1155 CFRelease(data_rate_limits);
1156
1157 if (status) {
1158 av_log(avctx, AV_LOG_ERROR, "Error setting max bitrate property: %d\n", status);
1159 return AVERROR_EXTERNAL;
1160 }
1161 }
1162
1163 if (vtctx->codec_id == AV_CODEC_ID_HEVC) {
1164 if (avctx->pix_fmt == AV_PIX_FMT_BGRA && vtctx->alpha_quality > 0.0) {
1165 CFNumberRef alpha_quality_num = CFNumberCreate(kCFAllocatorDefault,
1166 kCFNumberDoubleType,
1167 &vtctx->alpha_quality);
1168 if (!alpha_quality_num) return AVERROR(ENOMEM);
1169
1170 status = VTSessionSetProperty(vtctx->session,
1171 compat_keys.kVTCompressionPropertyKey_TargetQualityForAlpha,
1172 alpha_quality_num);
1173 CFRelease(alpha_quality_num);
1174 }
1175 }
1176
1177 if (profile_level) {
1178 status = VTSessionSetProperty(vtctx->session,
1179 kVTCompressionPropertyKey_ProfileLevel,
1180 profile_level);
1181 if (status) {
1182 av_log(avctx, AV_LOG_ERROR, "Error setting profile/level property: %d. Output will be encoded using a supported profile/level combination.\n", status);
1183 }
1184 }
1185
1186 if (avctx->gop_size > 0) {
1187 CFNumberRef interval = CFNumberCreate(kCFAllocatorDefault,
1188 kCFNumberIntType,
1189 &avctx->gop_size);
1190 if (!interval) {
1191 return AVERROR(ENOMEM);
1192 }
1193
1194 status = VTSessionSetProperty(vtctx->session,
1195 kVTCompressionPropertyKey_MaxKeyFrameInterval,
1196 interval);
1197 CFRelease(interval);
1198
1199 if (status) {
1200 av_log(avctx, AV_LOG_ERROR, "Error setting 'max key-frame interval' property: %d\n", status);
1201 return AVERROR_EXTERNAL;
1202 }
1203 }
1204
1205 if (vtctx->frames_before) {
1206 status = VTSessionSetProperty(vtctx->session,
1207 kVTCompressionPropertyKey_MoreFramesBeforeStart,
1208 kCFBooleanTrue);
1209
1210 if (status == kVTPropertyNotSupportedErr) {
1211 av_log(avctx, AV_LOG_WARNING, "frames_before property is not supported on this device. Ignoring.\n");
1212 } else if (status) {
1213 av_log(avctx, AV_LOG_ERROR, "Error setting frames_before property: %d\n", status);
1214 }
1215 }
1216
1217 if (vtctx->frames_after) {
1218 status = VTSessionSetProperty(vtctx->session,
1219 kVTCompressionPropertyKey_MoreFramesAfterEnd,
1220 kCFBooleanTrue);
1221
1222 if (status == kVTPropertyNotSupportedErr) {
1223 av_log(avctx, AV_LOG_WARNING, "frames_after property is not supported on this device. Ignoring.\n");
1224 } else if (status) {
1225 av_log(avctx, AV_LOG_ERROR, "Error setting frames_after property: %d\n", status);
1226 }
1227 }
1228
1229 if (avctx->sample_aspect_ratio.num != 0) {
1230 CFNumberRef num;
1231 CFNumberRef den;
1232 CFMutableDictionaryRef par;
1233 AVRational *avpar = &avctx->sample_aspect_ratio;
1234
1235 av_reduce(&avpar->num, &avpar->den,
1236 avpar->num, avpar->den,
1237 0xFFFFFFFF);
1238
1239 num = CFNumberCreate(kCFAllocatorDefault,
1240 kCFNumberIntType,
1241 &avpar->num);
1242
1243 den = CFNumberCreate(kCFAllocatorDefault,
1244 kCFNumberIntType,
1245 &avpar->den);
1246
1247
1248
1249 par = CFDictionaryCreateMutable(kCFAllocatorDefault,
1250 2,
1251 &kCFCopyStringDictionaryKeyCallBacks,
1252 &kCFTypeDictionaryValueCallBacks);
1253
1254 if (!par || !num || !den) {
1255 if (par) CFRelease(par);
1256 if (num) CFRelease(num);
1257 if (den) CFRelease(den);
1258
1259 return AVERROR(ENOMEM);
1260 }
1261
1262 CFDictionarySetValue(
1263 par,
1264 kCMFormatDescriptionKey_PixelAspectRatioHorizontalSpacing,
1265 num);
1266
1267 CFDictionarySetValue(
1268 par,
1269 kCMFormatDescriptionKey_PixelAspectRatioVerticalSpacing,
1270 den);
1271
1272 status = VTSessionSetProperty(vtctx->session,
1273 kVTCompressionPropertyKey_PixelAspectRatio,
1274 par);
1275
1276 CFRelease(par);
1277 CFRelease(num);
1278 CFRelease(den);
1279
1280 if (status) {
1281 av_log(avctx,
1282 AV_LOG_ERROR,
1283 "Error setting pixel aspect ratio to %d:%d: %d.\n",
1284 avctx->sample_aspect_ratio.num,
1285 avctx->sample_aspect_ratio.den,
1286 status);
1287
1288 return AVERROR_EXTERNAL;
1289 }
1290 }
1291
1292
1293 if (vtctx->transfer_function) {
1294 status = VTSessionSetProperty(vtctx->session,
1295 kVTCompressionPropertyKey_TransferFunction,
1296 vtctx->transfer_function);
1297
1298 if (status) {
1299 av_log(avctx, AV_LOG_WARNING, "Could not set transfer function: %d\n", status);
1300 }
1301 }
1302
1303
1304 if (vtctx->ycbcr_matrix) {
1305 status = VTSessionSetProperty(vtctx->session,
1306 kVTCompressionPropertyKey_YCbCrMatrix,
1307 vtctx->ycbcr_matrix);
1308
1309 if (status) {
1310 av_log(avctx, AV_LOG_WARNING, "Could not set ycbcr matrix: %d\n", status);
1311 }
1312 }
1313
1314
1315 if (vtctx->color_primaries) {
1316 status = VTSessionSetProperty(vtctx->session,
1317 kVTCompressionPropertyKey_ColorPrimaries,
1318 vtctx->color_primaries);
1319
1320 if (status) {
1321 av_log(avctx, AV_LOG_WARNING, "Could not set color primaries: %d\n", status);
1322 }
1323 }
1324
1325 if (gamma_level) {
1326 status = VTSessionSetProperty(vtctx->session,
1327 kCVImageBufferGammaLevelKey,
1328 gamma_level);
1329
1330 if (status) {
1331 av_log(avctx, AV_LOG_WARNING, "Could not set gamma level: %d\n", status);
1332 }
1333 }
1334
1335 if (!vtctx->has_b_frames) {
1336 status = VTSessionSetProperty(vtctx->session,
1337 kVTCompressionPropertyKey_AllowFrameReordering,
1338 kCFBooleanFalse);
1339
1340 if (status) {
1341 av_log(avctx, AV_LOG_ERROR, "Error setting 'allow frame reordering' property: %d\n", status);
1342 return AVERROR_EXTERNAL;
1343 }
1344 }
1345
1346 if (vtctx->entropy != VT_ENTROPY_NOT_SET) {
1347 CFStringRef entropy = vtctx->entropy == VT_CABAC ?
1348 compat_keys.kVTH264EntropyMode_CABAC:
1349 compat_keys.kVTH264EntropyMode_CAVLC;
1350
1351 status = VTSessionSetProperty(vtctx->session,
1352 compat_keys.kVTCompressionPropertyKey_H264EntropyMode,
1353 entropy);
1354
1355 if (status) {
1356 av_log(avctx, AV_LOG_ERROR, "Error setting entropy property: %d\n", status);
1357 }
1358 }
1359
1360 if (vtctx->realtime) {
1361 status = VTSessionSetProperty(vtctx->session,
1362 compat_keys.kVTCompressionPropertyKey_RealTime,
1363 kCFBooleanTrue);
1364
1365 if (status) {
1366 av_log(avctx, AV_LOG_ERROR, "Error setting realtime property: %d\n", status);
1367 }
1368 }
1369
1370 status = VTCompressionSessionPrepareToEncodeFrames(vtctx->session);
1371 if (status) {
1372 av_log(avctx, AV_LOG_ERROR, "Error: cannot prepare encoder: %d\n", status);
1373 return AVERROR_EXTERNAL;
1374 }
1375
1376 return 0;
1377 }
1378
vtenc_configure_encoder(AVCodecContext * avctx)1379 static int vtenc_configure_encoder(AVCodecContext *avctx)
1380 {
1381 CFMutableDictionaryRef enc_info;
1382 CFMutableDictionaryRef pixel_buffer_info;
1383 CMVideoCodecType codec_type;
1384 VTEncContext *vtctx = avctx->priv_data;
1385 CFStringRef profile_level;
1386 CFNumberRef gamma_level = NULL;
1387 int status;
1388
1389 codec_type = get_cm_codec_type(avctx->codec_id, avctx->pix_fmt, vtctx->alpha_quality);
1390 if (!codec_type) {
1391 av_log(avctx, AV_LOG_ERROR, "Error: no mapping for AVCodecID %d\n", avctx->codec_id);
1392 return AVERROR(EINVAL);
1393 }
1394
1395 vtctx->codec_id = avctx->codec_id;
1396 avctx->max_b_frames = 16;
1397
1398 if (vtctx->codec_id == AV_CODEC_ID_H264) {
1399 vtctx->get_param_set_func = CMVideoFormatDescriptionGetH264ParameterSetAtIndex;
1400
1401 vtctx->has_b_frames = avctx->max_b_frames > 0;
1402 if(vtctx->has_b_frames && vtctx->profile == H264_PROF_BASELINE){
1403 av_log(avctx, AV_LOG_WARNING, "Cannot use B-frames with baseline profile. Output will not contain B-frames.\n");
1404 vtctx->has_b_frames = 0;
1405 }
1406
1407 if (vtctx->entropy == VT_CABAC && vtctx->profile == H264_PROF_BASELINE) {
1408 av_log(avctx, AV_LOG_WARNING, "CABAC entropy requires 'main' or 'high' profile, but baseline was requested. Encode will not use CABAC entropy.\n");
1409 vtctx->entropy = VT_ENTROPY_NOT_SET;
1410 }
1411
1412 if (!get_vt_h264_profile_level(avctx, &profile_level)) return AVERROR(EINVAL);
1413 } else {
1414 vtctx->get_param_set_func = compat_keys.CMVideoFormatDescriptionGetHEVCParameterSetAtIndex;
1415 if (!vtctx->get_param_set_func) return AVERROR(EINVAL);
1416 if (!get_vt_hevc_profile_level(avctx, &profile_level)) return AVERROR(EINVAL);
1417 // HEVC has b-byramid
1418 vtctx->has_b_frames = avctx->max_b_frames > 0 ? 2 : 0;
1419 }
1420
1421 enc_info = CFDictionaryCreateMutable(
1422 kCFAllocatorDefault,
1423 20,
1424 &kCFCopyStringDictionaryKeyCallBacks,
1425 &kCFTypeDictionaryValueCallBacks
1426 );
1427
1428 if (!enc_info) return AVERROR(ENOMEM);
1429
1430 #if !TARGET_OS_IPHONE
1431 if(vtctx->require_sw) {
1432 CFDictionarySetValue(enc_info,
1433 compat_keys.kVTVideoEncoderSpecification_EnableHardwareAcceleratedVideoEncoder,
1434 kCFBooleanFalse);
1435 } else if (!vtctx->allow_sw) {
1436 CFDictionarySetValue(enc_info,
1437 compat_keys.kVTVideoEncoderSpecification_RequireHardwareAcceleratedVideoEncoder,
1438 kCFBooleanTrue);
1439 } else {
1440 CFDictionarySetValue(enc_info,
1441 compat_keys.kVTVideoEncoderSpecification_EnableHardwareAcceleratedVideoEncoder,
1442 kCFBooleanTrue);
1443 }
1444 #endif
1445
1446 if (avctx->pix_fmt != AV_PIX_FMT_VIDEOTOOLBOX) {
1447 status = create_cv_pixel_buffer_info(avctx, &pixel_buffer_info);
1448 if (status)
1449 goto init_cleanup;
1450 } else {
1451 pixel_buffer_info = NULL;
1452 }
1453
1454 vtctx->dts_delta = vtctx->has_b_frames ? -1 : 0;
1455
1456 get_cv_transfer_function(avctx, &vtctx->transfer_function, &gamma_level);
1457 get_cv_ycbcr_matrix(avctx, &vtctx->ycbcr_matrix);
1458 get_cv_color_primaries(avctx, &vtctx->color_primaries);
1459
1460
1461 if (avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER) {
1462 status = vtenc_populate_extradata(avctx,
1463 codec_type,
1464 profile_level,
1465 gamma_level,
1466 enc_info,
1467 pixel_buffer_info);
1468 if (status)
1469 goto init_cleanup;
1470 }
1471
1472 status = vtenc_create_encoder(avctx,
1473 codec_type,
1474 profile_level,
1475 gamma_level,
1476 enc_info,
1477 pixel_buffer_info,
1478 &vtctx->session);
1479
1480 init_cleanup:
1481 if (gamma_level)
1482 CFRelease(gamma_level);
1483
1484 if (pixel_buffer_info)
1485 CFRelease(pixel_buffer_info);
1486
1487 CFRelease(enc_info);
1488
1489 return status;
1490 }
1491
vtenc_init(AVCodecContext * avctx)1492 static av_cold int vtenc_init(AVCodecContext *avctx)
1493 {
1494 VTEncContext *vtctx = avctx->priv_data;
1495 CFBooleanRef has_b_frames_cfbool;
1496 int status;
1497
1498 pthread_once(&once_ctrl, loadVTEncSymbols);
1499
1500 pthread_mutex_init(&vtctx->lock, NULL);
1501 pthread_cond_init(&vtctx->cv_sample_sent, NULL);
1502
1503 vtctx->session = NULL;
1504 status = vtenc_configure_encoder(avctx);
1505 if (status) return status;
1506
1507 status = VTSessionCopyProperty(vtctx->session,
1508 kVTCompressionPropertyKey_AllowFrameReordering,
1509 kCFAllocatorDefault,
1510 &has_b_frames_cfbool);
1511
1512 if (!status && has_b_frames_cfbool) {
1513 //Some devices don't output B-frames for main profile, even if requested.
1514 // HEVC has b-pyramid
1515 vtctx->has_b_frames = (CFBooleanGetValue(has_b_frames_cfbool) && avctx->codec_id == AV_CODEC_ID_HEVC) ? 2 : 1;
1516 CFRelease(has_b_frames_cfbool);
1517 }
1518 avctx->has_b_frames = vtctx->has_b_frames;
1519
1520 return 0;
1521 }
1522
vtenc_get_frame_info(CMSampleBufferRef buffer,bool * is_key_frame)1523 static void vtenc_get_frame_info(CMSampleBufferRef buffer, bool *is_key_frame)
1524 {
1525 CFArrayRef attachments;
1526 CFDictionaryRef attachment;
1527 CFBooleanRef not_sync;
1528 CFIndex len;
1529
1530 attachments = CMSampleBufferGetSampleAttachmentsArray(buffer, false);
1531 len = !attachments ? 0 : CFArrayGetCount(attachments);
1532
1533 if (!len) {
1534 *is_key_frame = true;
1535 return;
1536 }
1537
1538 attachment = CFArrayGetValueAtIndex(attachments, 0);
1539
1540 if (CFDictionaryGetValueIfPresent(attachment,
1541 kCMSampleAttachmentKey_NotSync,
1542 (const void **)¬_sync))
1543 {
1544 *is_key_frame = !CFBooleanGetValue(not_sync);
1545 } else {
1546 *is_key_frame = true;
1547 }
1548 }
1549
is_post_sei_nal_type(int nal_type)1550 static int is_post_sei_nal_type(int nal_type){
1551 return nal_type != H264_NAL_SEI &&
1552 nal_type != H264_NAL_SPS &&
1553 nal_type != H264_NAL_PPS &&
1554 nal_type != H264_NAL_AUD;
1555 }
1556
1557 /*
1558 * Finds the sei message start/size of type find_sei_type.
1559 * If more than one of that type exists, the last one is returned.
1560 */
find_sei_end(AVCodecContext * avctx,uint8_t * nal_data,size_t nal_size,uint8_t ** sei_end)1561 static int find_sei_end(AVCodecContext *avctx,
1562 uint8_t *nal_data,
1563 size_t nal_size,
1564 uint8_t **sei_end)
1565 {
1566 int nal_type;
1567 size_t sei_payload_size = 0;
1568 int sei_payload_type = 0;
1569 *sei_end = NULL;
1570 uint8_t *nal_start = nal_data;
1571
1572 if (!nal_size)
1573 return 0;
1574
1575 nal_type = *nal_data & 0x1F;
1576 if (nal_type != H264_NAL_SEI)
1577 return 0;
1578
1579 nal_data++;
1580 nal_size--;
1581
1582 if (nal_data[nal_size - 1] == 0x80)
1583 nal_size--;
1584
1585 while (nal_size > 0 && *nal_data > 0) {
1586 do{
1587 sei_payload_type += *nal_data;
1588 nal_data++;
1589 nal_size--;
1590 } while (nal_size > 0 && *nal_data == 0xFF);
1591
1592 if (!nal_size) {
1593 av_log(avctx, AV_LOG_ERROR, "Unexpected end of SEI NAL Unit parsing type.\n");
1594 return AVERROR_INVALIDDATA;
1595 }
1596
1597 do{
1598 sei_payload_size += *nal_data;
1599 nal_data++;
1600 nal_size--;
1601 } while (nal_size > 0 && *nal_data == 0xFF);
1602
1603 if (nal_size < sei_payload_size) {
1604 av_log(avctx, AV_LOG_ERROR, "Unexpected end of SEI NAL Unit parsing size.\n");
1605 return AVERROR_INVALIDDATA;
1606 }
1607
1608 nal_data += sei_payload_size;
1609 nal_size -= sei_payload_size;
1610 }
1611
1612 *sei_end = nal_data;
1613
1614 return nal_data - nal_start + 1;
1615 }
1616
1617 /**
1618 * Copies the data inserting emulation prevention bytes as needed.
1619 * Existing data in the destination can be taken into account by providing
1620 * dst with a dst_offset > 0.
1621 *
1622 * @return The number of bytes copied on success. On failure, the negative of
1623 * the number of bytes needed to copy src is returned.
1624 */
copy_emulation_prev(const uint8_t * src,size_t src_size,uint8_t * dst,ssize_t dst_offset,size_t dst_size)1625 static int copy_emulation_prev(const uint8_t *src,
1626 size_t src_size,
1627 uint8_t *dst,
1628 ssize_t dst_offset,
1629 size_t dst_size)
1630 {
1631 int zeros = 0;
1632 int wrote_bytes;
1633 uint8_t* dst_start;
1634 uint8_t* dst_end = dst + dst_size;
1635 const uint8_t* src_end = src + src_size;
1636 int start_at = dst_offset > 2 ? dst_offset - 2 : 0;
1637 int i;
1638 for (i = start_at; i < dst_offset && i < dst_size; i++) {
1639 if (!dst[i])
1640 zeros++;
1641 else
1642 zeros = 0;
1643 }
1644
1645 dst += dst_offset;
1646 dst_start = dst;
1647 for (; src < src_end; src++, dst++) {
1648 if (zeros == 2) {
1649 int insert_ep3_byte = *src <= 3;
1650 if (insert_ep3_byte) {
1651 if (dst < dst_end)
1652 *dst = 3;
1653 dst++;
1654 }
1655
1656 zeros = 0;
1657 }
1658
1659 if (dst < dst_end)
1660 *dst = *src;
1661
1662 if (!*src)
1663 zeros++;
1664 else
1665 zeros = 0;
1666 }
1667
1668 wrote_bytes = dst - dst_start;
1669
1670 if (dst > dst_end)
1671 return -wrote_bytes;
1672
1673 return wrote_bytes;
1674 }
1675
write_sei(const ExtraSEI * sei,int sei_type,uint8_t * dst,size_t dst_size)1676 static int write_sei(const ExtraSEI *sei,
1677 int sei_type,
1678 uint8_t *dst,
1679 size_t dst_size)
1680 {
1681 uint8_t *sei_start = dst;
1682 size_t remaining_sei_size = sei->size;
1683 size_t remaining_dst_size = dst_size;
1684 int header_bytes;
1685 int bytes_written;
1686 ssize_t offset;
1687
1688 if (!remaining_dst_size)
1689 return AVERROR_BUFFER_TOO_SMALL;
1690
1691 while (sei_type && remaining_dst_size != 0) {
1692 int sei_byte = sei_type > 255 ? 255 : sei_type;
1693 *dst = sei_byte;
1694
1695 sei_type -= sei_byte;
1696 dst++;
1697 remaining_dst_size--;
1698 }
1699
1700 if (!dst_size)
1701 return AVERROR_BUFFER_TOO_SMALL;
1702
1703 while (remaining_sei_size && remaining_dst_size != 0) {
1704 int size_byte = remaining_sei_size > 255 ? 255 : remaining_sei_size;
1705 *dst = size_byte;
1706
1707 remaining_sei_size -= size_byte;
1708 dst++;
1709 remaining_dst_size--;
1710 }
1711
1712 if (remaining_dst_size < sei->size)
1713 return AVERROR_BUFFER_TOO_SMALL;
1714
1715 header_bytes = dst - sei_start;
1716
1717 offset = header_bytes;
1718 bytes_written = copy_emulation_prev(sei->data,
1719 sei->size,
1720 sei_start,
1721 offset,
1722 dst_size);
1723 if (bytes_written < 0)
1724 return AVERROR_BUFFER_TOO_SMALL;
1725
1726 bytes_written += header_bytes;
1727 return bytes_written;
1728 }
1729
1730 /**
1731 * Copies NAL units and replaces length codes with
1732 * H.264 Annex B start codes. On failure, the contents of
1733 * dst_data may have been modified.
1734 *
1735 * @param length_code_size Byte length of each length code
1736 * @param sample_buffer NAL units prefixed with length codes.
1737 * @param sei Optional A53 closed captions SEI data.
1738 * @param dst_data Must be zeroed before calling this function.
1739 * Contains the copied NAL units prefixed with
1740 * start codes when the function returns
1741 * successfully.
1742 * @param dst_size Length of dst_data
1743 * @return 0 on success
1744 * AVERROR_INVALIDDATA if length_code_size is invalid
1745 * AVERROR_BUFFER_TOO_SMALL if dst_data is too small
1746 * or if a length_code in src_data specifies data beyond
1747 * the end of its buffer.
1748 */
copy_replace_length_codes(AVCodecContext * avctx,size_t length_code_size,CMSampleBufferRef sample_buffer,ExtraSEI * sei,uint8_t * dst_data,size_t dst_size)1749 static int copy_replace_length_codes(
1750 AVCodecContext *avctx,
1751 size_t length_code_size,
1752 CMSampleBufferRef sample_buffer,
1753 ExtraSEI *sei,
1754 uint8_t *dst_data,
1755 size_t dst_size)
1756 {
1757 size_t src_size = CMSampleBufferGetTotalSampleSize(sample_buffer);
1758 size_t remaining_src_size = src_size;
1759 size_t remaining_dst_size = dst_size;
1760 size_t src_offset = 0;
1761 int wrote_sei = 0;
1762 int status;
1763 uint8_t size_buf[4];
1764 uint8_t nal_type;
1765 CMBlockBufferRef block = CMSampleBufferGetDataBuffer(sample_buffer);
1766
1767 if (length_code_size > 4) {
1768 return AVERROR_INVALIDDATA;
1769 }
1770
1771 while (remaining_src_size > 0) {
1772 size_t curr_src_len;
1773 size_t curr_dst_len;
1774 size_t box_len = 0;
1775 size_t i;
1776
1777 uint8_t *dst_box;
1778
1779 status = CMBlockBufferCopyDataBytes(block,
1780 src_offset,
1781 length_code_size,
1782 size_buf);
1783 if (status) {
1784 av_log(avctx, AV_LOG_ERROR, "Cannot copy length: %d\n", status);
1785 return AVERROR_EXTERNAL;
1786 }
1787
1788 status = CMBlockBufferCopyDataBytes(block,
1789 src_offset + length_code_size,
1790 1,
1791 &nal_type);
1792
1793 if (status) {
1794 av_log(avctx, AV_LOG_ERROR, "Cannot copy type: %d\n", status);
1795 return AVERROR_EXTERNAL;
1796 }
1797
1798 nal_type &= 0x1F;
1799
1800 for (i = 0; i < length_code_size; i++) {
1801 box_len <<= 8;
1802 box_len |= size_buf[i];
1803 }
1804
1805 if (sei && !wrote_sei && is_post_sei_nal_type(nal_type)) {
1806 //No SEI NAL unit - insert.
1807 int wrote_bytes;
1808
1809 memcpy(dst_data, start_code, sizeof(start_code));
1810 dst_data += sizeof(start_code);
1811 remaining_dst_size -= sizeof(start_code);
1812
1813 *dst_data = H264_NAL_SEI;
1814 dst_data++;
1815 remaining_dst_size--;
1816
1817 wrote_bytes = write_sei(sei,
1818 SEI_TYPE_USER_DATA_REGISTERED_ITU_T_T35,
1819 dst_data,
1820 remaining_dst_size);
1821
1822 if (wrote_bytes < 0)
1823 return wrote_bytes;
1824
1825 remaining_dst_size -= wrote_bytes;
1826 dst_data += wrote_bytes;
1827
1828 if (remaining_dst_size <= 0)
1829 return AVERROR_BUFFER_TOO_SMALL;
1830
1831 *dst_data = 0x80;
1832
1833 dst_data++;
1834 remaining_dst_size--;
1835
1836 wrote_sei = 1;
1837 }
1838
1839 curr_src_len = box_len + length_code_size;
1840 curr_dst_len = box_len + sizeof(start_code);
1841
1842 if (remaining_src_size < curr_src_len) {
1843 return AVERROR_BUFFER_TOO_SMALL;
1844 }
1845
1846 if (remaining_dst_size < curr_dst_len) {
1847 return AVERROR_BUFFER_TOO_SMALL;
1848 }
1849
1850 dst_box = dst_data + sizeof(start_code);
1851
1852 memcpy(dst_data, start_code, sizeof(start_code));
1853 status = CMBlockBufferCopyDataBytes(block,
1854 src_offset + length_code_size,
1855 box_len,
1856 dst_box);
1857
1858 if (status) {
1859 av_log(avctx, AV_LOG_ERROR, "Cannot copy data: %d\n", status);
1860 return AVERROR_EXTERNAL;
1861 }
1862
1863 if (sei && !wrote_sei && nal_type == H264_NAL_SEI) {
1864 //Found SEI NAL unit - append.
1865 int wrote_bytes;
1866 int old_sei_length;
1867 int extra_bytes;
1868 uint8_t *new_sei;
1869 old_sei_length = find_sei_end(avctx, dst_box, box_len, &new_sei);
1870 if (old_sei_length < 0)
1871 return status;
1872
1873 wrote_bytes = write_sei(sei,
1874 SEI_TYPE_USER_DATA_REGISTERED_ITU_T_T35,
1875 new_sei,
1876 remaining_dst_size - old_sei_length);
1877 if (wrote_bytes < 0)
1878 return wrote_bytes;
1879
1880 if (new_sei + wrote_bytes >= dst_data + remaining_dst_size)
1881 return AVERROR_BUFFER_TOO_SMALL;
1882
1883 new_sei[wrote_bytes++] = 0x80;
1884 extra_bytes = wrote_bytes - (dst_box + box_len - new_sei);
1885
1886 dst_data += extra_bytes;
1887 remaining_dst_size -= extra_bytes;
1888
1889 wrote_sei = 1;
1890 }
1891
1892 src_offset += curr_src_len;
1893 dst_data += curr_dst_len;
1894
1895 remaining_src_size -= curr_src_len;
1896 remaining_dst_size -= curr_dst_len;
1897 }
1898
1899 return 0;
1900 }
1901
1902 /**
1903 * Returns a sufficient number of bytes to contain the sei data.
1904 * It may be greater than the minimum required.
1905 */
get_sei_msg_bytes(const ExtraSEI * sei,int type)1906 static int get_sei_msg_bytes(const ExtraSEI* sei, int type){
1907 int copied_size;
1908 if (sei->size == 0)
1909 return 0;
1910
1911 copied_size = -copy_emulation_prev(sei->data,
1912 sei->size,
1913 NULL,
1914 0,
1915 0);
1916
1917 if ((sei->size % 255) == 0) //may result in an extra byte
1918 copied_size++;
1919
1920 return copied_size + sei->size / 255 + 1 + type / 255 + 1;
1921 }
1922
vtenc_cm_to_avpacket(AVCodecContext * avctx,CMSampleBufferRef sample_buffer,AVPacket * pkt,ExtraSEI * sei)1923 static int vtenc_cm_to_avpacket(
1924 AVCodecContext *avctx,
1925 CMSampleBufferRef sample_buffer,
1926 AVPacket *pkt,
1927 ExtraSEI *sei)
1928 {
1929 VTEncContext *vtctx = avctx->priv_data;
1930
1931 int status;
1932 bool is_key_frame;
1933 bool add_header;
1934 size_t length_code_size;
1935 size_t header_size = 0;
1936 size_t in_buf_size;
1937 size_t out_buf_size;
1938 size_t sei_nalu_size = 0;
1939 int64_t dts_delta;
1940 int64_t time_base_num;
1941 int nalu_count;
1942 CMTime pts;
1943 CMTime dts;
1944 CMVideoFormatDescriptionRef vid_fmt;
1945
1946
1947 vtenc_get_frame_info(sample_buffer, &is_key_frame);
1948 status = get_length_code_size(avctx, sample_buffer, &length_code_size);
1949 if (status) return status;
1950
1951 add_header = is_key_frame && !(avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER);
1952
1953 if (add_header) {
1954 vid_fmt = CMSampleBufferGetFormatDescription(sample_buffer);
1955 if (!vid_fmt) {
1956 av_log(avctx, AV_LOG_ERROR, "Cannot get format description.\n");
1957 return AVERROR_EXTERNAL;
1958 }
1959
1960 int status = get_params_size(avctx, vid_fmt, &header_size);
1961 if (status) return status;
1962 }
1963
1964 status = count_nalus(length_code_size, sample_buffer, &nalu_count);
1965 if(status)
1966 return status;
1967
1968 if (sei) {
1969 size_t msg_size = get_sei_msg_bytes(sei,
1970 SEI_TYPE_USER_DATA_REGISTERED_ITU_T_T35);
1971
1972 sei_nalu_size = sizeof(start_code) + 1 + msg_size + 1;
1973 }
1974
1975 in_buf_size = CMSampleBufferGetTotalSampleSize(sample_buffer);
1976 out_buf_size = header_size +
1977 in_buf_size +
1978 sei_nalu_size +
1979 nalu_count * ((int)sizeof(start_code) - (int)length_code_size);
1980
1981 status = ff_alloc_packet2(avctx, pkt, out_buf_size, out_buf_size);
1982 if (status < 0)
1983 return status;
1984
1985 if (add_header) {
1986 status = copy_param_sets(avctx, vid_fmt, pkt->data, out_buf_size);
1987 if(status) return status;
1988 }
1989
1990 status = copy_replace_length_codes(
1991 avctx,
1992 length_code_size,
1993 sample_buffer,
1994 sei,
1995 pkt->data + header_size,
1996 pkt->size - header_size
1997 );
1998
1999 if (status) {
2000 av_log(avctx, AV_LOG_ERROR, "Error copying packet data: %d\n", status);
2001 return status;
2002 }
2003
2004 if (is_key_frame) {
2005 pkt->flags |= AV_PKT_FLAG_KEY;
2006 }
2007
2008 pts = CMSampleBufferGetPresentationTimeStamp(sample_buffer);
2009 dts = CMSampleBufferGetDecodeTimeStamp (sample_buffer);
2010
2011 if (CMTIME_IS_INVALID(dts)) {
2012 if (!vtctx->has_b_frames) {
2013 dts = pts;
2014 } else {
2015 av_log(avctx, AV_LOG_ERROR, "DTS is invalid.\n");
2016 return AVERROR_EXTERNAL;
2017 }
2018 }
2019
2020 dts_delta = vtctx->dts_delta >= 0 ? vtctx->dts_delta : 0;
2021 time_base_num = avctx->time_base.num;
2022 pkt->pts = pts.value / time_base_num;
2023 pkt->dts = dts.value / time_base_num - dts_delta;
2024 pkt->size = out_buf_size;
2025
2026 return 0;
2027 }
2028
2029 /*
2030 * contiguous_buf_size is 0 if not contiguous, and the size of the buffer
2031 * containing all planes if so.
2032 */
get_cv_pixel_info(AVCodecContext * avctx,const AVFrame * frame,int * color,int * plane_count,size_t * widths,size_t * heights,size_t * strides,size_t * contiguous_buf_size)2033 static int get_cv_pixel_info(
2034 AVCodecContext *avctx,
2035 const AVFrame *frame,
2036 int *color,
2037 int *plane_count,
2038 size_t *widths,
2039 size_t *heights,
2040 size_t *strides,
2041 size_t *contiguous_buf_size)
2042 {
2043 VTEncContext *vtctx = avctx->priv_data;
2044 int av_format = frame->format;
2045 int av_color_range = frame->color_range;
2046 int i;
2047 int range_guessed;
2048 int status;
2049
2050 status = get_cv_pixel_format(avctx, av_format, av_color_range, color, &range_guessed);
2051 if (status) {
2052 av_log(avctx,
2053 AV_LOG_ERROR,
2054 "Could not get pixel format for color format '%s' range '%s'.\n",
2055 av_get_pix_fmt_name(av_format),
2056 av_color_range > AVCOL_RANGE_UNSPECIFIED &&
2057 av_color_range < AVCOL_RANGE_NB ?
2058 av_color_range_name(av_color_range) :
2059 "Unknown");
2060
2061 return AVERROR(EINVAL);
2062 }
2063
2064 if (range_guessed) {
2065 if (!vtctx->warned_color_range) {
2066 vtctx->warned_color_range = true;
2067 av_log(avctx,
2068 AV_LOG_WARNING,
2069 "Color range not set for %s. Using MPEG range.\n",
2070 av_get_pix_fmt_name(av_format));
2071 }
2072 }
2073
2074 switch (av_format) {
2075 case AV_PIX_FMT_NV12:
2076 *plane_count = 2;
2077
2078 widths [0] = avctx->width;
2079 heights[0] = avctx->height;
2080 strides[0] = frame ? frame->linesize[0] : avctx->width;
2081
2082 widths [1] = (avctx->width + 1) / 2;
2083 heights[1] = (avctx->height + 1) / 2;
2084 strides[1] = frame ? frame->linesize[1] : (avctx->width + 1) & -2;
2085 break;
2086
2087 case AV_PIX_FMT_YUV420P:
2088 *plane_count = 3;
2089
2090 widths [0] = avctx->width;
2091 heights[0] = avctx->height;
2092 strides[0] = frame ? frame->linesize[0] : avctx->width;
2093
2094 widths [1] = (avctx->width + 1) / 2;
2095 heights[1] = (avctx->height + 1) / 2;
2096 strides[1] = frame ? frame->linesize[1] : (avctx->width + 1) / 2;
2097
2098 widths [2] = (avctx->width + 1) / 2;
2099 heights[2] = (avctx->height + 1) / 2;
2100 strides[2] = frame ? frame->linesize[2] : (avctx->width + 1) / 2;
2101 break;
2102
2103 case AV_PIX_FMT_BGRA:
2104 *plane_count = 1;
2105
2106 widths [0] = avctx->width;
2107 heights[0] = avctx->height;
2108 strides[0] = frame ? frame->linesize[0] : avctx->width * 4;
2109 break;
2110
2111 case AV_PIX_FMT_P010LE:
2112 *plane_count = 2;
2113 widths[0] = avctx->width;
2114 heights[0] = avctx->height;
2115 strides[0] = frame ? frame->linesize[0] : (avctx->width * 2 + 63) & -64;
2116
2117 widths[1] = (avctx->width + 1) / 2;
2118 heights[1] = (avctx->height + 1) / 2;
2119 strides[1] = frame ? frame->linesize[1] : ((avctx->width + 1) / 2 + 63) & -64;
2120 break;
2121
2122 default:
2123 av_log(
2124 avctx,
2125 AV_LOG_ERROR,
2126 "Could not get frame format info for color %d range %d.\n",
2127 av_format,
2128 av_color_range);
2129
2130 return AVERROR(EINVAL);
2131 }
2132
2133 *contiguous_buf_size = 0;
2134 for (i = 0; i < *plane_count; i++) {
2135 if (i < *plane_count - 1 &&
2136 frame->data[i] + strides[i] * heights[i] != frame->data[i + 1]) {
2137 *contiguous_buf_size = 0;
2138 break;
2139 }
2140
2141 *contiguous_buf_size += strides[i] * heights[i];
2142 }
2143
2144 return 0;
2145 }
2146
2147 //Not used on OSX - frame is never copied.
copy_avframe_to_pixel_buffer(AVCodecContext * avctx,const AVFrame * frame,CVPixelBufferRef cv_img,const size_t * plane_strides,const size_t * plane_rows)2148 static int copy_avframe_to_pixel_buffer(AVCodecContext *avctx,
2149 const AVFrame *frame,
2150 CVPixelBufferRef cv_img,
2151 const size_t *plane_strides,
2152 const size_t *plane_rows)
2153 {
2154 int i, j;
2155 size_t plane_count;
2156 int status;
2157 int rows;
2158 int src_stride;
2159 int dst_stride;
2160 uint8_t *src_addr;
2161 uint8_t *dst_addr;
2162 size_t copy_bytes;
2163
2164 status = CVPixelBufferLockBaseAddress(cv_img, 0);
2165 if (status) {
2166 av_log(
2167 avctx,
2168 AV_LOG_ERROR,
2169 "Error: Could not lock base address of CVPixelBuffer: %d.\n",
2170 status
2171 );
2172 }
2173
2174 if (CVPixelBufferIsPlanar(cv_img)) {
2175 plane_count = CVPixelBufferGetPlaneCount(cv_img);
2176 for (i = 0; frame->data[i]; i++) {
2177 if (i == plane_count) {
2178 CVPixelBufferUnlockBaseAddress(cv_img, 0);
2179 av_log(avctx,
2180 AV_LOG_ERROR,
2181 "Error: different number of planes in AVFrame and CVPixelBuffer.\n"
2182 );
2183
2184 return AVERROR_EXTERNAL;
2185 }
2186
2187 dst_addr = (uint8_t*)CVPixelBufferGetBaseAddressOfPlane(cv_img, i);
2188 src_addr = (uint8_t*)frame->data[i];
2189 dst_stride = CVPixelBufferGetBytesPerRowOfPlane(cv_img, i);
2190 src_stride = plane_strides[i];
2191 rows = plane_rows[i];
2192
2193 if (dst_stride == src_stride) {
2194 memcpy(dst_addr, src_addr, src_stride * rows);
2195 } else {
2196 copy_bytes = dst_stride < src_stride ? dst_stride : src_stride;
2197
2198 for (j = 0; j < rows; j++) {
2199 memcpy(dst_addr + j * dst_stride, src_addr + j * src_stride, copy_bytes);
2200 }
2201 }
2202 }
2203 } else {
2204 if (frame->data[1]) {
2205 CVPixelBufferUnlockBaseAddress(cv_img, 0);
2206 av_log(avctx,
2207 AV_LOG_ERROR,
2208 "Error: different number of planes in AVFrame and non-planar CVPixelBuffer.\n"
2209 );
2210
2211 return AVERROR_EXTERNAL;
2212 }
2213
2214 dst_addr = (uint8_t*)CVPixelBufferGetBaseAddress(cv_img);
2215 src_addr = (uint8_t*)frame->data[0];
2216 dst_stride = CVPixelBufferGetBytesPerRow(cv_img);
2217 src_stride = plane_strides[0];
2218 rows = plane_rows[0];
2219
2220 if (dst_stride == src_stride) {
2221 memcpy(dst_addr, src_addr, src_stride * rows);
2222 } else {
2223 copy_bytes = dst_stride < src_stride ? dst_stride : src_stride;
2224
2225 for (j = 0; j < rows; j++) {
2226 memcpy(dst_addr + j * dst_stride, src_addr + j * src_stride, copy_bytes);
2227 }
2228 }
2229 }
2230
2231 status = CVPixelBufferUnlockBaseAddress(cv_img, 0);
2232 if (status) {
2233 av_log(avctx, AV_LOG_ERROR, "Error: Could not unlock CVPixelBuffer base address: %d.\n", status);
2234 return AVERROR_EXTERNAL;
2235 }
2236
2237 return 0;
2238 }
2239
create_cv_pixel_buffer(AVCodecContext * avctx,const AVFrame * frame,CVPixelBufferRef * cv_img)2240 static int create_cv_pixel_buffer(AVCodecContext *avctx,
2241 const AVFrame *frame,
2242 CVPixelBufferRef *cv_img)
2243 {
2244 int plane_count;
2245 int color;
2246 size_t widths [AV_NUM_DATA_POINTERS];
2247 size_t heights[AV_NUM_DATA_POINTERS];
2248 size_t strides[AV_NUM_DATA_POINTERS];
2249 int status;
2250 size_t contiguous_buf_size;
2251 CVPixelBufferPoolRef pix_buf_pool;
2252 VTEncContext* vtctx = avctx->priv_data;
2253
2254 if (avctx->pix_fmt == AV_PIX_FMT_VIDEOTOOLBOX) {
2255 av_assert0(frame->format == AV_PIX_FMT_VIDEOTOOLBOX);
2256
2257 *cv_img = (CVPixelBufferRef)frame->data[3];
2258 av_assert0(*cv_img);
2259
2260 CFRetain(*cv_img);
2261 return 0;
2262 }
2263
2264 memset(widths, 0, sizeof(widths));
2265 memset(heights, 0, sizeof(heights));
2266 memset(strides, 0, sizeof(strides));
2267
2268 status = get_cv_pixel_info(
2269 avctx,
2270 frame,
2271 &color,
2272 &plane_count,
2273 widths,
2274 heights,
2275 strides,
2276 &contiguous_buf_size
2277 );
2278
2279 if (status) {
2280 av_log(
2281 avctx,
2282 AV_LOG_ERROR,
2283 "Error: Cannot convert format %d color_range %d: %d\n",
2284 frame->format,
2285 frame->color_range,
2286 status
2287 );
2288
2289 return AVERROR_EXTERNAL;
2290 }
2291
2292 pix_buf_pool = VTCompressionSessionGetPixelBufferPool(vtctx->session);
2293 if (!pix_buf_pool) {
2294 /* On iOS, the VT session is invalidated when the APP switches from
2295 * foreground to background and vice versa. Fetch the actual error code
2296 * of the VT session to detect that case and restart the VT session
2297 * accordingly. */
2298 OSStatus vtstatus;
2299
2300 vtstatus = VTCompressionSessionPrepareToEncodeFrames(vtctx->session);
2301 if (vtstatus == kVTInvalidSessionErr) {
2302 CFRelease(vtctx->session);
2303 vtctx->session = NULL;
2304 status = vtenc_configure_encoder(avctx);
2305 if (status == 0)
2306 pix_buf_pool = VTCompressionSessionGetPixelBufferPool(vtctx->session);
2307 }
2308 if (!pix_buf_pool) {
2309 av_log(avctx, AV_LOG_ERROR, "Could not get pixel buffer pool.\n");
2310 return AVERROR_EXTERNAL;
2311 }
2312 else
2313 av_log(avctx, AV_LOG_WARNING, "VT session restarted because of a "
2314 "kVTInvalidSessionErr error.\n");
2315 }
2316
2317 status = CVPixelBufferPoolCreatePixelBuffer(NULL,
2318 pix_buf_pool,
2319 cv_img);
2320
2321
2322 if (status) {
2323 av_log(avctx, AV_LOG_ERROR, "Could not create pixel buffer from pool: %d.\n", status);
2324 return AVERROR_EXTERNAL;
2325 }
2326
2327 status = copy_avframe_to_pixel_buffer(avctx, frame, *cv_img, strides, heights);
2328 if (status) {
2329 CFRelease(*cv_img);
2330 *cv_img = NULL;
2331 return status;
2332 }
2333
2334 return 0;
2335 }
2336
create_encoder_dict_h264(const AVFrame * frame,CFDictionaryRef * dict_out)2337 static int create_encoder_dict_h264(const AVFrame *frame,
2338 CFDictionaryRef* dict_out)
2339 {
2340 CFDictionaryRef dict = NULL;
2341 if (frame->pict_type == AV_PICTURE_TYPE_I) {
2342 const void *keys[] = { kVTEncodeFrameOptionKey_ForceKeyFrame };
2343 const void *vals[] = { kCFBooleanTrue };
2344
2345 dict = CFDictionaryCreate(NULL, keys, vals, 1, NULL, NULL);
2346 if(!dict) return AVERROR(ENOMEM);
2347 }
2348
2349 *dict_out = dict;
2350 return 0;
2351 }
2352
vtenc_send_frame(AVCodecContext * avctx,VTEncContext * vtctx,const AVFrame * frame)2353 static int vtenc_send_frame(AVCodecContext *avctx,
2354 VTEncContext *vtctx,
2355 const AVFrame *frame)
2356 {
2357 CMTime time;
2358 CFDictionaryRef frame_dict;
2359 CVPixelBufferRef cv_img = NULL;
2360 AVFrameSideData *side_data = NULL;
2361 ExtraSEI *sei = NULL;
2362 int status = create_cv_pixel_buffer(avctx, frame, &cv_img);
2363
2364 if (status) return status;
2365
2366 status = create_encoder_dict_h264(frame, &frame_dict);
2367 if (status) {
2368 CFRelease(cv_img);
2369 return status;
2370 }
2371
2372 side_data = av_frame_get_side_data(frame, AV_FRAME_DATA_A53_CC);
2373 if (vtctx->a53_cc && side_data && side_data->size) {
2374 sei = av_mallocz(sizeof(*sei));
2375 if (!sei) {
2376 av_log(avctx, AV_LOG_ERROR, "Not enough memory for closed captions, skipping\n");
2377 } else {
2378 int ret = ff_alloc_a53_sei(frame, 0, &sei->data, &sei->size);
2379 if (ret < 0) {
2380 av_log(avctx, AV_LOG_ERROR, "Not enough memory for closed captions, skipping\n");
2381 av_free(sei);
2382 sei = NULL;
2383 }
2384 }
2385 }
2386
2387 time = CMTimeMake(frame->pts * avctx->time_base.num, avctx->time_base.den);
2388 status = VTCompressionSessionEncodeFrame(
2389 vtctx->session,
2390 cv_img,
2391 time,
2392 kCMTimeInvalid,
2393 frame_dict,
2394 sei,
2395 NULL
2396 );
2397
2398 if (frame_dict) CFRelease(frame_dict);
2399 CFRelease(cv_img);
2400
2401 if (status) {
2402 av_log(avctx, AV_LOG_ERROR, "Error: cannot encode frame: %d\n", status);
2403 return AVERROR_EXTERNAL;
2404 }
2405
2406 return 0;
2407 }
2408
vtenc_frame(AVCodecContext * avctx,AVPacket * pkt,const AVFrame * frame,int * got_packet)2409 static av_cold int vtenc_frame(
2410 AVCodecContext *avctx,
2411 AVPacket *pkt,
2412 const AVFrame *frame,
2413 int *got_packet)
2414 {
2415 VTEncContext *vtctx = avctx->priv_data;
2416 bool get_frame;
2417 int status;
2418 CMSampleBufferRef buf = NULL;
2419 ExtraSEI *sei = NULL;
2420
2421 if (frame) {
2422 status = vtenc_send_frame(avctx, vtctx, frame);
2423
2424 if (status) {
2425 status = AVERROR_EXTERNAL;
2426 goto end_nopkt;
2427 }
2428
2429 if (vtctx->frame_ct_in == 0) {
2430 vtctx->first_pts = frame->pts;
2431 } else if(vtctx->frame_ct_in == vtctx->has_b_frames) {
2432 vtctx->dts_delta = frame->pts - vtctx->first_pts;
2433 }
2434
2435 vtctx->frame_ct_in++;
2436 } else if(!vtctx->flushing) {
2437 vtctx->flushing = true;
2438
2439 status = VTCompressionSessionCompleteFrames(vtctx->session,
2440 kCMTimeIndefinite);
2441
2442 if (status) {
2443 av_log(avctx, AV_LOG_ERROR, "Error flushing frames: %d\n", status);
2444 status = AVERROR_EXTERNAL;
2445 goto end_nopkt;
2446 }
2447 }
2448
2449 *got_packet = 0;
2450 get_frame = vtctx->dts_delta >= 0 || !frame;
2451 if (!get_frame) {
2452 status = 0;
2453 goto end_nopkt;
2454 }
2455
2456 status = vtenc_q_pop(vtctx, !frame, &buf, &sei);
2457 if (status) goto end_nopkt;
2458 if (!buf) goto end_nopkt;
2459
2460 status = vtenc_cm_to_avpacket(avctx, buf, pkt, sei);
2461 if (sei) {
2462 if (sei->data) av_free(sei->data);
2463 av_free(sei);
2464 }
2465 CFRelease(buf);
2466 if (status) goto end_nopkt;
2467
2468 *got_packet = 1;
2469 return 0;
2470
2471 end_nopkt:
2472 av_packet_unref(pkt);
2473 return status;
2474 }
2475
vtenc_populate_extradata(AVCodecContext * avctx,CMVideoCodecType codec_type,CFStringRef profile_level,CFNumberRef gamma_level,CFDictionaryRef enc_info,CFDictionaryRef pixel_buffer_info)2476 static int vtenc_populate_extradata(AVCodecContext *avctx,
2477 CMVideoCodecType codec_type,
2478 CFStringRef profile_level,
2479 CFNumberRef gamma_level,
2480 CFDictionaryRef enc_info,
2481 CFDictionaryRef pixel_buffer_info)
2482 {
2483 VTEncContext *vtctx = avctx->priv_data;
2484 int status;
2485 CVPixelBufferPoolRef pool = NULL;
2486 CVPixelBufferRef pix_buf = NULL;
2487 CMTime time;
2488 CMSampleBufferRef buf = NULL;
2489
2490 status = vtenc_create_encoder(avctx,
2491 codec_type,
2492 profile_level,
2493 gamma_level,
2494 enc_info,
2495 pixel_buffer_info,
2496 &vtctx->session);
2497 if (status)
2498 goto pe_cleanup;
2499
2500 pool = VTCompressionSessionGetPixelBufferPool(vtctx->session);
2501 if(!pool){
2502 av_log(avctx, AV_LOG_ERROR, "Error getting pixel buffer pool.\n");
2503 goto pe_cleanup;
2504 }
2505
2506 status = CVPixelBufferPoolCreatePixelBuffer(NULL,
2507 pool,
2508 &pix_buf);
2509
2510 if(status != kCVReturnSuccess){
2511 av_log(avctx, AV_LOG_ERROR, "Error creating frame from pool: %d\n", status);
2512 goto pe_cleanup;
2513 }
2514
2515 time = CMTimeMake(0, avctx->time_base.den);
2516 status = VTCompressionSessionEncodeFrame(vtctx->session,
2517 pix_buf,
2518 time,
2519 kCMTimeInvalid,
2520 NULL,
2521 NULL,
2522 NULL);
2523
2524 if (status) {
2525 av_log(avctx,
2526 AV_LOG_ERROR,
2527 "Error sending frame for extradata: %d\n",
2528 status);
2529
2530 goto pe_cleanup;
2531 }
2532
2533 //Populates extradata - output frames are flushed and param sets are available.
2534 status = VTCompressionSessionCompleteFrames(vtctx->session,
2535 kCMTimeIndefinite);
2536
2537 if (status)
2538 goto pe_cleanup;
2539
2540 status = vtenc_q_pop(vtctx, 0, &buf, NULL);
2541 if (status) {
2542 av_log(avctx, AV_LOG_ERROR, "popping: %d\n", status);
2543 goto pe_cleanup;
2544 }
2545
2546 CFRelease(buf);
2547
2548
2549
2550 pe_cleanup:
2551 if(vtctx->session)
2552 CFRelease(vtctx->session);
2553
2554 vtctx->session = NULL;
2555 vtctx->frame_ct_out = 0;
2556
2557 av_assert0(status != 0 || (avctx->extradata && avctx->extradata_size > 0));
2558
2559 return status;
2560 }
2561
vtenc_close(AVCodecContext * avctx)2562 static av_cold int vtenc_close(AVCodecContext *avctx)
2563 {
2564 VTEncContext *vtctx = avctx->priv_data;
2565
2566 if(!vtctx->session) {
2567 pthread_cond_destroy(&vtctx->cv_sample_sent);
2568 pthread_mutex_destroy(&vtctx->lock);
2569 return 0;
2570 }
2571
2572 VTCompressionSessionCompleteFrames(vtctx->session,
2573 kCMTimeIndefinite);
2574 clear_frame_queue(vtctx);
2575 pthread_cond_destroy(&vtctx->cv_sample_sent);
2576 pthread_mutex_destroy(&vtctx->lock);
2577 CFRelease(vtctx->session);
2578 vtctx->session = NULL;
2579
2580 if (vtctx->color_primaries) {
2581 CFRelease(vtctx->color_primaries);
2582 vtctx->color_primaries = NULL;
2583 }
2584
2585 if (vtctx->transfer_function) {
2586 CFRelease(vtctx->transfer_function);
2587 vtctx->transfer_function = NULL;
2588 }
2589
2590 if (vtctx->ycbcr_matrix) {
2591 CFRelease(vtctx->ycbcr_matrix);
2592 vtctx->ycbcr_matrix = NULL;
2593 }
2594
2595 return 0;
2596 }
2597
2598 static const enum AVPixelFormat avc_pix_fmts[] = {
2599 AV_PIX_FMT_VIDEOTOOLBOX,
2600 AV_PIX_FMT_NV12,
2601 AV_PIX_FMT_YUV420P,
2602 AV_PIX_FMT_NONE
2603 };
2604
2605 static const enum AVPixelFormat hevc_pix_fmts[] = {
2606 AV_PIX_FMT_VIDEOTOOLBOX,
2607 AV_PIX_FMT_NV12,
2608 AV_PIX_FMT_YUV420P,
2609 AV_PIX_FMT_BGRA,
2610 AV_PIX_FMT_P010LE,
2611 AV_PIX_FMT_NONE
2612 };
2613
2614 #define VE AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM
2615 #define COMMON_OPTIONS \
2616 { "allow_sw", "Allow software encoding", OFFSET(allow_sw), AV_OPT_TYPE_BOOL, \
2617 { .i64 = 0 }, 0, 1, VE }, \
2618 { "require_sw", "Require software encoding", OFFSET(require_sw), AV_OPT_TYPE_BOOL, \
2619 { .i64 = 0 }, 0, 1, VE }, \
2620 { "realtime", "Hint that encoding should happen in real-time if not faster (e.g. capturing from camera).", \
2621 OFFSET(realtime), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, VE }, \
2622 { "frames_before", "Other frames will come before the frames in this session. This helps smooth concatenation issues.", \
2623 OFFSET(frames_before), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, VE }, \
2624 { "frames_after", "Other frames will come after the frames in this session. This helps smooth concatenation issues.", \
2625 OFFSET(frames_after), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, VE },
2626
2627 #define OFFSET(x) offsetof(VTEncContext, x)
2628 static const AVOption h264_options[] = {
2629 { "profile", "Profile", OFFSET(profile), AV_OPT_TYPE_INT, { .i64 = H264_PROF_AUTO }, H264_PROF_AUTO, H264_PROF_COUNT, VE, "profile" },
2630 { "baseline", "Baseline Profile", 0, AV_OPT_TYPE_CONST, { .i64 = H264_PROF_BASELINE }, INT_MIN, INT_MAX, VE, "profile" },
2631 { "main", "Main Profile", 0, AV_OPT_TYPE_CONST, { .i64 = H264_PROF_MAIN }, INT_MIN, INT_MAX, VE, "profile" },
2632 { "high", "High Profile", 0, AV_OPT_TYPE_CONST, { .i64 = H264_PROF_HIGH }, INT_MIN, INT_MAX, VE, "profile" },
2633 { "extended", "Extend Profile", 0, AV_OPT_TYPE_CONST, { .i64 = H264_PROF_EXTENDED }, INT_MIN, INT_MAX, VE, "profile" },
2634
2635 { "level", "Level", OFFSET(level), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, 52, VE, "level" },
2636 { "1.3", "Level 1.3, only available with Baseline Profile", 0, AV_OPT_TYPE_CONST, { .i64 = 13 }, INT_MIN, INT_MAX, VE, "level" },
2637 { "3.0", "Level 3.0", 0, AV_OPT_TYPE_CONST, { .i64 = 30 }, INT_MIN, INT_MAX, VE, "level" },
2638 { "3.1", "Level 3.1", 0, AV_OPT_TYPE_CONST, { .i64 = 31 }, INT_MIN, INT_MAX, VE, "level" },
2639 { "3.2", "Level 3.2", 0, AV_OPT_TYPE_CONST, { .i64 = 32 }, INT_MIN, INT_MAX, VE, "level" },
2640 { "4.0", "Level 4.0", 0, AV_OPT_TYPE_CONST, { .i64 = 40 }, INT_MIN, INT_MAX, VE, "level" },
2641 { "4.1", "Level 4.1", 0, AV_OPT_TYPE_CONST, { .i64 = 41 }, INT_MIN, INT_MAX, VE, "level" },
2642 { "4.2", "Level 4.2", 0, AV_OPT_TYPE_CONST, { .i64 = 42 }, INT_MIN, INT_MAX, VE, "level" },
2643 { "5.0", "Level 5.0", 0, AV_OPT_TYPE_CONST, { .i64 = 50 }, INT_MIN, INT_MAX, VE, "level" },
2644 { "5.1", "Level 5.1", 0, AV_OPT_TYPE_CONST, { .i64 = 51 }, INT_MIN, INT_MAX, VE, "level" },
2645 { "5.2", "Level 5.2", 0, AV_OPT_TYPE_CONST, { .i64 = 52 }, INT_MIN, INT_MAX, VE, "level" },
2646
2647 { "coder", "Entropy coding", OFFSET(entropy), AV_OPT_TYPE_INT, { .i64 = VT_ENTROPY_NOT_SET }, VT_ENTROPY_NOT_SET, VT_CABAC, VE, "coder" },
2648 { "cavlc", "CAVLC entropy coding", 0, AV_OPT_TYPE_CONST, { .i64 = VT_CAVLC }, INT_MIN, INT_MAX, VE, "coder" },
2649 { "vlc", "CAVLC entropy coding", 0, AV_OPT_TYPE_CONST, { .i64 = VT_CAVLC }, INT_MIN, INT_MAX, VE, "coder" },
2650 { "cabac", "CABAC entropy coding", 0, AV_OPT_TYPE_CONST, { .i64 = VT_CABAC }, INT_MIN, INT_MAX, VE, "coder" },
2651 { "ac", "CABAC entropy coding", 0, AV_OPT_TYPE_CONST, { .i64 = VT_CABAC }, INT_MIN, INT_MAX, VE, "coder" },
2652
2653 { "a53cc", "Use A53 Closed Captions (if available)", OFFSET(a53_cc), AV_OPT_TYPE_BOOL, {.i64 = 1}, 0, 1, VE },
2654
2655 COMMON_OPTIONS
2656 { NULL },
2657 };
2658
2659 static const AVClass h264_videotoolbox_class = {
2660 .class_name = "h264_videotoolbox",
2661 .item_name = av_default_item_name,
2662 .option = h264_options,
2663 .version = LIBAVUTIL_VERSION_INT,
2664 };
2665
2666 AVCodec ff_h264_videotoolbox_encoder = {
2667 .name = "h264_videotoolbox",
2668 .long_name = NULL_IF_CONFIG_SMALL("VideoToolbox H.264 Encoder"),
2669 .type = AVMEDIA_TYPE_VIDEO,
2670 .id = AV_CODEC_ID_H264,
2671 .priv_data_size = sizeof(VTEncContext),
2672 .pix_fmts = avc_pix_fmts,
2673 .init = vtenc_init,
2674 .encode2 = vtenc_frame,
2675 .close = vtenc_close,
2676 .capabilities = AV_CODEC_CAP_DELAY,
2677 .priv_class = &h264_videotoolbox_class,
2678 .caps_internal = FF_CODEC_CAP_INIT_THREADSAFE |
2679 FF_CODEC_CAP_INIT_CLEANUP,
2680 };
2681
2682 static const AVOption hevc_options[] = {
2683 { "profile", "Profile", OFFSET(profile), AV_OPT_TYPE_INT, { .i64 = HEVC_PROF_AUTO }, HEVC_PROF_AUTO, HEVC_PROF_COUNT, VE, "profile" },
2684 { "main", "Main Profile", 0, AV_OPT_TYPE_CONST, { .i64 = HEVC_PROF_MAIN }, INT_MIN, INT_MAX, VE, "profile" },
2685 { "main10", "Main10 Profile", 0, AV_OPT_TYPE_CONST, { .i64 = HEVC_PROF_MAIN10 }, INT_MIN, INT_MAX, VE, "profile" },
2686
2687 { "alpha_quality", "Compression quality for the alpha channel", OFFSET(alpha_quality), AV_OPT_TYPE_DOUBLE, { .dbl = 0.0 }, 0.0, 1.0, VE },
2688
2689 COMMON_OPTIONS
2690 { NULL },
2691 };
2692
2693 static const AVClass hevc_videotoolbox_class = {
2694 .class_name = "hevc_videotoolbox",
2695 .item_name = av_default_item_name,
2696 .option = hevc_options,
2697 .version = LIBAVUTIL_VERSION_INT,
2698 };
2699
2700 AVCodec ff_hevc_videotoolbox_encoder = {
2701 .name = "hevc_videotoolbox",
2702 .long_name = NULL_IF_CONFIG_SMALL("VideoToolbox H.265 Encoder"),
2703 .type = AVMEDIA_TYPE_VIDEO,
2704 .id = AV_CODEC_ID_HEVC,
2705 .priv_data_size = sizeof(VTEncContext),
2706 .pix_fmts = hevc_pix_fmts,
2707 .init = vtenc_init,
2708 .encode2 = vtenc_frame,
2709 .close = vtenc_close,
2710 .capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_HARDWARE,
2711 .priv_class = &hevc_videotoolbox_class,
2712 .caps_internal = FF_CODEC_CAP_INIT_THREADSAFE |
2713 FF_CODEC_CAP_INIT_CLEANUP,
2714 .wrapper_name = "videotoolbox",
2715 };
2716