1 /*
2 * AVFoundation input device
3 * Copyright (c) 2014 Thilo Borgmann <thilo.borgmann@mail.de>
4 *
5 * This file is part of FFmpeg.
6 *
7 * FFmpeg is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2.1 of the License, or (at your option) any later version.
11 *
12 * FFmpeg is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
16 *
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with FFmpeg; if not, write to the Free Software
19 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
20 */
21
22 /**
23 * @file
24 * AVFoundation input device
25 * @author Thilo Borgmann <thilo.borgmann@mail.de>
26 */
27
28 #import <AVFoundation/AVFoundation.h>
29 #include <pthread.h>
30
39
41
45 };
46
50 };
51
75 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1080
77 #endif
79 };
80
82 {
84
93
96
99
105
108
110
118
121
123
130
132 {
134 }
135
137 {
139 }
140
141 /** FrameReciever class - delegate for AVCaptureSession
142 */
144 {
146 }
147
149
150 - (
void) captureOutput:(AVCaptureOutput *)captureOutput
151 didOutputSampleBuffer:(CMSampleBufferRef)videoFrame
152 fromConnection:(AVCaptureConnection *)connection;
153
154 @end
155
157
159 {
160 if (
self = [super
init]) {
162 }
163 return self;
164 }
165
166 - (
void) captureOutput:(AVCaptureOutput *)captureOutput
167 didOutputSampleBuffer:(CMSampleBufferRef)videoFrame
168 fromConnection:(AVCaptureConnection *)connection
169 {
171
174 }
175
177
179
181
183 }
184
185 @end
186
187 /** AudioReciever class - delegate for AVCaptureSession
188 */
190 {
192 }
193
195
196 - (
void) captureOutput:(AVCaptureOutput *)captureOutput
197 didOutputSampleBuffer:(CMSampleBufferRef)audioFrame
198 fromConnection:(AVCaptureConnection *)connection;
199
200 @end
201
203
205 {
206 if (
self = [super
init]) {
208 }
209 return self;
210 }
211
212 - (
void) captureOutput:(AVCaptureOutput *)captureOutput
213 didOutputSampleBuffer:(CMSampleBufferRef)audioFrame
214 fromConnection:(AVCaptureConnection *)connection
215 {
217
220 }
221
223
225
227
229 }
230
231 @end
232
234 {
235 [ctx->capture_session stopRunning];
236
237 [ctx->capture_session release];
238 [ctx->video_output release];
239 [ctx->audio_output release];
240 [ctx->avf_delegate release];
241 [ctx->avf_audio_delegate release];
242
248
250
253
256 }
257 }
258
260 {
263 char *save;
264
265 if (tmp[0] != ':') {
268 } else {
270 }
271 }
272
273 /**
274 * Configure the video device.
275 *
276 * Configure the video device using a run-time approach to access properties
277 * since formats, activeFormat are available since iOS >= 7.0 or OSX >= 10.7
278 * and activeVideoMaxFrameDuration is available since i0S >= 7.0 and OSX >= 10.9.
279 *
280 * The NSUndefinedKeyException must be handled by the caller of this function.
281 *
282 */
284 {
286
288 NSObject *range = nil;
290 NSObject *selected_range = nil;
291 NSObject *selected_format = nil;
292
293 for (format
in [video_device valueForKey:
@"formats"]) {
294 CMFormatDescriptionRef formatDescription;
295 CMVideoDimensions dimensions;
296
297 formatDescription = (CMFormatDescriptionRef) [format performSelector:@selector(formatDescription)];
298 dimensions = CMVideoFormatDescriptionGetDimensions(formatDescription);
299
301 (dimensions.width == ctx->
width && dimensions.height == ctx->
height)) {
302
304
305 for (range
in [format valueForKey:
@"videoSupportedFrameRateRanges"]) {
306 double max_framerate;
307
308 [[range valueForKey:@"maxFrameRate"] getValue:&max_framerate];
309 if (fabs (framerate - max_framerate) < 0.01) {
310 selected_range = range;
311 break;
312 }
313 }
314 }
315 }
316
317 if (!selected_format) {
318 av_log(s,
AV_LOG_ERROR,
"Selected video size (%dx%d) is not supported by the device\n",
320 goto unsupported_format;
321 }
322
323 if (!selected_range) {
325 framerate);
326 goto unsupported_format;
327 }
328
329 if ([video_device lockForConfiguration:
NULL] == YES) {
330 NSValue *min_frame_duration = [selected_range valueForKey:@"minFrameDuration"];
331
332 [video_device setValue:selected_format forKey:@"activeFormat"];
333 [video_device setValue:min_frame_duration forKey:@"activeVideoMinFrameDuration"];
334 [video_device setValue:min_frame_duration forKey:@"activeVideoMaxFrameDuration"];
335 } else {
338 }
339
340 return 0;
341
342 unsupported_format:
343
345 for (format
in [video_device valueForKey:
@"formats"]) {
346 CMFormatDescriptionRef formatDescription;
347 CMVideoDimensions dimensions;
348
349 formatDescription = (CMFormatDescriptionRef) [format performSelector:@selector(formatDescription)];
350 dimensions = CMVideoFormatDescriptionGetDimensions(formatDescription);
351
352 for (range
in [format valueForKey:
@"videoSupportedFrameRateRanges"]) {
353 double min_framerate;
354 double max_framerate;
355
356 [[range valueForKey:@"minFrameRate"] getValue:&min_framerate];
357 [[range valueForKey:@"maxFrameRate"] getValue:&max_framerate];
359 dimensions.width, dimensions.height,
360 min_framerate, max_framerate);
361 }
362 }
364 }
365
367 {
369 int ret;
370 NSError *error = nil;
371 AVCaptureInput* capture_input = nil;
373 NSNumber *pixel_format;
374 NSDictionary *capture_dict;
375 dispatch_queue_t queue;
376
378 capture_input = (AVCaptureInput*) [[[AVCaptureDeviceInput alloc] initWithDevice:video_device error:&error] autorelease];
379 } else {
380 capture_input = (AVCaptureInput*) video_device;
381 }
382
383 if (!capture_input) {
385 [[error localizedDescription] UTF8String]);
386 return 1;
387 }
388
390 [ctx->capture_session addInput:capture_input];
391 } else {
393 return 1;
394 }
395
396 // Attaching output
397 ctx->
video_output = [[AVCaptureVideoDataOutput alloc] init];
398
401 return 1;
402 }
403
404 // Configure device framerate and video size
405 @try {
407 return ret;
408 }
409 } @catch (NSException *exception) {
410 if (![[exception
name] isEqualToString:NSUndefinedKeyException]) {
413 }
414 }
415
416 // select pixel format
418
421 pxl_fmt_spec = avf_pixel_formats[i];
422 break;
423 }
424 }
425
426 // check if selected pixel format is supported by AVFoundation
428 av_log(s,
AV_LOG_ERROR,
"Selected pixel format (%s) is not supported by AVFoundation.\n",
430 return 1;
431 }
432
433 // check if the pixel format is available for this device
434 if ([[ctx->
video_output availableVideoCVPixelFormatTypes] indexOfObject:[NSNumber numberWithInt:pxl_fmt_spec.avf_id]] == NSNotFound) {
435 av_log(s,
AV_LOG_ERROR,
"Selected pixel format (%s) is not supported by the input device.\n",
437
439
441 for (NSNumber *pxl_fmt
in [ctx->
video_output availableVideoCVPixelFormatTypes]) {
445 if ([pxl_fmt intValue] == avf_pixel_formats[i].
avf_id) {
446 pxl_fmt_dummy = avf_pixel_formats[i];
447 break;
448 }
449 }
450
453
454 // select first supported pixel format instead of user selected (or default) pixel format
456 pxl_fmt_spec = pxl_fmt_dummy;
457 }
458 }
459 }
460
461 // fail if there is no appropriate pixel format or print a warning about overriding the pixel format
463 return 1;
464 } else {
467 }
468 }
469
471 pixel_format = [NSNumber numberWithUnsignedInt:pxl_fmt_spec.avf_id];
472 capture_dict = [NSDictionary dictionaryWithObject:pixel_format
473 forKey:(id)kCVPixelBufferPixelFormatTypeKey];
474
475 [ctx->video_output setVideoSettings:capture_dict];
476 [ctx->video_output setAlwaysDiscardsLateVideoFrames:YES];
477
479
480 queue = dispatch_queue_create(
"avf_queue",
NULL);
481 [ctx->video_output setSampleBufferDelegate:ctx->avf_delegate queue:queue];
482 dispatch_release(queue);
483
485 [ctx->capture_session addOutput:ctx->video_output];
486 } else {
488 return 1;
489 }
490
491 return 0;
492 }
493
495 {
497 NSError *error = nil;
498 AVCaptureDeviceInput* audio_dev_input = [[[AVCaptureDeviceInput alloc] initWithDevice:audio_device error:&error] autorelease];
499 dispatch_queue_t queue;
500
501 if (!audio_dev_input) {
503 [[error localizedDescription] UTF8String]);
504 return 1;
505 }
506
508 [ctx->capture_session addInput:audio_dev_input];
509 } else {
511 return 1;
512 }
513
514 // Attaching output
515 ctx->
audio_output = [[AVCaptureAudioDataOutput alloc] init];
516
519 return 1;
520 }
521
523
524 queue = dispatch_queue_create(
"avf_audio_queue",
NULL);
525 [ctx->audio_output setSampleBufferDelegate:ctx->avf_audio_delegate queue:queue];
526 dispatch_release(queue);
527
529 [ctx->capture_session addOutput:ctx->audio_output];
530 } else {
532 return 1;
533 }
534
535 return 0;
536 }
537
539 {
541 CVImageBufferRef image_buffer;
542 CGSize image_buffer_size;
544
545 if (!stream) {
546 return 1;
547 }
548
549 // Take stream info from the first frame.
551 CFRunLoopRunInMode(kCFRunLoopDefaultMode, 0.1, YES);
552 }
553
555
557
559
560 image_buffer = CMSampleBufferGetImageBuffer(ctx->
current_frame);
561 image_buffer_size = CVImageBufferGetEncodedSize(image_buffer);
562
565 stream->codec->width = (int)image_buffer_size.width;
566 stream->codec->height = (int)image_buffer_size.height;
568
571
573
574 return 0;
575 }
576
578 {
580 CMFormatDescriptionRef format_desc;
582
583 if (!stream) {
584 return 1;
585 }
586
587 // Take stream info from the first frame.
589 CFRunLoopRunInMode(kCFRunLoopDefaultMode, 0.1, YES);
590 }
591
593
595
597
599 const AudioStreamBasicDescription *basic_desc = CMAudioFormatDescriptionGetStreamBasicDescription(format_desc);
600
601 if (!basic_desc) {
603 return 1;
604 }
605
607 stream->codec->sample_rate = basic_desc->mSampleRate;
608 stream->codec->channels = basic_desc->mChannelsPerFrame;
610
613 ctx->
audio_float = basic_desc->mFormatFlags & kAudioFormatFlagIsFloat;
614 ctx->
audio_be = basic_desc->mFormatFlags & kAudioFormatFlagIsBigEndian;
616 ctx->
audio_packed = basic_desc->mFormatFlags & kAudioFormatFlagIsPacked;
618
619 if (basic_desc->mFormatID == kAudioFormatLinearPCM &&
624 } else if (basic_desc->mFormatID == kAudioFormatLinearPCM &&
629 } else if (basic_desc->mFormatID == kAudioFormatLinearPCM &&
634 } else if (basic_desc->mFormatID == kAudioFormatLinearPCM &&
639 } else {
641 return 1;
642 }
643
650 return 1;
651 }
652 }
653
656
658
659 return 0;
660 }
661
663 {
664 NSAutoreleasePool *pool = [[NSAutoreleasePool alloc] init];
666 uint32_t num_screens = 0;
668 AVCaptureDevice *video_device = nil;
669 AVCaptureDevice *audio_device = nil;
670 // Find capture device
671 NSArray *devices = [AVCaptureDevice devicesWithMediaType:AVMediaTypeVideo];
673
676
679
680 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1070
681 CGGetActiveDisplayList(0,
NULL, &num_screens);
682 #endif
683
684 // List devices if requested
688 for (AVCaptureDevice *device
in devices) {
689 const char *
name = [[device localizedName] UTF8String];
690 index = [devices indexOfObject:device];
692 index++;
693 }
694 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1070
695 if (num_screens > 0) {
696 CGDirectDisplayID screens[num_screens];
697 CGGetActiveDisplayList(num_screens, screens, &num_screens);
698 for (int i = 0; i < num_screens; i++) {
700 }
701 }
702 #endif
703
705 devices = [AVCaptureDevice devicesWithMediaType:AVMediaTypeAudio];
706 for (AVCaptureDevice *device
in devices) {
707 const char *
name = [[device localizedName] UTF8String];
708 int index = [devices indexOfObject:device];
710 }
712 }
713
714 // parse input filename for video and audio device
716
717 // check for device index given in filename
720 }
723 }
724
727 video_device = [devices objectAtIndex:ctx->video_device_index];
729 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1070
730 CGDirectDisplayID screens[num_screens];
731 CGGetActiveDisplayList(num_screens, screens, &num_screens);
732 AVCaptureScreenInput* capture_screen_input = [[[AVCaptureScreenInput alloc] initWithDisplayID:screens[ctx->video_device_index - ctx->
num_video_devices]] autorelease];
733
736 }
737
738 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1080
740 capture_screen_input.capturesCursor = YES;
741 } else {
742 capture_screen_input.capturesCursor = NO;
743 }
744 #endif
745
747 capture_screen_input.capturesMouseClicks = YES;
748 } else {
749 capture_screen_input.capturesMouseClicks = NO;
750 }
751
752 video_device = (AVCaptureDevice*) capture_screen_input;
753 capture_screen = 1;
754 #endif
755 } else {
758 }
762 video_device = [AVCaptureDevice defaultDeviceWithMediaType:AVMediaTypeVideo];
763 } else {
764 // looking for video inputs
765 for (AVCaptureDevice *device
in devices) {
767 video_device = device;
768 break;
769 }
770 }
771
772 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1070
773 // looking for screen inputs
774 if (!video_device) {
775 int idx;
776 if(sscanf(ctx->
video_filename,
"Capture screen %d", &idx) && idx < num_screens) {
777 CGDirectDisplayID screens[num_screens];
778 CGGetActiveDisplayList(num_screens, screens, &num_screens);
779 AVCaptureScreenInput* capture_screen_input = [[[AVCaptureScreenInput alloc] initWithDisplayID:screens[idx]] autorelease];
780 video_device = (AVCaptureDevice*) capture_screen_input;
782 capture_screen = 1;
783
786 }
787
788 #if !TARGET_OS_IPHONE && __MAC_OS_X_VERSION_MIN_REQUIRED >= 1080
790 capture_screen_input.capturesCursor = YES;
791 } else {
792 capture_screen_input.capturesCursor = NO;
793 }
794 #endif
795
797 capture_screen_input.capturesMouseClicks = YES;
798 } else {
799 capture_screen_input.capturesMouseClicks = NO;
800 }
801 }
802 }
803 #endif
804 }
805
806 if (!video_device) {
809 }
810 }
811
812 // get audio device
814 NSArray *devices = [AVCaptureDevice devicesWithMediaType:AVMediaTypeAudio];
815
819 }
820
821 audio_device = [devices objectAtIndex:ctx->audio_device_index];
825 audio_device = [AVCaptureDevice defaultDeviceWithMediaType:AVMediaTypeAudio];
826 } else {
827 NSArray *devices = [AVCaptureDevice devicesWithMediaType:AVMediaTypeAudio];
828
829 for (AVCaptureDevice *device
in devices) {
831 audio_device = device;
832 break;
833 }
834 }
835 }
836
837 if (!audio_device) {
840 }
841 }
842
843 // Video nor Audio capture device not found, looking for AVMediaTypeVideo/Audio
844 if (!video_device && !audio_device) {
847 }
848
849 if (video_device) {
852 } else {
854 }
855 }
856 if (audio_device) {
857 av_log(s,
AV_LOG_DEBUG,
"audio device '%s' opened\n", [[audio_device localizedName] UTF8String]);
858 }
859
860 // Initialize capture session
862
865 }
867 }
868
869 [ctx->capture_session startRunning];
870
871 /* Unlock device configuration only after the session is started so it
872 * does not reset the capture formats */
873 if (!capture_screen) {
874 [video_device unlockForConfiguration];
875 }
876
879 }
880
881 // set audio stream
884 }
885
886 [pool release];
887 return 0;
888
890 [pool release];
893 }
894
896 {
898
899 do {
900 CVImageBufferRef image_buffer;
902
903 image_buffer = CMSampleBufferGetImageBuffer(ctx->
current_frame);
904
907 if (
av_new_packet(pkt, (
int)CVPixelBufferGetDataSize(image_buffer)) < 0) {
909 }
910
912 CMSampleTimingInfo timing_info;
913
914 if (CMSampleBufferGetOutputSampleTimingInfoArray(ctx->
current_frame, 1, &timing_info, &count) == noErr) {
916 pkt->
pts = pkt->
dts =
av_rescale_q(timing_info.presentationTimeStamp.value, timebase_q, avf_time_base_q);
917 }
918
921
922 CVPixelBufferLockBaseAddress(image_buffer, 0);
923
924 data = CVPixelBufferGetBaseAddress(image_buffer);
926
927 CVPixelBufferUnlockBaseAddress(image_buffer, 0);
932 int block_buffer_size = CMBlockBufferGetDataLength(block_buffer);
933
934 if (!block_buffer || !block_buffer_size) {
936 }
937
940 }
941
944 }
945
947 CMSampleTimingInfo timing_info;
948
949 if (CMSampleBufferGetOutputSampleTimingInfoArray(ctx->
current_audio_frame, 1, &timing_info, &count) == noErr) {
951 pkt->
pts = pkt->
dts =
av_rescale_q(timing_info.presentationTimeStamp.value, timebase_q, avf_time_base_q);
952 }
953
956
959
960 OSStatus ret = CMBlockBufferCopyDataBytes(block_buffer, 0, pkt->
size, ctx->
audio_buffer);
961 if (ret != kCMBlockBufferNoErr) {
963 }
964
966
967 // transform decoded frame into output format
968 #define INTERLEAVE_OUTPUT(bps) \
969 { \
970 int##bps##_t **src; \
971 int##bps##_t *dest; \
972 src = av_malloc(ctx->audio_channels * sizeof(int##bps##_t*)); \
973 if (!src) return AVERROR(EIO); \
974 for (c = 0; c < ctx->audio_channels; c++) { \
975 src[c] = ((int##bps##_t*)ctx->audio_buffer) + c * num_samples; \
976 } \
977 dest = (int##bps##_t*)pkt->data; \
978 shift = bps - ctx->audio_bits_per_sample; \
979 for (sample = 0; sample < num_samples; sample++) \
980 for (c = 0; c < ctx->audio_channels; c++) \
981 *dest++ = src[c][sample] << shift; \
982 av_freep(&src); \
983 }
984
987 } else {
989 }
990 } else {
991 OSStatus ret = CMBlockBufferCopyDataBytes(block_buffer, 0, pkt->
size, pkt->
data);
992 if (ret != kCMBlockBufferNoErr) {
994 }
995 }
996
999 } else {
1002 }
1003
1006
1007 return 0;
1008 }
1009
1011 {
1014 return 0;
1015 }
1016
1018 {
"list_devices",
"list available devices", offsetof(
AVFContext, list_devices),
AV_OPT_TYPE_INT, {.i64=0}, 0, 1,
AV_OPT_FLAG_DECODING_PARAM,
"list_devices" },
1028
1030 };
1031
1038 };
1039
1041 .
name =
"avfoundation",
1048 .priv_class = &avf_class,
1049 };
packed YUV 4:2:2, 16bpp, Cb Y0 Cr Y1
static const struct AVFPixelFormatSpec avf_pixel_formats[]
static void lock_frames(AVFContext *ctx)
static int shift(int a, int b)
static av_always_inline int pthread_mutex_destroy(pthread_mutex_t *mutex)
AVCaptureSession * capture_session
static av_always_inline int pthread_cond_wait(pthread_cond_t *cond, pthread_mutex_t *mutex)
ptrdiff_t const GLvoid * data
planar YUV 4:4:4, 24bpp, (1 Cr & Cb sample per 1x1 Y samples)
#define AV_LOG_WARNING
Something somehow does not look correct.
#define LIBAVUTIL_VERSION_INT
packed RGB 8:8:8, 24bpp, RGBRGB...
static av_cold int init(AVCodecContext *avctx)
int audio_frames_captured
static const AVOption options[]
packed RGB 5:5:5, 16bpp, (msb)1X 5R 5G 5B(lsb), little-endian, X=unused/undefined ...
static int add_video_device(AVFormatContext *s, AVCaptureDevice *video_device)
CMSampleBufferRef current_audio_frame
static int get_audio_config(AVFormatContext *s)
int audio_non_interleaved
packed BGR 8:8:8, 32bpp, XBGRXBGR... X=unused/undefined
static AVRational av_make_q(int num, int den)
Create a rational.
AudioReciever class - delegate for AVCaptureSession.
static av_always_inline int pthread_cond_destroy(pthread_cond_t *cond)
packed RGB 5:6:5, 16bpp, (msb) 5R 6G 5B(lsb), little-endian
const char * class_name
The name of the class; usually it is the same name as the context structure type to which the AVClass...
packed RGB 8:8:8, 32bpp, RGBXRGBX... X=unused/undefined
static int configure_video_device(AVFormatContext *s, AVCaptureDevice *video_device)
Configure the video device.
AVStream * avformat_new_stream(AVFormatContext *s, const AVCodec *c)
Add a new stream to a media file.
packed RGB 5:6:5, 16bpp, (msb) 5R 6G 5B(lsb), big-endian
static double av_q2d(AVRational a)
Convert rational to double.
static av_cold int read_close(AVFormatContext *ctx)
static av_always_inline int pthread_cond_signal(pthread_cond_t *cond)
static void parse_device_name(AVFormatContext *s)
#define AV_PKT_FLAG_KEY
The packet contains a keyframe.
AVInputFormat ff_avfoundation_demuxer
int64_t av_rescale_q(int64_t a, AVRational bq, AVRational cq)
Rescale a 64-bit integer by 2 rational numbers.
Main libavdevice API header.
int av_new_packet(AVPacket *pkt, int size)
Allocate the payload of a packet and initialize its fields with default values.
#define AV_LOG_ERROR
Something went wrong and cannot losslessly be recovered.
#define NULL_IF_CONFIG_SMALL(x)
Return NULL if CONFIG_SMALL is true, otherwise the argument without modification. ...
#define AV_LOG_DEBUG
Stuff which is only useful for libav* developers.
planar YUV 4:2:0, 12bpp, 1 plane for Y and 1 plane for the UV components, which are interleaved (firs...
static int avf_read_header(AVFormatContext *s)
#define AV_PIX_FMT_YUV444P10
int flags
A combination of AV_PKT_FLAG values.
common internal API header
char filename[1024]
input or output filename
#define AVERROR_BUFFER_TOO_SMALL
Buffer too small.
typedef void(APIENTRY *FF_PFNGLACTIVETEXTUREPROC)(GLenum texture)
static const int avf_time_base
static int capture_screen(CCaptionSubContext *ctx)
static av_always_inline int pthread_mutex_init(pthread_mutex_t *mutex, const pthread_mutexattr_t *attr)
packed RGB 8:8:8, 24bpp, BGRBGR...
int audio_bits_per_sample
packed RGB 16:16:16, 48bpp, 16B, 16G, 16R, the 2-byte value for each R/G/B component is stored as big...
static int read_header(FFV1Context *f)
int64_t av_gettime(void)
Get the current time in microseconds.
static int read_packet(void *opaque, uint8_t *buf, int buf_size)
static int add_audio_device(AVFormatContext *s, AVCaptureDevice *audio_device)
#define AV_LOG_INFO
Standard information.
char * av_strdup(const char *s)
Duplicate the string s.
planar YUV 4:4:4 32bpp, (1 Cr & Cb sample per 1x1 Y & A samples)
static const char * format
static const AVRational avf_time_base_q
packed YUV 4:2:2, 16bpp, Y0 Cb Y1 Cr
AVCaptureVideoDataOutput * video_output
uint8_t pi<< 24) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_U8, uint8_t,(*(constuint8_t *) pi-0x80)*(1.0f/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_U8, uint8_t,(*(constuint8_t *) pi-0x80)*(1.0/(1<< 7))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S16, int16_t,(*(constint16_t *) pi >>8)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S16, int16_t,*(constint16_t *) pi *(1.0f/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S16, int16_t,*(constint16_t *) pi *(1.0/(1<< 15))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_S32, int32_t,(*(constint32_t *) pi >>24)+0x80) CONV_FUNC_GROUP(AV_SAMPLE_FMT_FLT, float, AV_SAMPLE_FMT_S32, int32_t,*(constint32_t *) pi *(1.0f/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_DBL, double, AV_SAMPLE_FMT_S32, int32_t,*(constint32_t *) pi *(1.0/(1U<< 31))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_FLT, float, av_clip_uint8(lrintf(*(constfloat *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_FLT, float, av_clip_int16(lrintf(*(constfloat *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_FLT, float, av_clipl_int32(llrintf(*(constfloat *) pi *(1U<< 31)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_U8, uint8_t, AV_SAMPLE_FMT_DBL, double, av_clip_uint8(lrint(*(constdouble *) pi *(1<< 7))+0x80)) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S16, int16_t, AV_SAMPLE_FMT_DBL, double, av_clip_int16(lrint(*(constdouble *) pi *(1<< 15)))) CONV_FUNC_GROUP(AV_SAMPLE_FMT_S32, int32_t, AV_SAMPLE_FMT_DBL, double, av_clipl_int32(llrint(*(constdouble *) pi *(1U<< 31))))#defineSET_CONV_FUNC_GROUP(ofmt, ifmt) staticvoidset_generic_function(AudioConvert *ac){}voidff_audio_convert_free(AudioConvert **ac){if(!*ac) return;ff_dither_free(&(*ac) ->dc);av_freep(ac);}AudioConvert *ff_audio_convert_alloc(AVAudioResampleContext *avr, enumAVSampleFormatout_fmt, enumAVSampleFormatin_fmt, intchannels, intsample_rate, intapply_map){AudioConvert *ac;intin_planar, out_planar;ac=av_mallocz(sizeof(*ac));if(!ac) returnNULL;ac->avr=avr;ac->out_fmt=out_fmt;ac->in_fmt=in_fmt;ac->channels=channels;ac->apply_map=apply_map;if(avr->dither_method!=AV_RESAMPLE_DITHER_NONE &&av_get_packed_sample_fmt(out_fmt)==AV_SAMPLE_FMT_S16 &&av_get_bytes_per_sample(in_fmt)>2){ac->dc=ff_dither_alloc(avr, out_fmt, in_fmt, channels, sample_rate, apply_map);if(!ac->dc){av_free(ac);returnNULL;}returnac;}in_planar=ff_sample_fmt_is_planar(in_fmt, channels);out_planar=ff_sample_fmt_is_planar(out_fmt, channels);if(in_planar==out_planar){ac->func_type=CONV_FUNC_TYPE_FLAT;ac->planes=in_planar?ac->channels:1;}elseif(in_planar) ac->func_type=CONV_FUNC_TYPE_INTERLEAVE;elseac->func_type=CONV_FUNC_TYPE_DEINTERLEAVE;set_generic_function(ac);if(ARCH_AARCH64) ff_audio_convert_init_aarch64(ac);if(ARCH_ARM) ff_audio_convert_init_arm(ac);if(ARCH_X86) ff_audio_convert_init_x86(ac);returnac;}intff_audio_convert(AudioConvert *ac, AudioData *out, AudioData *in){intuse_generic=1;intlen=in->nb_samples;intp;if(ac->dc){av_log(ac->avr, AV_LOG_TRACE,"%dsamples-audio_convert:%sto%s(dithered)\n", len, av_get_sample_fmt_name(ac->in_fmt), av_get_sample_fmt_name(ac->out_fmt));returnff_convert_dither(ac-> in
pthread_mutex_t frame_lock
static int get_video_config(AVFormatContext *s)
static const AVClass avf_class
BYTE int const BYTE int int int height
Describe the class of an AVClass context structure.
planar YUV 4:4:4 64bpp, (1 Cr & Cb sample per 1x1 Y & A samples, little-endian)
#define INTERLEAVE_OUTPUT(bps)
rational number numerator/denominator
#define AV_OPT_FLAG_DECODING_PARAM
a generic parameter which can be set by the user for demuxing or decoding
static void unlock_frames(AVFContext *ctx)
offset must point to AVRational
packed BGR 8:8:8, 32bpp, BGRXBGRX... X=unused/undefined
FrameReciever class - delegate for AVCaptureSession.
offset must point to two consecutive integers
CMSampleBufferRef current_frame
#define AV_PIX_FMT_YUV422P10
packed RGB 5:5:5, 16bpp, (msb)1X 5R 5G 5B(lsb), big-endian , X=unused/undefined
enum AVPixelFormat pixel_format
char * av_strtok(char *s, const char *delim, char **saveptr)
Split the string into several tokens which can be accessed by successive calls to av_strtok()...
Y , 1bpp, 0 is black, 1 is white, in each byte pixels are ordered from the msb to the lsb...
planar YUV 4:2:0, 12bpp, (1 Cr & Cb sample per 2x2 Y samples)
AVCaptureAudioDataOutput * audio_output
static void destroy_context(AVFContext *ctx)
static av_always_inline int pthread_cond_init(pthread_cond_t *cond, const pthread_condattr_t *attr)
static int avf_read_packet(AVFormatContext *s, AVPacket *pkt)
static av_always_inline int pthread_mutex_unlock(pthread_mutex_t *mutex)
void * priv_data
Format private data.
static int avf_close(AVFormatContext *s)
pthread_cond_t frame_wait_cond
int64_t dts
Decompression timestamp in AVStream->time_base units; the time at which the packet is decompressed...
int64_t av_get_default_channel_layout(int nb_channels)
Return default channel layout for a given number of channels.
static av_always_inline int pthread_mutex_lock(pthread_mutex_t *mutex)
const char * av_get_pix_fmt_name(enum AVPixelFormat pix_fmt)
Return the short name for a pixel format, NULL in case pix_fmt is unknown.
packed RGB 8:8:8, 32bpp, XRGBXRGB... X=unused/undefined
#define AVERROR_EXTERNAL
Generic error in an external library.
AVPixelFormat
Pixel format.
This structure stores compressed data.
#define AV_PIX_FMT_YUV422P16
int64_t pts
Presentation timestamp in AVStream->time_base units; the time at which the decompressed packet will b...