mirror of
https://github.com/pion/mediadevices.git
synced 2026-04-22 15:57:27 +08:00
03900dcb1b
* wip * wip * Organize * Remove unnecessary change in camera_darwin.go filtering * wip * Make observer stop safe during startup * wip IsAvailable impl * Fix non-darwin builds * Lock bg loop to main thread and add comment * Remove fmt prints * Simplify isAvailable; Add timeout for Read darwin * Match comment with code * Change to singleton pattern; Add clearer safer state machine states; Change language from Stop to Destroy; Add new error for when observer is unavailable; * Add stubs for linux * Move cancel() up so its not dead code sometimes * Add stubs for Windows too * Remove StopObserver usage * Add camera tests * Add device observer tests * Fix multiple destroy calls bug; Call setup in start * Improve isAvailable * Improve string handling in device observer c * Add error handling in example * Add comment about setup vs start * Rename and organize device observer darwin * Explicitly case initial state for setup * Fix potential destroy goroutine leak; Use only modern build tag; Return err not nil for stubs; Improve comments * Close startDone channel on device observer stop not wait
434 lines
15 KiB
Objective-C
434 lines
15 KiB
Objective-C
// MIT License
|
|
//
|
|
// Copyright (c) 2019-2020 Pion
|
|
//
|
|
// Permission is hereby granted, free of charge, to any person obtaining a copy
|
|
// of this software and associated documentation files (the "Software"), to deal
|
|
// in the Software without restriction, including without limitation the rights
|
|
// to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
|
// copies of the Software, and to permit persons to whom the Software is
|
|
// furnished to do so, subject to the following conditions:
|
|
//
|
|
// The above copyright notice and this permission notice shall be included in all
|
|
// copies or substantial portions of the Software.
|
|
//
|
|
// THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
// IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
// FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
|
// AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
// LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
|
// OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
// SOFTWARE.
|
|
|
|
// Naming Convention (let "name" as an actual variable name):
|
|
// - mName: "name" is a member of an Objective C object
|
|
// - pName: "name" is a C pointer
|
|
// - refName: "name" is an Objective C object reference
|
|
|
|
#import <Foundation/Foundation.h>
|
|
#import <AVFoundation/AVFoundation.h>
|
|
#import "AVFoundationBind.h"
|
|
#include <string.h>
|
|
|
|
// AVFoundationBind.m is the entry point for cgo compilation (included by avfoundation_darwin.go).
|
|
// Including DeviceObserver.m here compiles both into a single compilation unit,
|
|
// making all symbols available to the linker.
|
|
#include "DeviceObserver.m"
|
|
|
|
|
|
#define CHK(condition, status) \
|
|
do { \
|
|
if(!(condition)) { \
|
|
retStatus = status; \
|
|
goto cleanup; \
|
|
} \
|
|
} while(0)
|
|
|
|
#define CHK_STATUS(status) \
|
|
do { \
|
|
if(status != STATUS_OK) { \
|
|
retStatus = status; \
|
|
goto cleanup; \
|
|
} \
|
|
} while(0)
|
|
|
|
static NSString *const UnrecognizedMacOSVersionException = @"UnrecognizedMacOSVersionException";
|
|
|
|
@interface VideoDataDelegate : NSObject<AVCaptureVideoDataOutputSampleBufferDelegate>
|
|
|
|
@property (readonly) AVBindDataCallback mCallback;
|
|
@property (readonly) void *mPUserData;
|
|
|
|
- (void)captureOutput:(AVCaptureOutput *)captureOutput
|
|
didOutputSampleBuffer:(CMSampleBufferRef)sampleBuffer
|
|
fromConnection:(AVCaptureConnection *)connection;
|
|
|
|
@end
|
|
|
|
@implementation VideoDataDelegate
|
|
|
|
- (id) init: (AVBindDataCallback) callback
|
|
withUserData: (void*) pUserData {
|
|
self = [super init];
|
|
_mCallback = callback;
|
|
_mPUserData = pUserData;
|
|
|
|
return self;
|
|
}
|
|
|
|
- (void)captureOutput:(AVCaptureOutput *)captureOutput
|
|
didOutputSampleBuffer:(CMSampleBufferRef)sampleBuffer
|
|
fromConnection:(AVCaptureConnection *)connection {
|
|
if (CMSampleBufferGetNumSamples(sampleBuffer) != 1 ||
|
|
!CMSampleBufferIsValid(sampleBuffer) ||
|
|
!CMSampleBufferDataIsReady(sampleBuffer)) {
|
|
return;
|
|
}
|
|
|
|
CVImageBufferRef imageBuffer = CMSampleBufferGetImageBuffer(sampleBuffer);
|
|
if (imageBuffer == NULL) {
|
|
return;
|
|
}
|
|
|
|
CVBufferRetain(imageBuffer);
|
|
CVReturn ret =
|
|
CVPixelBufferLockBaseAddress(imageBuffer, kCVPixelBufferLock_ReadOnly);
|
|
if (ret != kCVReturnSuccess) {
|
|
CVBufferRelease(imageBuffer);
|
|
return;
|
|
}
|
|
|
|
// Handle NV12 special case
|
|
OSType pixelFormat = CVPixelBufferGetPixelFormatType(imageBuffer);
|
|
if (pixelFormat == kCVPixelFormatType_420YpCbCr8BiPlanarVideoRange) {
|
|
// Get actual dimensions of image (without padding)
|
|
size_t width = CVPixelBufferGetWidth(imageBuffer);
|
|
size_t height = CVPixelBufferGetHeight(imageBuffer);
|
|
size_t totalSize = /*Y plane*/ width * height + /*UV plane*/ width * height / 2;
|
|
|
|
size_t bytesPerRowY = CVPixelBufferGetBytesPerRowOfPlane(imageBuffer, 0);
|
|
size_t bytesPerRowUV = CVPixelBufferGetBytesPerRowOfPlane(imageBuffer, 1);
|
|
|
|
void *mergedBuffer = malloc(totalSize);
|
|
if (!mergedBuffer) {
|
|
NSLog(@"Failed to allocate memory for merged buffer");
|
|
CVPixelBufferUnlockBaseAddress(imageBuffer, kCVPixelBufferLock_ReadOnly);
|
|
CVBufferRelease(imageBuffer);
|
|
return;
|
|
}
|
|
|
|
// Truncate data where we know it should end to strip padding
|
|
void *yPlaneBuf = CVPixelBufferGetBaseAddressOfPlane(imageBuffer, 0);
|
|
for (size_t row = 0; row < height; ++row) {
|
|
memcpy(mergedBuffer + row * width, yPlaneBuf + row * bytesPerRowY, width);
|
|
}
|
|
|
|
void *uvPlaneBuf = CVPixelBufferGetBaseAddressOfPlane(imageBuffer, 1);
|
|
for (size_t row = 0; row < height / 2; ++row) {
|
|
memcpy(mergedBuffer + width * height + row * width, uvPlaneBuf + row * bytesPerRowUV, width);
|
|
}
|
|
|
|
_mCallback(_mPUserData, mergedBuffer, (int)totalSize);
|
|
free(mergedBuffer);
|
|
} else {
|
|
void *buf = CVPixelBufferGetBaseAddress(imageBuffer);
|
|
size_t dataSize = CVPixelBufferGetDataSize(imageBuffer);
|
|
_mCallback(_mPUserData, buf, (int)dataSize);
|
|
}
|
|
|
|
CVPixelBufferUnlockBaseAddress(imageBuffer, kCVPixelBufferLock_ReadOnly);
|
|
CVBufferRelease(imageBuffer);
|
|
}
|
|
|
|
@end
|
|
|
|
@interface AudioDataDelegate : NSObject<AVCaptureAudioDataOutputSampleBufferDelegate>
|
|
|
|
@property (readonly) AVBindDataCallback mCallback;
|
|
|
|
- (void)captureOutput:(AVCaptureOutput *)captureOutput
|
|
didOutputSampleBuffer:(CMSampleBufferRef)sampleBuffer
|
|
fromConnection:(AVCaptureConnection *)connection;
|
|
|
|
@end
|
|
|
|
@implementation AudioDataDelegate
|
|
|
|
- (id) init: (AVBindDataCallback) callback {
|
|
self = [super init];
|
|
_mCallback = callback;
|
|
return self;
|
|
}
|
|
|
|
- (void)captureOutput:(AVCaptureOutput *)captureOutput
|
|
didOutputSampleBuffer:(CMSampleBufferRef)sampleBuffer
|
|
fromConnection:(AVCaptureConnection *)connection {
|
|
// TODO
|
|
}
|
|
|
|
@end
|
|
|
|
STATUS frameFormatToFourCC(AVBindFrameFormat format, FourCharCode *pFourCC) {
|
|
STATUS retStatus = STATUS_OK;
|
|
// Useful mapping reference from ffmpeg:
|
|
// https://github.com/FFmpeg/FFmpeg/blob/c810a9502cebe32e1dd08ee3d0d17053dde44aa9/libavdevice/avfoundation.m#L53-L80
|
|
switch (format) {
|
|
case AVBindFrameFormatI420:
|
|
*pFourCC = kCVPixelFormatType_420YpCbCr8Planar;
|
|
break;
|
|
case AVBindFrameFormatNV12:
|
|
*pFourCC = kCVPixelFormatType_420YpCbCr8BiPlanarVideoRange;
|
|
break;
|
|
case AVBindFrameFormatUYVY:
|
|
*pFourCC = kCVPixelFormatType_422YpCbCr8;
|
|
break;
|
|
case AVBindFrameFormatYUYV:
|
|
*pFourCC = kCVPixelFormatType_422YpCbCr8_yuvs;
|
|
break;
|
|
// TODO: Add the rest of frame formats
|
|
default:
|
|
retStatus = STATUS_UNSUPPORTED_FRAME_FORMAT;
|
|
}
|
|
return retStatus;
|
|
}
|
|
|
|
STATUS frameFormatFromFourCC(FourCharCode fourCC, AVBindFrameFormat *pFormat) {
|
|
STATUS retStatus = STATUS_OK;
|
|
switch (fourCC) {
|
|
case kCVPixelFormatType_420YpCbCr8Planar:
|
|
*pFormat = AVBindFrameFormatI420;
|
|
break;
|
|
case kCVPixelFormatType_420YpCbCr8BiPlanarFullRange:
|
|
case kCVPixelFormatType_420YpCbCr8BiPlanarVideoRange:
|
|
*pFormat = AVBindFrameFormatNV12;
|
|
break;
|
|
case kCVPixelFormatType_422YpCbCr8:
|
|
*pFormat = AVBindFrameFormatUYVY;
|
|
break;
|
|
case kCVPixelFormatType_422YpCbCr8_yuvs:
|
|
*pFormat = AVBindFrameFormatYUYV;
|
|
break;
|
|
// TODO: Add the rest of frame formats
|
|
default:
|
|
retStatus = STATUS_UNSUPPORTED_FRAME_FORMAT;
|
|
}
|
|
return retStatus;
|
|
}
|
|
|
|
|
|
STATUS AVBindDevices(AVBindMediaType mediaType, PAVBindDevice *ppDevices, int *pLen) {
|
|
static AVBindDevice devices[MAX_DEVICES];
|
|
STATUS retStatus = STATUS_OK;
|
|
NSAutoreleasePool *refPool = [[NSAutoreleasePool alloc] init];
|
|
CHK(mediaType == AVBindMediaTypeVideo || mediaType == AVBindMediaTypeAudio, STATUS_UNSUPPORTED_MEDIA_TYPE);
|
|
CHK(ppDevices != NULL && pLen != NULL, STATUS_NULL_ARG);
|
|
|
|
PAVBindDevice pDevice;
|
|
AVMediaType _mediaType = mediaType == AVBindMediaTypeVideo ? AVMediaTypeVideo : AVMediaTypeAudio;
|
|
|
|
NSArray *refAllTypes;
|
|
#if defined(MAC_OS_VERSION_14_0)
|
|
if (@available(macOS 14.0, *)) {
|
|
refAllTypes = @[
|
|
AVCaptureDeviceTypeBuiltInWideAngleCamera,
|
|
AVCaptureDeviceTypeMicrophone,
|
|
AVCaptureDeviceTypeExternal,
|
|
];
|
|
} else {
|
|
@throw [NSException exceptionWithName:UnrecognizedMacOSVersionException
|
|
reason:@"Unrecognized or unsupported macOS version detected."
|
|
userInfo:nil];
|
|
}
|
|
#else
|
|
refAllTypes = @[
|
|
AVCaptureDeviceTypeBuiltInWideAngleCamera,
|
|
AVCaptureDeviceTypeBuiltInMicrophone,
|
|
AVCaptureDeviceTypeExternalUnknown,
|
|
];
|
|
#endif
|
|
|
|
AVCaptureDeviceDiscoverySession *refSession = [AVCaptureDeviceDiscoverySession
|
|
discoverySessionWithDeviceTypes: refAllTypes
|
|
mediaType: _mediaType
|
|
position: AVCaptureDevicePositionUnspecified];
|
|
|
|
int i = 0;
|
|
for (AVCaptureDevice *refDevice in refSession.devices) {
|
|
if (i >= MAX_DEVICES) {
|
|
break;
|
|
}
|
|
|
|
pDevice = devices + i;
|
|
strncpy(pDevice->uid, refDevice.uniqueID.UTF8String, MAX_DEVICE_UID_CHARS);
|
|
pDevice->uid[MAX_DEVICE_UID_CHARS] = '\0';
|
|
strncpy(pDevice->name, refDevice.localizedName.UTF8String, MAX_DEVICE_NAME_CHARS);
|
|
pDevice->name[MAX_DEVICE_NAME_CHARS] = '\0';
|
|
i++;
|
|
}
|
|
|
|
*ppDevices = devices;
|
|
*pLen = i;
|
|
|
|
cleanup:
|
|
[refPool drain];
|
|
return retStatus;
|
|
}
|
|
|
|
struct AVBindSession {
|
|
AVBindDevice device;
|
|
AVCaptureSession *refCaptureSession;
|
|
AVBindMediaProperty properties[MAX_PROPERTIES];
|
|
};
|
|
|
|
|
|
STATUS AVBindSessionInit(AVBindDevice device, PAVBindSession *ppSessionResult) {
|
|
STATUS retStatus = STATUS_OK;
|
|
CHK(ppSessionResult != NULL, STATUS_NULL_ARG);
|
|
PAVBindSession pSession = malloc(sizeof(AVBindSession));
|
|
pSession->device = device;
|
|
pSession->refCaptureSession = NULL;
|
|
*ppSessionResult = pSession;
|
|
|
|
cleanup:
|
|
return retStatus;
|
|
}
|
|
|
|
STATUS AVBindSessionFree(PAVBindSession *ppSession) {
|
|
STATUS retStatus = STATUS_OK;
|
|
CHK(ppSession != NULL, STATUS_NULL_ARG);
|
|
PAVBindSession pSession = *ppSession;
|
|
if (pSession->refCaptureSession != NULL) {
|
|
[pSession->refCaptureSession release];
|
|
pSession->refCaptureSession = NULL;
|
|
}
|
|
free(pSession);
|
|
*ppSession = NULL;
|
|
|
|
cleanup:
|
|
return retStatus;
|
|
}
|
|
|
|
STATUS AVBindSessionOpen(PAVBindSession pSession,
|
|
AVBindMediaProperty property,
|
|
AVBindDataCallback dataCallback,
|
|
void *pUserData) {
|
|
STATUS retStatus = STATUS_OK;
|
|
NSAutoreleasePool *refPool = [[NSAutoreleasePool alloc] init];
|
|
CHK(pSession != NULL && dataCallback != NULL, STATUS_NULL_ARG);
|
|
|
|
AVCaptureDeviceInput *refInput;
|
|
NSError *refErr = NULL;
|
|
NSString *refUID = [NSString stringWithUTF8String: pSession->device.uid];
|
|
AVCaptureDevice *refDevice = [AVCaptureDevice deviceWithUniqueID: refUID];
|
|
|
|
refInput = [[AVCaptureDeviceInput alloc] initWithDevice: refDevice error: &refErr];
|
|
CHK(refErr == NULL, STATUS_DEVICE_INIT_FAILED);
|
|
|
|
AVCaptureSession *refCaptureSession = [[AVCaptureSession alloc] init];
|
|
refCaptureSession.sessionPreset = AVCaptureSessionPresetMedium;
|
|
[refCaptureSession addInput: refInput];
|
|
|
|
if ([refDevice hasMediaType: AVMediaTypeVideo]) {
|
|
VideoDataDelegate *pDelegate = [[VideoDataDelegate alloc]
|
|
init: dataCallback
|
|
withUserData: pUserData];
|
|
|
|
AVCaptureVideoDataOutput *pOutput = [[AVCaptureVideoDataOutput alloc] init];
|
|
FourCharCode fourCC;
|
|
CHK_STATUS(frameFormatToFourCC(property.frameFormat, &fourCC));
|
|
|
|
pOutput.videoSettings = @{
|
|
(id)kCVPixelBufferWidthKey: @(property.width),
|
|
(id)kCVPixelBufferHeightKey: @(property.height),
|
|
(id)kCVPixelBufferPixelFormatTypeKey: @(fourCC),
|
|
};
|
|
pOutput.alwaysDiscardsLateVideoFrames = YES;
|
|
dispatch_queue_t queue =
|
|
dispatch_queue_create("captureQueue", DISPATCH_QUEUE_SERIAL);
|
|
[pOutput setSampleBufferDelegate:pDelegate queue:queue];
|
|
[refCaptureSession addOutput: pOutput];
|
|
} else {
|
|
// TODO: implement audio pipeline
|
|
}
|
|
|
|
pSession->refCaptureSession = [refCaptureSession retain];
|
|
[refCaptureSession startRunning];
|
|
|
|
cleanup:
|
|
[refPool drain];
|
|
return retStatus;
|
|
}
|
|
|
|
|
|
STATUS AVBindSessionClose(PAVBindSession pSession) {
|
|
STATUS retStatus = STATUS_OK;
|
|
CHK(pSession != NULL, STATUS_NULL_ARG);
|
|
CHK(pSession->refCaptureSession != NULL, STATUS_OK);
|
|
|
|
[pSession->refCaptureSession stopRunning];
|
|
[pSession->refCaptureSession release];
|
|
pSession->refCaptureSession = NULL;
|
|
|
|
cleanup:
|
|
return retStatus;
|
|
}
|
|
|
|
static NSString* FourCCString(FourCharCode code) {
|
|
NSString *result = [NSString stringWithFormat:@"%c%c%c%c",
|
|
(code >> 24) & 0xff,
|
|
(code >> 16) & 0xff,
|
|
(code >> 8) & 0xff,
|
|
code & 0xff];
|
|
NSCharacterSet *characterSet = [NSCharacterSet whitespaceCharacterSet];
|
|
return [result stringByTrimmingCharactersInSet:characterSet];
|
|
}
|
|
|
|
STATUS AVBindSessionProperties(PAVBindSession pSession, PAVBindMediaProperty *ppProperties, int *pLen) {
|
|
STATUS retStatus = STATUS_OK;
|
|
NSAutoreleasePool *refPool = [[NSAutoreleasePool alloc] init];
|
|
CHK(pSession != NULL && ppProperties != NULL && pLen != NULL, STATUS_NULL_ARG);
|
|
|
|
NSString *refDeviceUID = [NSString stringWithUTF8String: pSession->device.uid];
|
|
AVCaptureDevice *refDevice = [AVCaptureDevice deviceWithUniqueID: refDeviceUID];
|
|
FourCharCode fourCC;
|
|
CMVideoFormatDescriptionRef videoFormat;
|
|
CMVideoDimensions videoDimensions;
|
|
|
|
memset(pSession->properties, 0, sizeof(pSession->properties));
|
|
PAVBindMediaProperty pProperty = pSession->properties;
|
|
int len = 0;
|
|
for (AVCaptureDeviceFormat *refFormat in refDevice.formats) {
|
|
// TODO: Probably gives a warn to the user
|
|
if (len >= MAX_PROPERTIES) {
|
|
NSLog(@"[WARNING] skipping the rest of properties due to MAX_PROPERTIES");
|
|
break;
|
|
}
|
|
|
|
if ([refFormat.mediaType isEqual:AVMediaTypeVideo]) {
|
|
fourCC = CMFormatDescriptionGetMediaSubType(refFormat.formatDescription);
|
|
if (frameFormatFromFourCC(fourCC, &pProperty->frameFormat) != STATUS_OK) {
|
|
NSLog(@"[WARNING] skipping %@ %dx%d since it's not supported", FourCCString(fourCC), videoDimensions.width, videoDimensions.height);
|
|
continue;
|
|
}
|
|
|
|
videoFormat = (CMVideoFormatDescriptionRef) refFormat.formatDescription;
|
|
videoDimensions = CMVideoFormatDescriptionGetDimensions(videoFormat);
|
|
pProperty->height = videoDimensions.height;
|
|
pProperty->width = videoDimensions.width;
|
|
} else {
|
|
// TODO: Get audio properties
|
|
}
|
|
|
|
pProperty++;
|
|
len++;
|
|
}
|
|
|
|
*ppProperties = pSession->properties;
|
|
*pLen = len;
|
|
|
|
cleanup:
|
|
|
|
[refPool drain];
|
|
return retStatus;
|
|
}
|