mirror of https://github.com/opencv/opencv.git
Open Source Computer Vision Library
https://opencv.org/
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
619 lines
22 KiB
619 lines
22 KiB
/* |
|
* cap_ios_video_camera.mm |
|
* For iOS video I/O |
|
* by Eduard Feicho on 29/07/12 |
|
* by Alexander Shishkov on 17/07/13 |
|
* Copyright 2012. All rights reserved. |
|
* |
|
* Redistribution and use in source and binary forms, with or without |
|
* modification, are permitted provided that the following conditions are met: |
|
* |
|
* 1. Redistributions of source code must retain the above copyright notice, |
|
* this list of conditions and the following disclaimer. |
|
* 2. Redistributions in binary form must reproduce the above copyright notice, |
|
* this list of conditions and the following disclaimer in the documentation |
|
* and/or other materials provided with the distribution. |
|
* 3. The name of the author may not be used to endorse or promote products |
|
* derived from this software without specific prior written permission. |
|
* |
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR "AS IS" AND ANY EXPRESS OR IMPLIED |
|
* WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF |
|
* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO |
|
* EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
|
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; |
|
* OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, |
|
* WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR |
|
* OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF |
|
* ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
|
* |
|
*/ |
|
|
|
#import "opencv2/highgui/cap_ios.h" |
|
#include "precomp.hpp" |
|
#import <AssetsLibrary/AssetsLibrary.h> |
|
|
|
|
|
static CGFloat DegreesToRadians(CGFloat degrees) {return degrees * M_PI / 180;}; |
|
|
|
#pragma mark - Private Interface |
|
|
|
|
|
|
|
|
|
@interface CvVideoCamera () |
|
|
|
- (void)createVideoDataOutput; |
|
- (void)createVideoFileOutput; |
|
|
|
|
|
@property (nonatomic, retain) CALayer *customPreviewLayer; |
|
@property (nonatomic, retain) AVCaptureVideoDataOutput *videoDataOutput; |
|
|
|
@end |
|
|
|
|
|
|
|
#pragma mark - Implementation |
|
|
|
|
|
|
|
@implementation CvVideoCamera |
|
|
|
|
|
|
|
|
|
@synthesize delegate; |
|
@synthesize grayscaleMode; |
|
|
|
@synthesize customPreviewLayer; |
|
@synthesize videoDataOutput; |
|
|
|
@synthesize recordVideo; |
|
@synthesize rotateVideo; |
|
//@synthesize videoFileOutput; |
|
@synthesize recordAssetWriterInput; |
|
@synthesize recordPixelBufferAdaptor; |
|
@synthesize recordAssetWriter; |
|
|
|
|
|
|
|
#pragma mark - Constructors |
|
|
|
- (id)initWithParentView:(UIView*)parent; |
|
{ |
|
self = [super initWithParentView:parent]; |
|
if (self) { |
|
self.useAVCaptureVideoPreviewLayer = NO; |
|
self.recordVideo = NO; |
|
self.rotateVideo = NO; |
|
} |
|
return self; |
|
} |
|
|
|
|
|
|
|
#pragma mark - Public interface |
|
|
|
|
|
- (void)start; |
|
{ |
|
[super start]; |
|
|
|
if (self.recordVideo == YES) { |
|
NSError* error; |
|
if ([[NSFileManager defaultManager] fileExistsAtPath:[self videoFileString]]) { |
|
[[NSFileManager defaultManager] removeItemAtPath:[self videoFileString] error:&error]; |
|
} |
|
if (error == nil) { |
|
NSLog(@"[Camera] Delete file %@", [self videoFileString]); |
|
} |
|
} |
|
} |
|
|
|
|
|
|
|
- (void)stop; |
|
{ |
|
[super stop]; |
|
|
|
self.videoDataOutput = nil; |
|
if (videoDataOutputQueue) { |
|
dispatch_release(videoDataOutputQueue); |
|
} |
|
|
|
if (self.recordVideo == YES) { |
|
|
|
if (self.recordAssetWriter.status == AVAssetWriterStatusWriting) { |
|
[self.recordAssetWriter finishWriting]; |
|
NSLog(@"[Camera] recording stopped"); |
|
} else { |
|
NSLog(@"[Camera] Recording Error: asset writer status is not writing"); |
|
} |
|
|
|
self.recordAssetWriter = nil; |
|
self.recordAssetWriterInput = nil; |
|
self.recordPixelBufferAdaptor = nil; |
|
} |
|
|
|
[self.customPreviewLayer removeFromSuperlayer]; |
|
self.customPreviewLayer = nil; |
|
} |
|
|
|
// TODO fix |
|
- (void)adjustLayoutToInterfaceOrientation:(UIInterfaceOrientation)interfaceOrientation; |
|
{ |
|
|
|
NSLog(@"layout preview layer"); |
|
if (self.parentView != nil) { |
|
|
|
CALayer* layer = self.customPreviewLayer; |
|
CGRect bounds = self.customPreviewLayer.bounds; |
|
int rotation_angle = 0; |
|
bool flip_bounds = false; |
|
|
|
switch (interfaceOrientation) { |
|
case UIInterfaceOrientationPortrait: |
|
NSLog(@"to Portrait"); |
|
rotation_angle = 270; |
|
break; |
|
case UIInterfaceOrientationPortraitUpsideDown: |
|
rotation_angle = 90; |
|
NSLog(@"to UpsideDown"); |
|
break; |
|
case UIInterfaceOrientationLandscapeLeft: |
|
rotation_angle = 0; |
|
NSLog(@"to LandscapeLeft"); |
|
break; |
|
case UIInterfaceOrientationLandscapeRight: |
|
rotation_angle = 180; |
|
NSLog(@"to LandscapeRight"); |
|
break; |
|
default: |
|
break; // leave the layer in its last known orientation |
|
} |
|
|
|
switch (defaultAVCaptureVideoOrientation) { |
|
case AVCaptureVideoOrientationLandscapeRight: |
|
rotation_angle += 180; |
|
break; |
|
case AVCaptureVideoOrientationPortraitUpsideDown: |
|
rotation_angle += 270; |
|
break; |
|
case AVCaptureVideoOrientationPortrait: |
|
rotation_angle += 90; |
|
case AVCaptureVideoOrientationLandscapeLeft: |
|
break; |
|
default: |
|
break; |
|
} |
|
rotation_angle = rotation_angle % 360; |
|
|
|
if (rotation_angle == 90 || rotation_angle == 270) { |
|
flip_bounds = true; |
|
} |
|
|
|
if (flip_bounds) { |
|
NSLog(@"flip bounds"); |
|
bounds = CGRectMake(0, 0, bounds.size.height, bounds.size.width); |
|
} |
|
|
|
layer.position = CGPointMake(self.parentView.frame.size.width/2., self.parentView.frame.size.height/2.); |
|
self.customPreviewLayer.bounds = CGRectMake(0, 0, self.parentView.frame.size.width, self.parentView.frame.size.height); |
|
|
|
layer.affineTransform = CGAffineTransformMakeRotation( DegreesToRadians(rotation_angle) ); |
|
layer.bounds = bounds; |
|
} |
|
|
|
} |
|
|
|
// TODO fix |
|
- (void)layoutPreviewLayer; |
|
{ |
|
NSLog(@"layout preview layer"); |
|
if (self.parentView != nil) { |
|
|
|
CALayer* layer = self.customPreviewLayer; |
|
CGRect bounds = self.customPreviewLayer.bounds; |
|
int rotation_angle = 0; |
|
bool flip_bounds = false; |
|
|
|
switch (currentDeviceOrientation) { |
|
case UIDeviceOrientationPortrait: |
|
rotation_angle = 270; |
|
break; |
|
case UIDeviceOrientationPortraitUpsideDown: |
|
rotation_angle = 90; |
|
break; |
|
case UIDeviceOrientationLandscapeLeft: |
|
NSLog(@"left"); |
|
rotation_angle = 180; |
|
break; |
|
case UIDeviceOrientationLandscapeRight: |
|
NSLog(@"right"); |
|
rotation_angle = 0; |
|
break; |
|
case UIDeviceOrientationFaceUp: |
|
case UIDeviceOrientationFaceDown: |
|
default: |
|
break; // leave the layer in its last known orientation |
|
} |
|
|
|
switch (defaultAVCaptureVideoOrientation) { |
|
case AVCaptureVideoOrientationLandscapeRight: |
|
rotation_angle += 180; |
|
break; |
|
case AVCaptureVideoOrientationPortraitUpsideDown: |
|
rotation_angle += 270; |
|
break; |
|
case AVCaptureVideoOrientationPortrait: |
|
rotation_angle += 90; |
|
case AVCaptureVideoOrientationLandscapeLeft: |
|
break; |
|
default: |
|
break; |
|
} |
|
rotation_angle = rotation_angle % 360; |
|
|
|
if (rotation_angle == 90 || rotation_angle == 270) { |
|
flip_bounds = true; |
|
} |
|
|
|
if (flip_bounds) { |
|
NSLog(@"flip bounds"); |
|
bounds = CGRectMake(0, 0, bounds.size.height, bounds.size.width); |
|
} |
|
|
|
layer.position = CGPointMake(self.parentView.frame.size.width/2., self.parentView.frame.size.height/2.); |
|
layer.affineTransform = CGAffineTransformMakeRotation( DegreesToRadians(rotation_angle) ); |
|
layer.bounds = bounds; |
|
} |
|
|
|
} |
|
|
|
#pragma mark - Private Interface |
|
|
|
- (void)createVideoDataOutput; |
|
{ |
|
// Make a video data output |
|
self.videoDataOutput = [AVCaptureVideoDataOutput new]; |
|
|
|
// In grayscale mode we want YUV (YpCbCr 4:2:0) so we can directly access the graylevel intensity values (Y component) |
|
// In color mode we, BGRA format is used |
|
OSType format = self.grayscaleMode ? kCVPixelFormatType_420YpCbCr8BiPlanarFullRange : kCVPixelFormatType_32BGRA; |
|
|
|
self.videoDataOutput.videoSettings = [NSDictionary dictionaryWithObject:[NSNumber numberWithUnsignedInt:format] |
|
forKey:(id)kCVPixelBufferPixelFormatTypeKey]; |
|
|
|
// discard if the data output queue is blocked (as we process the still image) |
|
[self.videoDataOutput setAlwaysDiscardsLateVideoFrames:YES]; |
|
|
|
if ( [self.captureSession canAddOutput:self.videoDataOutput] ) { |
|
[self.captureSession addOutput:self.videoDataOutput]; |
|
} |
|
[[self.videoDataOutput connectionWithMediaType:AVMediaTypeVideo] setEnabled:YES]; |
|
|
|
|
|
// set default FPS |
|
if ([self.videoDataOutput connectionWithMediaType:AVMediaTypeVideo].supportsVideoMinFrameDuration) { |
|
[self.videoDataOutput connectionWithMediaType:AVMediaTypeVideo].videoMinFrameDuration = CMTimeMake(1, self.defaultFPS); |
|
} |
|
if ([self.videoDataOutput connectionWithMediaType:AVMediaTypeVideo].supportsVideoMaxFrameDuration) { |
|
[self.videoDataOutput connectionWithMediaType:AVMediaTypeVideo].videoMaxFrameDuration = CMTimeMake(1, self.defaultFPS); |
|
} |
|
|
|
// set video mirroring for front camera (more intuitive) |
|
if ([self.videoDataOutput connectionWithMediaType:AVMediaTypeVideo].supportsVideoMirroring) { |
|
if (self.defaultAVCaptureDevicePosition == AVCaptureDevicePositionFront) { |
|
[self.videoDataOutput connectionWithMediaType:AVMediaTypeVideo].videoMirrored = YES; |
|
} else { |
|
[self.videoDataOutput connectionWithMediaType:AVMediaTypeVideo].videoMirrored = NO; |
|
} |
|
} |
|
|
|
// set default video orientation |
|
if ([self.videoDataOutput connectionWithMediaType:AVMediaTypeVideo].supportsVideoOrientation) { |
|
[self.videoDataOutput connectionWithMediaType:AVMediaTypeVideo].videoOrientation = self.defaultAVCaptureVideoOrientation; |
|
} |
|
|
|
|
|
// create a custom preview layer |
|
self.customPreviewLayer = [CALayer layer]; |
|
self.customPreviewLayer.bounds = CGRectMake(0, 0, self.parentView.frame.size.width, self.parentView.frame.size.height); |
|
[self layoutPreviewLayer]; |
|
|
|
// create a serial dispatch queue used for the sample buffer delegate as well as when a still image is captured |
|
// a serial dispatch queue must be used to guarantee that video frames will be delivered in order |
|
// see the header doc for setSampleBufferDelegate:queue: for more information |
|
videoDataOutputQueue = dispatch_queue_create("VideoDataOutputQueue", DISPATCH_QUEUE_SERIAL); |
|
[self.videoDataOutput setSampleBufferDelegate:self queue:videoDataOutputQueue]; |
|
|
|
|
|
NSLog(@"[Camera] created AVCaptureVideoDataOutput at %d FPS", self.defaultFPS); |
|
} |
|
|
|
|
|
|
|
- (void)createVideoFileOutput; |
|
{ |
|
/* Video File Output in H.264, via AVAsserWriter */ |
|
NSLog(@"Create Video with dimensions %dx%d", self.imageWidth, self.imageHeight); |
|
|
|
NSDictionary *outputSettings |
|
= [NSDictionary dictionaryWithObjectsAndKeys:[NSNumber numberWithInt:self.imageWidth], AVVideoWidthKey, |
|
[NSNumber numberWithInt:self.imageHeight], AVVideoHeightKey, |
|
AVVideoCodecH264, AVVideoCodecKey, |
|
nil |
|
]; |
|
|
|
|
|
self.recordAssetWriterInput = [AVAssetWriterInput assetWriterInputWithMediaType:AVMediaTypeVideo outputSettings:outputSettings]; |
|
|
|
|
|
int pixelBufferFormat = (self.grayscaleMode == YES) ? kCVPixelFormatType_420YpCbCr8BiPlanarFullRange : kCVPixelFormatType_32BGRA; |
|
|
|
self.recordPixelBufferAdaptor = |
|
[[AVAssetWriterInputPixelBufferAdaptor alloc] |
|
initWithAssetWriterInput:self.recordAssetWriterInput |
|
sourcePixelBufferAttributes:[NSDictionary dictionaryWithObjectsAndKeys:[NSNumber numberWithInt:pixelBufferFormat], kCVPixelBufferPixelFormatTypeKey, nil]]; |
|
|
|
NSError* error = nil; |
|
NSLog(@"Create AVAssetWriter with url: %@", [self videoFileURL]); |
|
self.recordAssetWriter = [AVAssetWriter assetWriterWithURL:[self videoFileURL] |
|
fileType:AVFileTypeMPEG4 |
|
error:&error]; |
|
if (error != nil) { |
|
NSLog(@"[Camera] Unable to create AVAssetWriter: %@", error); |
|
} |
|
|
|
[self.recordAssetWriter addInput:self.recordAssetWriterInput]; |
|
self.recordAssetWriterInput.expectsMediaDataInRealTime = YES; |
|
|
|
NSLog(@"[Camera] created AVAssetWriter"); |
|
} |
|
|
|
|
|
- (void)createCaptureOutput; |
|
{ |
|
[self createVideoDataOutput]; |
|
if (self.recordVideo == YES) { |
|
[self createVideoFileOutput]; |
|
} |
|
} |
|
|
|
- (void)createCustomVideoPreview; |
|
{ |
|
[self.parentView.layer addSublayer:self.customPreviewLayer]; |
|
} |
|
|
|
- (CVPixelBufferRef) pixelBufferFromCGImage: (CGImageRef) image |
|
{ |
|
|
|
CGSize frameSize = CGSizeMake(CGImageGetWidth(image), CGImageGetHeight(image)); |
|
NSDictionary *options = [NSDictionary dictionaryWithObjectsAndKeys: |
|
[NSNumber numberWithBool:NO], kCVPixelBufferCGImageCompatibilityKey, |
|
[NSNumber numberWithBool:NO], kCVPixelBufferCGBitmapContextCompatibilityKey, |
|
nil]; |
|
CVPixelBufferRef pxbuffer = NULL; |
|
CVReturn status = CVPixelBufferCreate(kCFAllocatorDefault, frameSize.width, |
|
frameSize.height, kCVPixelFormatType_32ARGB, (CFDictionaryRef) CFBridgingRetain(options), |
|
&pxbuffer); |
|
NSParameterAssert(status == kCVReturnSuccess && pxbuffer != NULL); |
|
|
|
CVPixelBufferLockBaseAddress(pxbuffer, 0); |
|
void *pxdata = CVPixelBufferGetBaseAddress(pxbuffer); |
|
|
|
|
|
CGColorSpaceRef rgbColorSpace = CGColorSpaceCreateDeviceRGB(); |
|
CGContextRef context = CGBitmapContextCreate(pxdata, frameSize.width, |
|
frameSize.height, 8, 4*frameSize.width, rgbColorSpace, |
|
kCGImageAlphaPremultipliedFirst); |
|
|
|
CGContextDrawImage(context, CGRectMake(0, 0, CGImageGetWidth(image), |
|
CGImageGetHeight(image)), image); |
|
CGColorSpaceRelease(rgbColorSpace); |
|
CGContextRelease(context); |
|
|
|
CVPixelBufferUnlockBaseAddress(pxbuffer, 0); |
|
|
|
return pxbuffer; |
|
} |
|
|
|
#pragma mark - Protocol AVCaptureVideoDataOutputSampleBufferDelegate |
|
|
|
|
|
- (void)captureOutput:(AVCaptureOutput *)captureOutput didOutputSampleBuffer:(CMSampleBufferRef)sampleBuffer fromConnection:(AVCaptureConnection *)connection |
|
{ |
|
if (self.delegate) { |
|
|
|
// convert from Core Media to Core Video |
|
CVImageBufferRef imageBuffer = CMSampleBufferGetImageBuffer(sampleBuffer); |
|
CVPixelBufferLockBaseAddress(imageBuffer, 0); |
|
|
|
void* bufferAddress; |
|
size_t width; |
|
size_t height; |
|
size_t bytesPerRow; |
|
|
|
CGColorSpaceRef colorSpace; |
|
CGContextRef context; |
|
|
|
int format_opencv; |
|
|
|
OSType format = CVPixelBufferGetPixelFormatType(imageBuffer); |
|
if (format == kCVPixelFormatType_420YpCbCr8BiPlanarFullRange) { |
|
|
|
format_opencv = CV_8UC1; |
|
|
|
bufferAddress = CVPixelBufferGetBaseAddressOfPlane(imageBuffer, 0); |
|
width = CVPixelBufferGetWidthOfPlane(imageBuffer, 0); |
|
height = CVPixelBufferGetHeightOfPlane(imageBuffer, 0); |
|
bytesPerRow = CVPixelBufferGetBytesPerRowOfPlane(imageBuffer, 0); |
|
|
|
} else { // expect kCVPixelFormatType_32BGRA |
|
|
|
format_opencv = CV_8UC4; |
|
|
|
bufferAddress = CVPixelBufferGetBaseAddress(imageBuffer); |
|
width = CVPixelBufferGetWidth(imageBuffer); |
|
height = CVPixelBufferGetHeight(imageBuffer); |
|
bytesPerRow = CVPixelBufferGetBytesPerRow(imageBuffer); |
|
|
|
} |
|
|
|
// delegate image processing to the delegate |
|
cv::Mat image(height, width, format_opencv, bufferAddress, bytesPerRow); |
|
|
|
cv::Mat* result = NULL; |
|
CGImage* dstImage; |
|
|
|
if ([self.delegate respondsToSelector:@selector(processImage:)]) { |
|
[self.delegate processImage:image]; |
|
} |
|
|
|
// check if matrix data pointer or dimensions were changed by the delegate |
|
bool iOSimage = false; |
|
if (height == image.rows && width == image.cols && format_opencv == image.type() && bufferAddress == image.data && bytesPerRow == image.step) { |
|
iOSimage = true; |
|
} |
|
|
|
|
|
// (create color space, create graphics context, render buffer) |
|
CGBitmapInfo bitmapInfo; |
|
|
|
// basically we decide if it's a grayscale, rgb or rgba image |
|
if (image.channels() == 1) { |
|
colorSpace = CGColorSpaceCreateDeviceGray(); |
|
bitmapInfo = kCGImageAlphaNone; |
|
} else if (image.channels() == 3) { |
|
colorSpace = CGColorSpaceCreateDeviceRGB(); |
|
bitmapInfo = kCGImageAlphaNone; |
|
if (iOSimage) { |
|
bitmapInfo |= kCGBitmapByteOrder32Little; |
|
} else { |
|
bitmapInfo |= kCGBitmapByteOrder32Big; |
|
} |
|
} else { |
|
colorSpace = CGColorSpaceCreateDeviceRGB(); |
|
bitmapInfo = kCGImageAlphaPremultipliedFirst; |
|
if (iOSimage) { |
|
bitmapInfo |= kCGBitmapByteOrder32Little; |
|
} else { |
|
bitmapInfo |= kCGBitmapByteOrder32Big; |
|
} |
|
} |
|
|
|
if (iOSimage) { |
|
context = CGBitmapContextCreate(bufferAddress, width, height, 8, bytesPerRow, colorSpace, bitmapInfo); |
|
dstImage = CGBitmapContextCreateImage(context); |
|
CGContextRelease(context); |
|
} else { |
|
|
|
NSData *data = [NSData dataWithBytes:image.data length:image.elemSize()*image.total()]; |
|
CGDataProviderRef provider = CGDataProviderCreateWithCFData((__bridge CFDataRef)data); |
|
|
|
// Creating CGImage from cv::Mat |
|
dstImage = CGImageCreate(image.cols, // width |
|
image.rows, // height |
|
8, // bits per component |
|
8 * image.elemSize(), // bits per pixel |
|
image.step, // bytesPerRow |
|
colorSpace, // colorspace |
|
bitmapInfo, // bitmap info |
|
provider, // CGDataProviderRef |
|
NULL, // decode |
|
false, // should interpolate |
|
kCGRenderingIntentDefault // intent |
|
); |
|
|
|
CGDataProviderRelease(provider); |
|
} |
|
|
|
|
|
// render buffer |
|
dispatch_sync(dispatch_get_main_queue(), ^{ |
|
self.customPreviewLayer.contents = (__bridge id)dstImage; |
|
}); |
|
|
|
|
|
if (self.recordVideo == YES) { |
|
lastSampleTime = CMSampleBufferGetPresentationTimeStamp(sampleBuffer); |
|
// CMTimeShow(lastSampleTime); |
|
if (self.recordAssetWriter.status != AVAssetWriterStatusWriting) { |
|
[self.recordAssetWriter startWriting]; |
|
[self.recordAssetWriter startSessionAtSourceTime:lastSampleTime]; |
|
if (self.recordAssetWriter.status != AVAssetWriterStatusWriting) { |
|
NSLog(@"[Camera] Recording Error: asset writer status is not writing: %@", self.recordAssetWriter.error); |
|
return; |
|
} else { |
|
NSLog(@"[Camera] Video recording started"); |
|
} |
|
} |
|
|
|
if (self.recordAssetWriterInput.readyForMoreMediaData) { |
|
CVImageBufferRef pixelBuffer = [self pixelBufferFromCGImage:dstImage]; |
|
if (! [self.recordPixelBufferAdaptor appendPixelBuffer:pixelBuffer |
|
withPresentationTime:lastSampleTime] ) { |
|
NSLog(@"Video Writing Error"); |
|
} |
|
} |
|
|
|
} |
|
|
|
|
|
// cleanup |
|
CGImageRelease(dstImage); |
|
|
|
CGColorSpaceRelease(colorSpace); |
|
|
|
CVPixelBufferUnlockBaseAddress(imageBuffer, 0); |
|
} |
|
} |
|
|
|
|
|
- (void)updateOrientation; |
|
{ |
|
if (self.rotateVideo == YES) |
|
{ |
|
NSLog(@"rotate.."); |
|
self.customPreviewLayer.bounds = CGRectMake(0, 0, self.parentView.frame.size.width, self.parentView.frame.size.height); |
|
[self layoutPreviewLayer]; |
|
} |
|
} |
|
|
|
|
|
- (void)saveVideo; |
|
{ |
|
if (self.recordVideo == NO) { |
|
return; |
|
} |
|
|
|
ALAssetsLibrary *library = [[ALAssetsLibrary alloc] init]; |
|
if ([library videoAtPathIsCompatibleWithSavedPhotosAlbum:[self videoFileURL]]) { |
|
[library writeVideoAtPathToSavedPhotosAlbum:[self videoFileURL] |
|
completionBlock:^(NSURL *assetURL, NSError *error){}]; |
|
} |
|
} |
|
|
|
|
|
- (NSURL *)videoFileURL; |
|
{ |
|
NSString *outputPath = [[NSString alloc] initWithFormat:@"%@%@", NSTemporaryDirectory(), @"output.mov"]; |
|
NSURL *outputURL = [NSURL fileURLWithPath:outputPath]; |
|
NSFileManager *fileManager = [NSFileManager defaultManager]; |
|
if ([fileManager fileExistsAtPath:outputPath]) { |
|
NSLog(@"file exists"); |
|
} |
|
return outputURL; |
|
} |
|
|
|
|
|
|
|
- (NSString *)videoFileString; |
|
{ |
|
NSString *outputPath = [[NSString alloc] initWithFormat:@"%@%@", NSTemporaryDirectory(), @"output.mov"]; |
|
return outputPath; |
|
} |
|
|
|
@end |
|
|
|
|