现在的位置: 首页 > 编程语言 > 正文

iOS使用AVFoundation展示视频

2020年02月18日 编程语言 ⁄ 共 5087字 ⁄ 字号 评论关闭

本文实例为大家分享了iOS使用AVFoundation展示视频的具体代码,供大家参考,具体内容如下

//// Capter2ViewController.m// IosTest//// Created by garin on 13-7-19.// Copyright (c) 2013年 garin. All rights reserved.//#import "Capter2ViewController.h"@interface Capter2ViewController ()@end@implementation Capter2ViewController-(void) dealloc{ [session release]; [super dealloc];}- (id)initWithNibName:(NSString *)nibNameOrNil bundle:(NSBundle *)nibBundleOrNil{ self = [super initWithNibName:nibNameOrNil bundle:nibBundleOrNil]; if (self) { // Custom initialization } return self;}- (void)viewDidLoad{ [super viewDidLoad]; videoPreviewView=[[UIView alloc] initWithFrame:CGRectMake(10, 10, 320, 200)]; [self.view addSubview:videoPreviewView]; [videoPreviewView release]; // Do any additional setup after loading the view. //在viewdidload调用下面的函数显示摄像信息 [self setupCaptureSession];// imgView=[[UIImageView alloc] initWithFrame:CGRectMake(10, 230, 320, 100)];// imgView.backgroundColor=[UIColor grayColor];// [self.view addSubview:imgView];// [imgView release]; UIButton *cloeseBtn=[UIButton buttonWithType:UIButtonTypeRoundedRect]; cloeseBtn.frame=CGRectMake(10, 220, 300, 50); [cloeseBtn setTitle:@"Press" forState:UIControlStateNormal]; [cloeseBtn addTarget:self action:@selector(closeBtnClick:) forControlEvents:UIControlEventTouchUpInside]; [self.view addSubview:cloeseBtn];}-(void) closeBtnClick:(id) sender{ [session stopRunning];}- (void)didReceiveMemoryWarning{ [super didReceiveMemoryWarning]; // Dispose of any resources that can be recreated.}- (void)setupCaptureSession{ NSError *error = nil; // Create the session session = [[AVCaptureSession alloc] init]; // Configure the session to produce lower resolution video frames, if your // processing algorithm can cope. We'll specify medium quality for the // chosen device. session.sessionPreset = AVCaptureSessionPresetLow; // Find a suitable AVCaptureDevice AVCaptureDevice *device = [AVCaptureDevice defaultDeviceWithMediaType:AVMediaTypeVideo]; // Create a device input with the device and add it to the session. AVCaptureDeviceInput *input = [AVCaptureDeviceInput deviceInputWithDevice:device error:&error]; if (!input) { // Handling the error appropriately. } [session addInput:input]; // Create a VideoDataOutput and add it to the session AVCaptureVideoDataOutput *output = [[[AVCaptureVideoDataOutput alloc] init] autorelease]; [session addOutput:output]; // Configure your output. dispatch_queue_t queue = dispatch_queue_create("myQueue", NULL); [output setSampleBufferDelegate:self queue:queue]; dispatch_release(queue); // Specify the pixel format output.videoSettings = [NSDictionary dictionaryWithObject: [NSNumber numberWithInt:kCVPixelFormatType_32BGRA] forKey:(id)kCVPixelBufferPixelFormatTypeKey]; // If you wish to cap the frame rate to a known value, such as 15 fps, set // minFrameDuration. //output.minFrameDuration = CMTimeMake(1, 15); //AVCaptureConnection *avcaptureconn=[[AVCaptureConnection alloc] init]; //[avcaptureconn setVideoMinFrameDuration:CMTimeMake(1, 15)]; // Start the session running to start the flow of data [session startRunning]; AVCaptureVideoPreviewLayer* previewLayer = [AVCaptureVideoPreviewLayer layerWithSession: session]; previewLayer.frame = videoPreviewView.bounds; //视频显示到的UIView previewLayer.videoGravity = AVLayerVideoGravityResizeAspectFill;// [previewLayer setOrientation:AVCaptureVideoOrientationLandscapeRight]; // if(previewLayer.orientationSupported){ // previewLayer.orientation = mOrientation; // } [videoPreviewView.layer addSublayer: previewLayer]; if(![session isRunning]){ [session startRunning]; } // Assign session to an ivar. //[self setSession:session];}//得到视频流- (void)captureOutput:(AVCaptureOutput *)captureOutputdidOutputSampleBuffer:(CMSampleBufferRef)sampleBuffer fromConnection:(AVCaptureConnection *)connection{ // Create a UIImage from the sample buffer data return; UIImage *image = [self imageFromSampleBuffer:sampleBuffer]; //得到的视频流图片 imgView.image=image;}// Create a UIImage from sample buffer data- (UIImage *) imageFromSampleBuffer:(CMSampleBufferRef) sampleBuffer{ // Get a CMSampleBuffer's Core Video image buffer for the media data CVImageBufferRef imageBuffer = CMSampleBufferGetImageBuffer(sampleBuffer); // Lock the base address of the pixel buffer CVPixelBufferLockBaseAddress(imageBuffer, 0); // Get the number of bytes per row for the pixel buffer void *baseAddress = CVPixelBufferGetBaseAddress(imageBuffer); // Get the number of bytes per row for the pixel buffer size_t bytesPerRow = CVPixelBufferGetBytesPerRow(imageBuffer); // Get the pixel buffer width and height size_t width = CVPixelBufferGetWidth(imageBuffer); size_t height = CVPixelBufferGetHeight(imageBuffer); // Create a device-dependent RGB color space CGColorSpaceRef colorSpace = CGColorSpaceCreateDeviceRGB(); // Create a bitmap graphics context with the sample buffer data CGContextRef context = CGBitmapContextCreate(baseAddress, width, height, 8, bytesPerRow, colorSpace, kCGBitmapByteOrder32Little | kCGImageAlphaPremultipliedFirst); // Create a Quartz image from the pixel data in the bitmap graphics context CGImageRef quartzImage = CGBitmapContextCreateImage(context); // Unlock the pixel buffer CVPixelBufferUnlockBaseAddress(imageBuffer,0); // Free up the context and color space CGContextRelease(context); CGColorSpaceRelease(colorSpace); // Create an image object from the Quartz image UIImage *image = [UIImage imageWithCGImage:quartzImage]; // Release the Quartz image CGImageRelease(quartzImage); return (image);}@end

以上就是本文的全部内容,希望对大家的学习有所帮助,也希望大家多多支持我们。

以上就上有关iOS使用AVFoundation展示视频的相关介绍,要了解更多ios, 视频, AVFoundation内容请登录学步园。

抱歉!评论已关闭.