当前位置: 移动技术网 > 移动技术>移动开发>IOS > iOS仿微信相机拍照、视频录制功能

iOS仿微信相机拍照、视频录制功能

2019年07月24日  | 移动技术网移动技术  | 我要评论

网上有很多自定义相机的例子,这里只是我临时写的一个ios自定义相机(仿微信)拍照、视频录制demo,仅供参考:

用到了下面几个库:

#import <avfoundation/avfoundation.h> 
#import <assetslibrary/assetslibrary.h>

在使用的时候需要在info.plist中把相关权限写进去:

privacy - microphone usage description 
privacy - photo library usage description 
privacy - camera usage description

我在写这个demo时,是按照微信的样式写的,同样是点击拍照、长按录制视频,视频录制完直接进行播放,这里封装了一个简易的播放器:

m文件

#import "havplayer.h"
#import <avfoundation/avfoundation.h>

@interface havplayer ()

@property (nonatomic,strong) avplayer *player;//播放器对象

@end

@implementation havplayer

/*
// only override drawrect: if you perform custom drawing.
// an empty implementation adversely affects performance during animation.
- (void)drawrect:(cgrect)rect {
 // drawing code
}
*/

- (instancetype)initwithframe:(cgrect)frame withshowinview:(uiview *)bgview url:(nsurl *)url {
 if (self = [self initwithframe:frame]) {
  //创建播放器层
  avplayerlayer *playerlayer = [avplayerlayer playerlayerwithplayer:self.player];
  playerlayer.frame = self.bounds;

  [self.layer addsublayer:playerlayer];
  if (url) {
   self.videourl = url;
  }

  [bgview addsubview:self];
 }
 return self;
}

- (void)dealloc {
 [self removeavplayerntf];
 [self stopplayer];
 self.player = nil;
}

- (avplayer *)player {
 if (!_player) {
  _player = [avplayer playerwithplayeritem:[self getavplayeritem]];
  [self addavplayerntf:_player.currentitem];

 }

 return _player;
}

- (avplayeritem *)getavplayeritem {
 avplayeritem *playeritem=[avplayeritem playeritemwithurl:self.videourl];
 return playeritem;
}

- (void)setvideourl:(nsurl *)videourl {
 _videourl = videourl;
 [self removeavplayerntf];
 [self nextplayer];
}

- (void)nextplayer {
 [self.player seektotime:cmtimemakewithseconds(0, _player.currentitem.duration.timescale)];
 [self.player replacecurrentitemwithplayeritem:[self getavplayeritem]];
 [self addavplayerntf:self.player.currentitem];
 if (self.player.rate == 0) {
  [self.player play];
 }
}

- (void) addavplayerntf:(avplayeritem *)playeritem {
 //监控状态属性
 [playeritem addobserver:self forkeypath:@"status" options:nskeyvalueobservingoptionnew context:nil];
 //监控网络加载情况属性
 [playeritem addobserver:self forkeypath:@"loadedtimeranges" options:nskeyvalueobservingoptionnew context:nil];

 [[nsnotificationcenter defaultcenter] addobserver:self selector:@selector(playbackfinished:) name:avplayeritemdidplaytoendtimenotification object:self.player.currentitem];
}

- (void)removeavplayerntf {
 avplayeritem *playeritem = self.player.currentitem;
 [playeritem removeobserver:self forkeypath:@"status"];
 [playeritem removeobserver:self forkeypath:@"loadedtimeranges"];
 [[nsnotificationcenter defaultcenter] removeobserver:self];
}

- (void)stopplayer {
 if (self.player.rate == 1) {
  [self.player pause];//如果在播放状态就停止
 }
}

/**
 * 通过kvo监控播放器状态
 *
 * @param keypath 监控属性
 * @param object 监视器
 * @param change 状态改变
 * @param context 上下文
 */
-(void)observevalueforkeypath:(nsstring *)keypath ofobject:(id)object change:(nsdictionary *)change context:(void *)context{
 avplayeritem *playeritem = object;
 if ([keypath isequaltostring:@"status"]) {
  avplayerstatus status= [[change objectforkey:@"new"] intvalue];
  if(status==avplayerstatusreadytoplay){
   nslog(@"正在播放...,视频总长度:%.2f",cmtimegetseconds(playeritem.duration));
  }
 }else if([keypath isequaltostring:@"loadedtimeranges"]){
  nsarray *array=playeritem.loadedtimeranges;
  cmtimerange timerange = [array.firstobject cmtimerangevalue];//本次缓冲时间范围
  float startseconds = cmtimegetseconds(timerange.start);
  float durationseconds = cmtimegetseconds(timerange.duration);
  nstimeinterval totalbuffer = startseconds + durationseconds;//缓冲总长度
  nslog(@"共缓冲:%.2f",totalbuffer);
 }
}

- (void)playbackfinished:(nsnotification *)ntf {
 plog(@"视频播放完成");
 [self.player seektotime:cmtimemake(0, 1)];
 [self.player play];
}

@end

另外微信下面的按钮长按会出现圆弧时间条:

m文件

#import "hprogressview.h"

@interface hprogressview ()

/**
 * 进度值0-1.0之间
 */
@property (nonatomic,assign)cgfloat progressvalue;

@property (nonatomic, assign) cgfloat currenttime;

@end

@implementation hprogressview


// only override drawrect: if you perform custom drawing.
// an empty implementation adversely affects performance during animation.
- (void)drawrect:(cgrect)rect {
 // drawing code
 cgcontextref ctx = uigraphicsgetcurrentcontext();//获取上下文
 plog(@"width = %f",self.frame.size.width);
 cgpoint center = cgpointmake(self.frame.size.width/2.0, self.frame.size.width/2.0); //设置圆心位置
 cgfloat radius = self.frame.size.width/2.0-5; //设置半径
 cgfloat starta = - m_pi_2; //圆起点位置
 cgfloat enda = -m_pi_2 + m_pi * 2 * _progressvalue; //圆终点位置

 uibezierpath *path = [uibezierpath bezierpathwitharccenter:center radius:radius startangle:starta endangle:enda clockwise:yes];

 cgcontextsetlinewidth(ctx, 10); //设置线条宽度
 [[uicolor whitecolor] setstroke]; //设置描边颜色

 cgcontextaddpath(ctx, path.cgpath); //把路径添加到上下文

 cgcontextstrokepath(ctx); //渲染
}

- (void)settimemax:(nsinteger)timemax {
 _timemax = timemax;
 self.currenttime = 0;
 self.progressvalue = 0;
 [self setneedsdisplay];
 self.hidden = no;
 [self performselector:@selector(startprogress) withobject:nil afterdelay:0.1];
}

- (void)clearprogress {
 _currenttime = _timemax;
 self.hidden = yes;
}

- (void)startprogress {
 _currenttime += 0.1;
 if (_timemax > _currenttime) {
  _progressvalue = _currenttime/_timemax;
  plog(@"progress = %f",_progressvalue);
  [self setneedsdisplay];
  [self performselector:@selector(startprogress) withobject:nil afterdelay:0.1];
 }

 if (_timemax <= _currenttime) {
  [self clearprogress];

 }
}

@end

接下来就是相机的控制器了,由于是临时写的,所以用的xib,大家不要直接使用,直接上m文件代码吧:

#import "hvideoviewcontroller.h"
#import <avfoundation/avfoundation.h>
#import "havplayer.h"
#import "hprogressview.h"
#import <foundation/foundation.h>
#import <assetslibrary/assetslibrary.h>

typedef void(^propertychangeblock)(avcapturedevice *capturedevice);
@interface hvideoviewcontroller ()<avcapturefileoutputrecordingdelegate>

//轻触拍照,按住摄像
@property (strong, nonatomic) iboutlet uilabel *labeltiptitle;

//视频输出流
@property (strong,nonatomic) avcapturemoviefileoutput *capturemoviefileoutput;
//图片输出流
//@property (strong,nonatomic) avcapturestillimageoutput *capturestillimageoutput;//照片输出流
//负责从avcapturedevice获得输入数据
@property (strong,nonatomic) avcapturedeviceinput *capturedeviceinput;
//后台任务标识
@property (assign,nonatomic) uibackgroundtaskidentifier backgroundtaskidentifier;

@property (assign,nonatomic) uibackgroundtaskidentifier lastbackgroundtaskidentifier;

@property (weak, nonatomic) iboutlet uiimageview *focuscursor; //聚焦光标

//负责输入和输出设备之间的数据传递
@property(nonatomic)avcapturesession *session;

//图像预览层,实时显示捕获的图像
@property(nonatomic)avcapturevideopreviewlayer *previewlayer;

@property (strong, nonatomic) iboutlet uibutton *btnback;
//重新录制
@property (strong, nonatomic) iboutlet uibutton *btnafresh;
//确定
@property (strong, nonatomic) iboutlet uibutton *btnensure;
//摄像头切换
@property (strong, nonatomic) iboutlet uibutton *btncamera;

@property (strong, nonatomic) iboutlet uiimageview *bgview;
//记录录制的时间 默认最大60秒
@property (assign, nonatomic) nsinteger seconds;

//记录需要保存视频的路径
@property (strong, nonatomic) nsurl *savevideourl;

//是否在对焦
@property (assign, nonatomic) bool isfocus;
@property (strong, nonatomic) iboutlet nslayoutconstraint *afreshcenterx;
@property (strong, nonatomic) iboutlet nslayoutconstraint *ensurecenterx;
@property (strong, nonatomic) iboutlet nslayoutconstraint *backcenterx;

//视频播放
@property (strong, nonatomic) havplayer *player;

@property (strong, nonatomic) iboutlet hprogressview *progressview;

//是否是摄像 yes 代表是录制 no 表示拍照
@property (assign, nonatomic) bool isvideo;

@property (strong, nonatomic) uiimage *takeimage;
@property (strong, nonatomic) uiimageview *takeimageview;
@property (strong, nonatomic) iboutlet uiimageview *imgrecord;


@end

//时间大于这个就是视频,否则为拍照
#define timemax 1

@implementation hvideoviewcontroller


-(void)dealloc{
 [self removenotification];


}

- (void)viewdidload {
 [super viewdidload];
 // do any additional setup after loading the view.

 uiimage *image = [uiimage imagenamed:@"sc_btn_take.png"];
 self.backcenterx.constant = -(screen_width/2/2)-image.size.width/2/2;

 self.progressview.layer.cornerradius = self.progressview.frame.size.width/2;

 if (self.hseconds == 0) {
  self.hseconds = 60;
 }

 [self performselector:@selector(hiddentipslabel) withobject:nil afterdelay:4];
}

- (void)hiddentipslabel {
 self.labeltiptitle.hidden = yes;
}

- (void)didreceivememorywarning {
 [super didreceivememorywarning];
 // dispose of any resources that can be recreated.
}

- (void)viewwillappear:(bool)animated {
 [super viewwillappear:animated];
 [[uiapplication sharedapplication] setstatusbarhidden:yes];
 [self customcamera];
 [self.session startrunning];
}


-(void)viewdidappear:(bool)animated{
 [super viewdidappear:animated];
}

-(void)viewdiddisappear:(bool)animated{
 [super viewdiddisappear:animated];
 [self.session stoprunning];
}

- (void)viewwilldisappear:(bool)animated {
 [super viewwilldisappear:animated];
 [[uiapplication sharedapplication] setstatusbarhidden:no];
}

- (void)customcamera {

 //初始化会话,用来结合输入输出
 self.session = [[avcapturesession alloc] init];
 //设置分辨率 (设备支持的最高分辨率)
 if ([self.session cansetsessionpreset:avcapturesessionpresethigh]) {
  self.session.sessionpreset = avcapturesessionpresethigh;
 }
 //取得后置摄像头
 avcapturedevice *capturedevice = [self getcameradevicewithposition:avcapturedevicepositionback];
 //添加一个音频输入设备
 avcapturedevice *audiocapturedevice=[[avcapturedevice deviceswithmediatype:avmediatypeaudio] firstobject];

 //初始化输入设备
 nserror *error = nil;
 self.capturedeviceinput = [[avcapturedeviceinput alloc] initwithdevice:capturedevice error:&error];
 if (error) {
  plog(@"取得设备输入对象时出错,错误原因:%@",error.localizeddescription);
  return;
 }

 //添加音频
 error = nil;
 avcapturedeviceinput *audiocapturedeviceinput=[[avcapturedeviceinput alloc]initwithdevice:audiocapturedevice error:&error];
 if (error) {
  nslog(@"取得设备输入对象时出错,错误原因:%@",error.localizeddescription);
  return;
 }

 //输出对象
 self.capturemoviefileoutput = [[avcapturemoviefileoutput alloc] init];//视频输出

 //将输入设备添加到会话
 if ([self.session canaddinput:self.capturedeviceinput]) {
  [self.session addinput:self.capturedeviceinput];
  [self.session addinput:audiocapturedeviceinput];
  //设置视频防抖
  avcaptureconnection *connection = [self.capturemoviefileoutput connectionwithmediatype:avmediatypevideo];
  if ([connection isvideostabilizationsupported]) {
   connection.preferredvideostabilizationmode = avcapturevideostabilizationmodecinematic;
  }
 }

 //将输出设备添加到会话 (刚开始 是照片为输出对象)
 if ([self.session canaddoutput:self.capturemoviefileoutput]) {
  [self.session addoutput:self.capturemoviefileoutput];
 }

 //创建视频预览层,用于实时展示摄像头状态
 self.previewlayer = [[avcapturevideopreviewlayer alloc] initwithsession:self.session];
 self.previewlayer.frame = self.view.bounds;//cgrectmake(0, 0, self.view.width, self.view.height);
 self.previewlayer.videogravity = avlayervideogravityresizeaspectfill;//填充模式
 [self.bgview.layer addsublayer:self.previewlayer];

 [self addnotificationtocapturedevice:capturedevice];
 [self addgensturerecognizer];
}

 

- (ibaction)oncancelaction:(uibutton *)sender {
 [self dismissviewcontrolleranimated:yes completion:^{
  [utility hideprogressdialog];
 }];
}


- (void)touchesbegan:(nsset<uitouch *> *)touches withevent:(uievent *)event {
 if ([[touches anyobject] view] == self.imgrecord) {
  plog(@"开始录制");
  //根据设备输出获得连接
  avcaptureconnection *connection = [self.capturemoviefileoutput connectionwithmediatype:avmediatypeaudio];
  //根据连接取得设备输出的数据
  if (![self.capturemoviefileoutput isrecording]) {
   //如果支持多任务则开始多任务
   if ([[uidevice currentdevice] ismultitaskingsupported]) {
    self.backgroundtaskidentifier = [[uiapplication sharedapplication] beginbackgroundtaskwithexpirationhandler:nil];
   }
   if (self.savevideourl) {
    [[nsfilemanager defaultmanager] removeitematurl:self.savevideourl error:nil];
   }
   //预览图层和视频方向保持一致
   connection.videoorientation = [self.previewlayer connection].videoorientation;
   nsstring *outputfielpath=[nstemporarydirectory() stringbyappendingstring:@"mymovie.mov"];
   nslog(@"save path is :%@",outputfielpath);
   nsurl *fileurl=[nsurl fileurlwithpath:outputfielpath];
   nslog(@"fileurl:%@",fileurl);
   [self.capturemoviefileoutput startrecordingtooutputfileurl:fileurl recordingdelegate:self];
  } else {
   [self.capturemoviefileoutput stoprecording];
  }
 }
}


- (void)touchesended:(nsset<uitouch *> *)touches withevent:(uievent *)event {
 if ([[touches anyobject] view] == self.imgrecord) {
  plog(@"结束触摸");
  if (!self.isvideo) {
   [self performselector:@selector(endrecord) withobject:nil afterdelay:0.3];
  } else {
   [self endrecord];
  }
 }
}

- (void)endrecord {
 [self.capturemoviefileoutput stoprecording];//停止录制
}

- (ibaction)onafreshaction:(uibutton *)sender {
 plog(@"重新录制");
 [self recoverlayout];
}

- (ibaction)onensureaction:(uibutton *)sender {
 plog(@"确定 这里进行保存或者发送出去");
 if (self.savevideourl) {
  ws(weakself)
  [utility showprogressdialogtext:@"视频处理中..."];
  alassetslibrary *assetslibrary=[[alassetslibrary alloc]init];
  [assetslibrary writevideoatpathtosavedphotosalbum:self.savevideourl completionblock:^(nsurl *asseturl, nserror *error) {
   plog(@"outputurl:%@",weakself.savevideourl);
   [[nsfilemanager defaultmanager] removeitematurl:weakself.savevideourl error:nil];
   if (weakself.lastbackgroundtaskidentifier!= uibackgroundtaskinvalid) {
    [[uiapplication sharedapplication] endbackgroundtask:weakself.lastbackgroundtaskidentifier];
   }
   if (error) {
    plog(@"保存视频到相簿过程中发生错误,错误信息:%@",error.localizeddescription);
    [utility showalltextdialog:kappdelegate.window text:@"保存视频到相册发生错误"];
   } else {
    if (weakself.takeblock) {
     weakself.takeblock(asseturl);
    }
    plog(@"成功保存视频到相簿.");
    [weakself oncancelaction:nil];
   }
  }];
 } else {
  //照片
  uiimagewritetosavedphotosalbum(self.takeimage, self, nil, nil);
  if (self.takeblock) {
   self.takeblock(self.takeimage);
  }

  [self oncancelaction:nil];
 }
}

//前后摄像头的切换
- (ibaction)oncameraaction:(uibutton *)sender {
 plog(@"切换摄像头");
 avcapturedevice *currentdevice=[self.capturedeviceinput device];
 avcapturedeviceposition currentposition=[currentdevice position];
 [self removenotificationfromcapturedevice:currentdevice];
 avcapturedevice *tochangedevice;
 avcapturedeviceposition tochangeposition = avcapturedevicepositionfront;//前
 if (currentposition == avcapturedevicepositionunspecified || currentposition == avcapturedevicepositionfront) {
  tochangeposition = avcapturedevicepositionback;//后
 }
 tochangedevice=[self getcameradevicewithposition:tochangeposition];
 [self addnotificationtocapturedevice:tochangedevice];
 //获得要调整的设备输入对象
 avcapturedeviceinput *tochangedeviceinput=[[avcapturedeviceinput alloc]initwithdevice:tochangedevice error:nil];

 //改变会话的配置前一定要先开启配置,配置完成后提交配置改变
 [self.session beginconfiguration];
 //移除原有输入对象
 [self.session removeinput:self.capturedeviceinput];
 //添加新的输入对象
 if ([self.session canaddinput:tochangedeviceinput]) {
  [self.session addinput:tochangedeviceinput];
  self.capturedeviceinput = tochangedeviceinput;
 }
 //提交会话配置
 [self.session commitconfiguration];
}

- (void)onstarttranscribe:(nsurl *)fileurl {
 if ([self.capturemoviefileoutput isrecording]) {
  -- self.seconds;
  if (self.seconds > 0) {
   if (self.hseconds - self.seconds >= timemax && !self.isvideo) {
    self.isvideo = yes;//长按时间超过timemax 表示是视频录制
    self.progressview.timemax = self.seconds;
   }
   [self performselector:@selector(onstarttranscribe:) withobject:fileurl afterdelay:1.0];
  } else {
   if ([self.capturemoviefileoutput isrecording]) {
    [self.capturemoviefileoutput stoprecording];
   }
  }
 }
}


#pragma mark - 视频输出代理
-(void)captureoutput:(avcapturefileoutput *)captureoutput didstartrecordingtooutputfileaturl:(nsurl *)fileurl fromconnections:(nsarray *)connections{
 plog(@"开始录制...");
 self.seconds = self.hseconds;
 [self performselector:@selector(onstarttranscribe:) withobject:fileurl afterdelay:1.0];
}


-(void)captureoutput:(avcapturefileoutput *)captureoutput didfinishrecordingtooutputfileaturl:(nsurl *)outputfileurl fromconnections:(nsarray *)connections error:(nserror *)error{
 plog(@"视频录制完成.");
 [self changelayout];
 if (self.isvideo) {
  self.savevideourl = outputfileurl;
  if (!self.player) {
   self.player = [[havplayer alloc] initwithframe:self.bgview.bounds withshowinview:self.bgview url:outputfileurl];
  } else {
   if (outputfileurl) {
    self.player.videourl = outputfileurl;
    self.player.hidden = no;
   }
  }
 } else {
  //照片
  self.savevideourl = nil;
  [self videohandlephoto:outputfileurl];
 }

}

- (void)videohandlephoto:(nsurl *)url {
 avurlasset *urlset = [avurlasset assetwithurl:url];
 avassetimagegenerator *imagegenerator = [avassetimagegenerator assetimagegeneratorwithasset:urlset];
 imagegenerator.appliespreferredtracktransform = yes; // 截图的时候调整到正确的方向
 nserror *error = nil;
 cmtime time = cmtimemake(0,30);//缩略图创建时间 cmtime是表示电影时间信息的结构体,第一个参数表示是视频第几秒,第二个参数表示每秒帧数.(如果要获取某一秒的第几帧可以使用cmtimemake方法)
 cmtime actucaltime; //缩略图实际生成的时间
 cgimageref cgimage = [imagegenerator copycgimageattime:time actualtime:&actucaltime error:&error];
 if (error) {
  plog(@"截取视频图片失败:%@",error.localizeddescription);
 }
 cmtimeshow(actucaltime);
 uiimage *image = [uiimage imagewithcgimage:cgimage];

 cgimagerelease(cgimage);
 if (image) {
  plog(@"视频截取成功");
 } else {
  plog(@"视频截取失败");
 }


 self.takeimage = image;//[uiimage imagewithcgimage:cgimage];

 [[nsfilemanager defaultmanager] removeitematurl:url error:nil];

 if (!self.takeimageview) {
  self.takeimageview = [[uiimageview alloc] initwithframe:self.view.frame];
  [self.bgview addsubview:self.takeimageview];
 }
 self.takeimageview.hidden = no;
 self.takeimageview.image = self.takeimage;
}

#pragma mark - 通知

//注册通知
- (void)setupobservers
{
 nsnotificationcenter *notification = [nsnotificationcenter defaultcenter];
 [notification addobserver:self selector:@selector(applicationdidenterbackground:) name:uiapplicationwillresignactivenotification object:[uiapplication sharedapplication]];
}

//进入后台就退出视频录制
- (void)applicationdidenterbackground:(nsnotification *)notification {
 [self oncancelaction:nil];
}

/**
 * 给输入设备添加通知
 */
-(void)addnotificationtocapturedevice:(avcapturedevice *)capturedevice{
 //注意添加区域改变捕获通知必须首先设置设备允许捕获
 [self changedeviceproperty:^(avcapturedevice *capturedevice) {
  capturedevice.subjectareachangemonitoringenabled=yes;
 }];
 nsnotificationcenter *notificationcenter= [nsnotificationcenter defaultcenter];
 //捕获区域发生改变
 [notificationcenter addobserver:self selector:@selector(areachange:) name:avcapturedevicesubjectareadidchangenotification object:capturedevice];
}
-(void)removenotificationfromcapturedevice:(avcapturedevice *)capturedevice{
 nsnotificationcenter *notificationcenter= [nsnotificationcenter defaultcenter];
 [notificationcenter removeobserver:self name:avcapturedevicesubjectareadidchangenotification object:capturedevice];
}
/**
 * 移除所有通知
 */
-(void)removenotification{
 nsnotificationcenter *notificationcenter= [nsnotificationcenter defaultcenter];
 [notificationcenter removeobserver:self];
}

-(void)addnotificationtocapturesession:(avcapturesession *)capturesession{
 nsnotificationcenter *notificationcenter= [nsnotificationcenter defaultcenter];
 //会话出错
 [notificationcenter addobserver:self selector:@selector(sessionruntimeerror:) name:avcapturesessionruntimeerrornotification object:capturesession];
}

/**
 * 设备连接成功
 *
 * @param notification 通知对象
 */
-(void)deviceconnected:(nsnotification *)notification{
 nslog(@"设备已连接...");
}
/**
 * 设备连接断开
 *
 * @param notification 通知对象
 */
-(void)devicedisconnected:(nsnotification *)notification{
 nslog(@"设备已断开.");
}
/**
 * 捕获区域改变
 *
 * @param notification 通知对象
 */
-(void)areachange:(nsnotification *)notification{
 nslog(@"捕获区域改变...");
}

/**
 * 会话出错
 *
 * @param notification 通知对象
 */
-(void)sessionruntimeerror:(nsnotification *)notification{
 nslog(@"会话发生错误.");
}

 

/**
 * 取得指定位置的摄像头
 *
 * @param position 摄像头位置
 *
 * @return 摄像头设备
 */
-(avcapturedevice *)getcameradevicewithposition:(avcapturedeviceposition )position{
 nsarray *cameras= [avcapturedevice deviceswithmediatype:avmediatypevideo];
 for (avcapturedevice *camera in cameras) {
  if ([camera position] == position) {
   return camera;
  }
 }
 return nil;
}

/**
 * 改变设备属性的统一操作方法
 *
 * @param propertychange 属性改变操作
 */
-(void)changedeviceproperty:(propertychangeblock)propertychange{
 avcapturedevice *capturedevice= [self.capturedeviceinput device];
 nserror *error;
 //注意改变设备属性前一定要首先调用lockforconfiguration:调用完之后使用unlockforconfiguration方法解锁
 if ([capturedevice lockforconfiguration:&error]) {
  //自动白平衡
  if ([capturedevice iswhitebalancemodesupported:avcapturewhitebalancemodecontinuousautowhitebalance]) {
   [capturedevice setwhitebalancemode:avcapturewhitebalancemodecontinuousautowhitebalance];
  }
  //自动根据环境条件开启闪光灯
  if ([capturedevice isflashmodesupported:avcaptureflashmodeauto]) {
   [capturedevice setflashmode:avcaptureflashmodeauto];
  }

  propertychange(capturedevice);
  [capturedevice unlockforconfiguration];
 }else{
  nslog(@"设置设备属性过程发生错误,错误信息:%@",error.localizeddescription);
 }
}

/**
 * 设置闪光灯模式
 *
 * @param flashmode 闪光灯模式
 */
-(void)setflashmode:(avcaptureflashmode )flashmode{
 [self changedeviceproperty:^(avcapturedevice *capturedevice) {
  if ([capturedevice isflashmodesupported:flashmode]) {
   [capturedevice setflashmode:flashmode];
  }
 }];
}
/**
 * 设置聚焦模式
 *
 * @param focusmode 聚焦模式
 */
-(void)setfocusmode:(avcapturefocusmode )focusmode{
 [self changedeviceproperty:^(avcapturedevice *capturedevice) {
  if ([capturedevice isfocusmodesupported:focusmode]) {
   [capturedevice setfocusmode:focusmode];
  }
 }];
}
/**
 * 设置曝光模式
 *
 * @param exposuremode 曝光模式
 */
-(void)setexposuremode:(avcaptureexposuremode)exposuremode{
 [self changedeviceproperty:^(avcapturedevice *capturedevice) {
  if ([capturedevice isexposuremodesupported:exposuremode]) {
   [capturedevice setexposuremode:exposuremode];
  }
 }];
}
/**
 * 设置聚焦点
 *
 * @param point 聚焦点
 */
-(void)focuswithmode:(avcapturefocusmode)focusmode exposuremode:(avcaptureexposuremode)exposuremode atpoint:(cgpoint)point{
 [self changedeviceproperty:^(avcapturedevice *capturedevice) {
//  if ([capturedevice isfocuspointofinterestsupported]) {
//   [capturedevice setfocuspointofinterest:point];
//  }
//  if ([capturedevice isexposurepointofinterestsupported]) {
//   [capturedevice setexposurepointofinterest:point];
//  }
  if ([capturedevice isexposuremodesupported:exposuremode]) {
   [capturedevice setexposuremode:exposuremode];
  }
  if ([capturedevice isfocusmodesupported:focusmode]) {
   [capturedevice setfocusmode:focusmode];
  }
 }];
}

/**
 * 添加点按手势,点按时聚焦
 */
-(void)addgensturerecognizer{
 uitapgesturerecognizer *tapgesture=[[uitapgesturerecognizer alloc]initwithtarget:self action:@selector(tapscreen:)];
 [self.bgview addgesturerecognizer:tapgesture];
}

-(void)tapscreen:(uitapgesturerecognizer *)tapgesture{
 if ([self.session isrunning]) {
  cgpoint point= [tapgesture locationinview:self.bgview];
  //将ui坐标转化为摄像头坐标
  cgpoint camerapoint= [self.previewlayer capturedevicepointofinterestforpoint:point];
  [self setfocuscursorwithpoint:point];
  [self focuswithmode:avcapturefocusmodecontinuousautofocus exposuremode:avcaptureexposuremodecontinuousautoexposure atpoint:camerapoint];
 }
}

/**
 * 设置聚焦光标位置
 *
 * @param point 光标位置
 */
-(void)setfocuscursorwithpoint:(cgpoint)point{
 if (!self.isfocus) {
  self.isfocus = yes;
  self.focuscursor.center=point;
  self.focuscursor.transform = cgaffinetransformmakescale(1.25, 1.25);
  self.focuscursor.alpha = 1.0;
  [uiview animatewithduration:0.5 animations:^{
   self.focuscursor.transform = cgaffinetransformidentity;
  } completion:^(bool finished) {
   [self performselector:@selector(onhiddenfocuscursoraction) withobject:nil afterdelay:0.5];
  }];
 }
}

- (void)onhiddenfocuscursoraction {
 self.focuscursor.alpha=0;
 self.isfocus = no;
}

//拍摄完成时调用
- (void)changelayout {
 self.imgrecord.hidden = yes;
 self.btncamera.hidden = yes;
 self.btnafresh.hidden = no;
 self.btnensure.hidden = no;
 self.btnback.hidden = yes;
 if (self.isvideo) {
  [self.progressview clearprogress];
 }
 self.afreshcenterx.constant = -(screen_width/2/2);
 self.ensurecenterx.constant = screen_width/2/2;
 [uiview animatewithduration:0.25 animations:^{
  [self.view layoutifneeded];
 }];

 self.lastbackgroundtaskidentifier = self.backgroundtaskidentifier;
 self.backgroundtaskidentifier = uibackgroundtaskinvalid;
 [self.session stoprunning];
}


//重新拍摄时调用
- (void)recoverlayout {
 if (self.isvideo) {
  self.isvideo = no;
  [self.player stopplayer];
  self.player.hidden = yes;
 }
 [self.session startrunning];

 if (!self.takeimageview.hidden) {
  self.takeimageview.hidden = yes;
 }
// self.savevideourl = nil;
 self.afreshcenterx.constant = 0;
 self.ensurecenterx.constant = 0;
 self.imgrecord.hidden = no;
 self.btncamera.hidden = no;
 self.btnafresh.hidden = yes;
 self.btnensure.hidden = yes;
 self.btnback.hidden = no;
 [uiview animatewithduration:0.25 animations:^{
  [self.view layoutifneeded];
 }];
}

/*
#pragma mark - navigation

// in a storyboard-based application, you will often want to do a little preparation before navigation
- (void)prepareforsegue:(uistoryboardsegue *)segue sender:(id)sender {
 // get the new view controller using [segue destinationviewcontroller].
 // pass the selected object to the new view controller.
}
*/

@end

使用也挺简单:

- (ibaction)oncameraaction:(uibutton *)sender {
 //额 。。由于是demo,所以用的xib,大家根据需求自己更改,该demo只是提供一个思路,使用时不要直接拖入项目
 hvideoviewcontroller *ctrl = [[nsbundle mainbundle] loadnibnamed:@"hvideoviewcontroller" owner:nil options:nil].lastobject;
 ctrl.hseconds = 30;//设置可录制最长时间
 ctrl.takeblock = ^(id item) {
  if ([item iskindofclass:[nsurl class]]) {
   nsurl *videourl = item;
   //视频url

  } else {
   //图片

  }
 };
 [self presentviewcontroller:ctrl animated:yes completion:nil];
}

demo地址也给出来吧:不喜勿碰-_-\

kjcamera

自此就结束啦,写的比较简单,希望能帮助到大家,谢谢!

以上就是本文的全部内容,希望对大家的学习有所帮助,也希望大家多多支持移动技术网。

如对本文有疑问, 点击进行留言回复!!

相关文章:

验证码:
移动技术网