Si quieres hacer pruebas por ti mismo y aprender en el camino, te sugiero que sigas este artículo que encontré en Medium. This approach is using native Swift code to extract camera frame, then you can process that frame with your own code.
- import UIKit
- import AVFoundation
- protocol FrameExtractorDelegate: class {
- func captured(image: UIImage)
- }
- clase FrameExtractor: NSObject, AVCaptureVideoDataOutputSampleBufferDelegate {
- private let position = AVCaptureDevicePosition.front
- private let quality = AVCaptureSessionPresetMedium
- private var permissionGranted = false
- private let sessionQueue = DispatchQueue(label: "session queue")
- private let captureSession = AVCaptureSession()
- private let context = CIContext()
- weak var delegate: FrameExtractorDelegate?
- override init() {
- super.init()
- checkPermission()
- sessionQueue.async { [unowned self] in
- self.configureSession()
- self.captureSession.startRunning()
- }
- }
- // MARK: AVSession configuration
- private func checkPermission() {
- switch AVCaptureDevice.authorizationStatus(forMediaType: AVMediaTypeVideo) {
- case .authorized:
- permissionGranted = true
- case .notDetermined:
- requestPermission()
- default:
- permissionGranted = false
- }
- }
- private func requestPermission() {
- sessionQueue.suspend()
- AVCaptureDevice.requestAccess(forMediaType: AVMediaTypeVideo) { [unowned self] granted in
- self.permissionGranted = granted
- self.sessionQueue.resume()
- }
- }
- private func configureSession() {
- guard permissionGranted else { return }
- captureSession.sessionPreset = quality
- guard let captureDevice = selectCaptureDevice() else { return }
- guard let captureDeviceInput = try? AVCaptureDeviceInput(device: captureDevice) else { return }
- guard captureSession.canAddInput(captureDeviceInput) else { return }
- captureSession.addInput(captureDeviceInput)
- let videoOutput = AVCaptureVideoDataOutput()
- videoOutput.setSampleBufferDelegate(self, queue: DispatchQueue(label: "sample buffer"))
- guard captureSession.canAddOutput(videoOutput) else { return }
- captureSession.addOutput(videoOutput)
- guard let connection = videoOutput.connection(withMediaType: AVFoundation.AVMediaTypeVideo) else { return }
- guard connection.isVideoOrientationSupported else { return }
- guard connection.isVideoMirroringSupported else { return }
- connection.videoOrientation = .portrait
- connection.isVideoMirrored = position == .front
- }
- private func selectCaptureDevice() -> AVCaptureDevice? {
- return AVCaptureDevice.devices().filter {
- ($0 as AnyObject).hasMediaType(AVMediaTypeVideo) &&
- ($0 as AnyObject).position == position
- }.first as? AVCaptureDevice
- }
- // MARK: Sample buffer to UIImage conversion
- private func imageFromSampleBuffer(sampleBuffer: CMSampleBuffer) -> UIImage? {
- guard let imageBuffer = CMSampleBufferGetImageBuffer(sampleBuffer) else { return nil }
- let ciImage = CIImage(cvPixelBuffer: imageBuffer)
- guard let cgImage = context.createCGImage(ciImage, from: ciImage.extent) else { return nil }
- return UIImage(cgImage: cgImage)
- }
- // MARK: AVCaptureVideoDataOutputSampleBufferDelegate
- func captureOutput(_ captureOutput: AVCaptureOutput!, didOutputSampleBuffer sampleBuffer: CMSampleBuffer!, from connection: AVCaptureConnection!) {
- guard let uiImage = imageFromSampleBuffer(sampleBuffer: sampleBuffer) else { return }
- DispatchQueue.main.async { [unowned self] in
- self.delegate?.captured(image: uiImage)
- }
- }
- }
iOS — Camera Frames Extraction – iOS App Development – Medium