From 40b5d1d7c54a7a7283cbb57ae76ffb74161d9e61 Mon Sep 17 00:00:00 2001 From: bluesea <307723040@qq.com> Date: Wed, 13 Mar 2024 17:10:01 +0800 Subject: [PATCH] =?UTF-8?q?=E8=BE=B9=E8=BD=AC=E8=BE=B9=E6=92=AD=EF=BC=8C?= =?UTF-8?q?=E4=B8=8D=E8=83=BD=E7=AE=97=E4=BD=9C=E4=BC=98=E5=8C=96?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- .../CCSpatialVideoDisplayController.swift | 47 ++++++++++++++----- 1 file changed, 34 insertions(+), 13 deletions(-) diff --git a/SwiftProject/SwiftProject/Project/Controller/RecordingVideo/CCSpatialVideoDisplayController.swift b/SwiftProject/SwiftProject/Project/Controller/RecordingVideo/CCSpatialVideoDisplayController.swift index d84ec2c..a3b3871 100644 --- a/SwiftProject/SwiftProject/Project/Controller/RecordingVideo/CCSpatialVideoDisplayController.swift +++ b/SwiftProject/SwiftProject/Project/Controller/RecordingVideo/CCSpatialVideoDisplayController.swift @@ -286,39 +286,60 @@ class CCSpatialVideoDisplayController: BaseController, AVPlayerViewControllerDel // self.quickLoadAReader(timeRange: tr) - while let nextSampleBuffer = self.assetOutput!.copyNextSampleBuffer() { -// print("compositon......copy samplebuffer") - } -// if(self.assetReader?.status == .completed){ -// print("okkkssss...") - self.assetOutput?.reset(forReadingTimeRanges:[NSValue(timeRange: tr)]) -// } -// else{ -// -// print("status:\(self.assetReader?.error?.localizedDescription)") -// } - - var ciImg:CIImage? = nil switch self.selectedIndex { case .crossedEyes://交叉眼 + while let nextSampleBuffer = self.assetOutput!.copyNextSampleBuffer() { + } + self.assetOutput?.reset(forReadingTimeRanges:[NSValue(timeRange: tr)]) + + ciImg = videoTranserConvertor.convertVideo(asset: videoOriginalAsset, assetOutput: self.assetOutput!, type: self.selectedIndex, time: compositionTime) break + + case .fsbs: + while let nextSampleBuffer = self.assetOutput!.copyNextSampleBuffer() { + } + self.assetOutput?.reset(forReadingTimeRanges:[NSValue(timeRange: tr)]) + ciImg = videoTranserConvertor.convertVideo(asset: videoOriginalAsset, assetOutput: self.assetOutput!, type: self.selectedIndex, time: compositionTime) break + + case .hsbs: + while let nextSampleBuffer = self.assetOutput!.copyNextSampleBuffer() { + } + self.assetOutput?.reset(forReadingTimeRanges:[NSValue(timeRange: tr)]) + + ciImg = videoTranserConvertor.convertVideo(asset: videoOriginalAsset, assetOutput: self.assetOutput!, type: self.selectedIndex, time: compositionTime) break + + case .parallelEyes://平行眼 + while let nextSampleBuffer = self.assetOutput!.copyNextSampleBuffer() { +// print("平心眼...") + } + self.assetOutput?.reset(forReadingTimeRanges:[NSValue(timeRange: tr)]) + + ciImg = videoTranserConvertor.convertVideo(asset: videoOriginalAsset, assetOutput: self.assetOutput!, type: self.selectedIndex, time: compositionTime) break + + case .monocular2D: ciImg = request.sourceImage break + + case .redBlueSolid://红蓝立体 + while let nextSampleBuffer = self.assetOutput!.copyNextSampleBuffer() { + } + self.assetOutput?.reset(forReadingTimeRanges:[NSValue(timeRange: tr)]) + ciImg = videoTranserConvertor.convertVideo(asset: videoOriginalAsset, assetOutput: self.assetOutput!, type: self.selectedIndex, time: compositionTime) break