diff --git a/SwiftProject/SwiftProject.xcworkspace/xcuserdata/aaa.xcuserdatad/UserInterfaceState.xcuserstate b/SwiftProject/SwiftProject.xcworkspace/xcuserdata/aaa.xcuserdatad/UserInterfaceState.xcuserstate index 5771afd..11bf857 100644 Binary files a/SwiftProject/SwiftProject.xcworkspace/xcuserdata/aaa.xcuserdatad/UserInterfaceState.xcuserstate and b/SwiftProject/SwiftProject.xcworkspace/xcuserdata/aaa.xcuserdatad/UserInterfaceState.xcuserstate differ diff --git a/SwiftProject/SwiftProject.xcworkspace/xcuserdata/aaa.xcuserdatad/xcdebugger/Breakpoints_v2.xcbkptlist b/SwiftProject/SwiftProject.xcworkspace/xcuserdata/aaa.xcuserdatad/xcdebugger/Breakpoints_v2.xcbkptlist index 09ed0fc..e76c3a8 100644 --- a/SwiftProject/SwiftProject.xcworkspace/xcuserdata/aaa.xcuserdatad/xcdebugger/Breakpoints_v2.xcbkptlist +++ b/SwiftProject/SwiftProject.xcworkspace/xcuserdata/aaa.xcuserdatad/xcdebugger/Breakpoints_v2.xcbkptlist @@ -597,7 +597,7 @@ endingColumnNumber = "9223372036854775807" startingLineNumber = "20" endingLineNumber = "20" - landmarkName = "convertVideo(asset:assetOutput:type:time:)" + landmarkName = "convertVideo(asset:assetOutput:type:time:scale:ed:)" landmarkType = "7"> diff --git a/SwiftProject/SwiftProject/Project/Controller/RecordingVideo/CCSpatialVideoDisplayController/ZZHCustomPlayer.swift b/SwiftProject/SwiftProject/Project/Controller/RecordingVideo/CCSpatialVideoDisplayController/ZZHCustomPlayer.swift index d6d3de7..dc640a4 100644 --- a/SwiftProject/SwiftProject/Project/Controller/RecordingVideo/CCSpatialVideoDisplayController/ZZHCustomPlayer.swift +++ b/SwiftProject/SwiftProject/Project/Controller/RecordingVideo/CCSpatialVideoDisplayController/ZZHCustomPlayer.swift @@ -41,7 +41,7 @@ class ZZHCustomPlayer: UIView { var timerForMaskView:Timer?//标记隐藏倒计时 var temScale:CGFloat = 1.0 var currentScale:CGFloat = 1.0//当前缩放值 - + var currentED:Int = 5//当前的瞳距 //ui var playerLayerBgView:UIView? var avPlayer:AVPlayer? @@ -97,7 +97,7 @@ class ZZHCustomPlayer: UIView { NotificationCenter.default.addObserver(self, selector: #selector(notification_PlayerEndTime(notification:)), name: AVPlayerItem.didPlayToEndTimeNotification, object: nil) let playerWidth = KScreenWidth - let playerHeight = 200 + let playerHeight = 230 playerLayerBgView = UIView() self.addSubview(playerLayerBgView!) playerLayerBgView!.snp.makeConstraints { make in @@ -276,7 +276,7 @@ class ZZHCustomPlayer: UIView { currentScale = currentScale - (temScale - sender.scale) currentScale = min(currentScale, 1.5) - currentScale = max(currentScale, 0.5) + currentScale = max(currentScale, 0.6) temScale = sender.scale print("\(sender.scale) currentScale....\(currentScale)") } diff --git a/SwiftProject/SwiftProject/Project/Controller/RecordingVideo/CCSpatialVideoDisplayController/ZZHCustomPlayerForVideoTask.swift b/SwiftProject/SwiftProject/Project/Controller/RecordingVideo/CCSpatialVideoDisplayController/ZZHCustomPlayerForVideoTask.swift index 36682c9..1523c45 100644 --- a/SwiftProject/SwiftProject/Project/Controller/RecordingVideo/CCSpatialVideoDisplayController/ZZHCustomPlayerForVideoTask.swift +++ b/SwiftProject/SwiftProject/Project/Controller/RecordingVideo/CCSpatialVideoDisplayController/ZZHCustomPlayerForVideoTask.swift @@ -100,26 +100,26 @@ extension ZZHCustomPlayer { var presentTime:CMTime? = nil switch mode { case .crossedEyes://交叉眼 - ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime) + ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime,scale: currentScale,ed:currentED) break case .fsbs://3d全宽 - ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime) + ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime,scale: currentScale,ed:currentED) break case .hsbs://3d半宽 - ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime) + ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime,scale: currentScale,ed:currentED) break case .parallelEyes://平行眼 - ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime) + ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime,scale: currentScale,ed:currentED) break case .redBlueSolid://红蓝立体 - ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime) + ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime,scale: currentScale,ed:currentED) break default: break diff --git a/SwiftProject/SwiftProject/Project/Util/PlayByTransferConvertor.swift b/SwiftProject/SwiftProject/Project/Util/PlayByTransferConvertor.swift index 16a0c0f..1de13c2 100644 --- a/SwiftProject/SwiftProject/Project/Util/PlayByTransferConvertor.swift +++ b/SwiftProject/SwiftProject/Project/Util/PlayByTransferConvertor.swift @@ -14,7 +14,7 @@ import ImageIO class PlayByTransferConvertor { - func convertVideo(asset:AVAsset, assetOutput:AVAssetReaderTrackOutput,type:SpatialType,time: CMTime)->(CIImage?) { + func convertVideo(asset:AVAsset, assetOutput:AVAssetReaderTrackOutput,type:SpatialType,time: CMTime,scale:CGFloat,ed:Int)->(CIImage?) { var newpb:CIImage? = nil // var presentationTime:CMTime? = nil // print("sta.....>>>>>>>thread") @@ -65,10 +65,10 @@ class PlayByTransferConvertor { break case .parallelEyes://平行眼 - newpb = joinImages(leftImage: lciImage, rightImage: rciImage) + newpb = joinImages(leftImage: lciImage, rightImage: rciImage,scale: scale,ed: ed) break case .crossedEyes://交叉眼 - newpb = joinImages(leftImage: rciImage, rightImage: lciImage) + newpb = joinImages(leftImage: rciImage, rightImage: lciImage,scale: scale,ed: ed) break case .redBlueSolid://红蓝立体 newpb = joinImages_red_blue(lciImage: lciImage, rciImage: rciImage) @@ -175,9 +175,9 @@ class PlayByTransferConvertor { } //将两张图片合成一张图片 - func joinImages( leftImage:CIImage, rightImage:CIImage) -> CIImage { - let scale = 1.0 //缩放参数,此处必须大于0.5 - let ed = 5 //瞳距参数,暂定 瞳距与像素点的比例为1:2,瞳距的值范围为 -30 ~ +30 之间的整数 + func joinImages( leftImage:CIImage, rightImage:CIImage,scale:CGFloat,ed:Int) -> CIImage { +// let scale = 1.0 //缩放参数,此处必须大于0.5 +// let ed = 5 //瞳距参数,暂定 瞳距与像素点的比例为1:2,瞳距的值范围为 -30 ~ +30 之间的整数 let edS = 30 //比列 let left = UIImage(ciImage: leftImage )