将接口增加瞳距、scale参数

This commit is contained in:
bluesea 2024-04-22 18:25:57 +08:00
parent 1f1d2282e9
commit 1623bfdfd0
5 changed files with 15 additions and 15 deletions

View File

@ -597,7 +597,7 @@
endingColumnNumber = "9223372036854775807" endingColumnNumber = "9223372036854775807"
startingLineNumber = "20" startingLineNumber = "20"
endingLineNumber = "20" endingLineNumber = "20"
landmarkName = "convertVideo(asset:assetOutput:type:time:)" landmarkName = "convertVideo(asset:assetOutput:type:time:scale:ed:)"
landmarkType = "7"> landmarkType = "7">
</BreakpointContent> </BreakpointContent>
</BreakpointProxy> </BreakpointProxy>

View File

@ -41,7 +41,7 @@ class ZZHCustomPlayer: UIView {
var timerForMaskView:Timer?// var timerForMaskView:Timer?//
var temScale:CGFloat = 1.0 var temScale:CGFloat = 1.0
var currentScale:CGFloat = 1.0// var currentScale:CGFloat = 1.0//
var currentED:Int = 5//
//ui //ui
var playerLayerBgView:UIView? var playerLayerBgView:UIView?
var avPlayer:AVPlayer? var avPlayer:AVPlayer?
@ -97,7 +97,7 @@ class ZZHCustomPlayer: UIView {
NotificationCenter.default.addObserver(self, selector: #selector(notification_PlayerEndTime(notification:)), name: AVPlayerItem.didPlayToEndTimeNotification, object: nil) NotificationCenter.default.addObserver(self, selector: #selector(notification_PlayerEndTime(notification:)), name: AVPlayerItem.didPlayToEndTimeNotification, object: nil)
let playerWidth = KScreenWidth let playerWidth = KScreenWidth
let playerHeight = 200 let playerHeight = 230
playerLayerBgView = UIView() playerLayerBgView = UIView()
self.addSubview(playerLayerBgView!) self.addSubview(playerLayerBgView!)
playerLayerBgView!.snp.makeConstraints { make in playerLayerBgView!.snp.makeConstraints { make in
@ -276,7 +276,7 @@ class ZZHCustomPlayer: UIView {
currentScale = currentScale - (temScale - sender.scale) currentScale = currentScale - (temScale - sender.scale)
currentScale = min(currentScale, 1.5) currentScale = min(currentScale, 1.5)
currentScale = max(currentScale, 0.5) currentScale = max(currentScale, 0.6)
temScale = sender.scale temScale = sender.scale
print("\(sender.scale) currentScale....\(currentScale)") print("\(sender.scale) currentScale....\(currentScale)")
} }

View File

@ -100,26 +100,26 @@ extension ZZHCustomPlayer {
var presentTime:CMTime? = nil var presentTime:CMTime? = nil
switch mode { switch mode {
case .crossedEyes:// case .crossedEyes://
ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime) ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime,scale: currentScale,ed:currentED)
break break
case .fsbs://3d case .fsbs://3d
ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime) ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime,scale: currentScale,ed:currentED)
break break
case .hsbs://3d case .hsbs://3d
ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime) ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime,scale: currentScale,ed:currentED)
break break
case .parallelEyes:// case .parallelEyes://
ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime) ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime,scale: currentScale,ed:currentED)
break break
case .redBlueSolid:// case .redBlueSolid://
ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime) ciImg = self.videoTranserConvertor.convertVideo(asset: self.videoOriginalAsset, assetOutput: ao, type: self.selectedIndex, time: compositionTime,scale: currentScale,ed:currentED)
break break
default: default:
break break

View File

@ -14,7 +14,7 @@ import ImageIO
class PlayByTransferConvertor { class PlayByTransferConvertor {
func convertVideo(asset:AVAsset, assetOutput:AVAssetReaderTrackOutput,type:SpatialType,time: CMTime)->(CIImage?) { func convertVideo(asset:AVAsset, assetOutput:AVAssetReaderTrackOutput,type:SpatialType,time: CMTime,scale:CGFloat,ed:Int)->(CIImage?) {
var newpb:CIImage? = nil var newpb:CIImage? = nil
// var presentationTime:CMTime? = nil // var presentationTime:CMTime? = nil
// print("sta.....>>>>>>>thread") // print("sta.....>>>>>>>thread")
@ -65,10 +65,10 @@ class PlayByTransferConvertor {
break break
case .parallelEyes:// case .parallelEyes://
newpb = joinImages(leftImage: lciImage, rightImage: rciImage) newpb = joinImages(leftImage: lciImage, rightImage: rciImage,scale: scale,ed: ed)
break break
case .crossedEyes:// case .crossedEyes://
newpb = joinImages(leftImage: rciImage, rightImage: lciImage) newpb = joinImages(leftImage: rciImage, rightImage: lciImage,scale: scale,ed: ed)
break break
case .redBlueSolid:// case .redBlueSolid://
newpb = joinImages_red_blue(lciImage: lciImage, rciImage: rciImage) newpb = joinImages_red_blue(lciImage: lciImage, rciImage: rciImage)
@ -175,9 +175,9 @@ class PlayByTransferConvertor {
} }
// //
func joinImages( leftImage:CIImage, rightImage:CIImage) -> CIImage { func joinImages( leftImage:CIImage, rightImage:CIImage,scale:CGFloat,ed:Int) -> CIImage {
let scale = 1.0 //,0.5 // let scale = 1.0 //,0.5
let ed = 5 //, 1:2, -30 ~ +30 // let ed = 5 //, 1:2, -30 ~ +30
let edS = 30 // let edS = 30 //
let left = UIImage(ciImage: leftImage ) let left = UIImage(ciImage: leftImage )