计算机科学
人工智能
视频去噪
降噪
渲染(计算机图形)
卷积神经网络
失真(音乐)
计算机视觉
深度学习
视频质量
模式识别(心理学)
视频处理
公制(单位)
视频跟踪
多视点视频编码
带宽(计算)
经济
放大器
计算机网络
运营管理
作者
Huan Zhang,Yun Zhang,Linwei Zhu,Weisi Lin
标识
DOI:10.1109/tcsvt.2022.3147788
摘要
Due to occlusion among views and temporal inconsistency in depth video, spatio-temporal distortion occurs in 3D synthesized video with depth image-based rendering. In this paper, we propose a deep Convolutional Neural Network (CNN)-based synthesized video denoising algorithm to reduce temporal flicker distortion and improve perceptual quality of 3D synthesized video. First, we analyze the spatio-temporal distortion, and model eliminating spatio-temporal distortion as a perceptual video denoising problem. Then, a deep learning-based synthesized video denoising network is proposed, in which a CNN-friendly spatio-temporal loss function is derived from a synthesized video quality metric and integrated with a single image denoising network architecture. Finally, specific schemes, i.e., specific Synthesized Video Denoising Networks (SynVD-Nets), and a general scheme, i.e., General SynVD-Net (GSynVD-Net), based on existing CNN-based denoising models, are developed to handle synthesized video with different distortion levels more effectively. Experimental results show that the proposed SynVD-Net and GSynVD-Net can outperform deep learning-based counterparts and conventional denoising methods, and significantly enhance perceptual quality of 3D synthesized video.
科研通智能强力驱动
Strongly Powered by AbleSci AI