计算机科学
人工智能
一般化
深度图
RGB颜色模型
编码器
特征(语言学)
代表(政治)
计算机视觉
失真(音乐)
还原(数学)
模式识别(心理学)
图像(数学)
数学
数学分析
放大器
语言学
哲学
计算机网络
几何学
带宽(计算)
政治
政治学
法学
操作系统
作者
Xiangyin Meng,Jie Wen,Yang Li,Chenlong Wang,Jingzhen Zhang
标识
DOI:10.1016/j.cviu.2023.103914
摘要
Transparent objects play a vital role in modern industries and find widespread applications across various engineering scenarios. However, capturing accurate depth maps of transparent objects remains challenging due to their reflective and refractive properties, which pose difficulties for most commercial-grade optical sensors. In this paper, we propose a novel depth estimation method called DFNet-Trans, designed to estimate depth from a noisy RGB-D image input. Initially, a multiscale feature fusion module (FFM) is incorporated into the existing depth estimation network to generate the initial depth map. Subsequently, we enhance the network by adding a confidence branch and a mask branch on the same encoder, enabling improved distortion correction and real scene restoration in the depth estimation. Based on the framework representation, missing depth can be completed. Comprehensive experiments demonstrate that the proposed approach significantly outperforms the current state-of-the-art methods on the recently popular large-scale real dataset TransCG. the proposed approach achieves a remarkable 27.7% reduction in RMSE and a notable 34.6% reduction in REL. The generalization experiment shows that the proposed approach outperforms existing methods when generalized to an unknown real dataset.
科研通智能强力驱动
Strongly Powered by AbleSci AI