人工智能
计算机视觉
计算机科学
RGB颜色模型
匹配(统计)
平滑度
翻译(生物学)
基本事实
模式识别(心理学)
数学
生物化学
基因
信使核糖核酸
统计
数学分析
化学
作者
Xiaolong Liang,Cheolkon Jung
出处
期刊:IEEE robotics and automation letters
日期:2022-03-01
卷期号:7 (2): 5373-5380
被引量:12
标识
DOI:10.1109/lra.2022.3155202
摘要
Cross spectral stereo matching aims to estimate disparity from color (RGB) and near-infrared (NIR) image pairs. The main difference from traditional stereo matching is that there is a big gap between two spectral bands, which makes cross spectral stereo matching a challenging task. In this letter, we propose deep cross spectral stereo matching using multi-spectral image fusion. We adopt unsupervised learning to consider no ground truth in cross spectral stereo matching. We perform multi-spectral image fusion after cross spectral image translation to bridge the spectral gap between two images. First, we extract features from input RGB and NIR images to get fusion stereo pairs. Second, we get stereo correspondence robust to disparity variation based on parallax attention. In the loss function, we combine four losses: view reconstruction, material aware matching, cycle disparity consistency, and smoothness. We use a view reconstruction loss for spectral translation and fusion to warp stereo image pairs for appearance matching, while we use a material aware matching loss to take material property into consideration. Moreover, we utilize a cycle disparity consistency loss for disparity consistency between left and right predictions, and use a smoothness loss to enforce disparity smoothness. Experimental results show that the proposed network successfully estimates disparity with adaptability to materials and outperforms state-of-the-art models in terms of visual quality and quantitative measurements.
科研通智能强力驱动
Strongly Powered by AbleSci AI