Robust LiDAR-Camera Alignment With Modality Adapted Local-to-Global Representation

计算机科学 人工智能 激光雷达 计算机视觉 杠杆(统计) 模式识别(心理学) 遥感 地质学
作者
Angfan Zhu,Yang Xiao,Chengxin Liu,Zhiguo Cao
出处
期刊:IEEE Transactions on Circuits and Systems for Video Technology [Institute of Electrical and Electronics Engineers]
卷期号:33 (1): 59-73 被引量:10
标识
DOI:10.1109/tcsvt.2022.3197212
摘要

LiDAR-Camera alignment (LCA) is an important preprocessing procedure for fusing LiDAR and camera data. For it, one key issue is to extract unified cross-modality representation for characterizing the heterogeneous LiDAR and camera data effectively and robustly. The main challenge is to resist the modality gap and visual data degradation during feature learning, while still maintaining strong representative power. To address this, a novel modality adapted local-to-global representation learning method is proposed. The research efforts are paid in 2 main folders via modality adaptation and capturing global spatial context. First for modality gap resistance, LiDAR and camera data is projected into the same depth map domain for unified representation learning. Particularly, LiDAR data is converted to depth map according to pre-acquired extrinsic parameters. Thanks to the recent advantage of deep learning based monocular depth estimation, camera data is transformed into depth map in data driven manner, which is jointly optimized with LCA. Secondly to capture global spatial context, ViT (vision transformer) is introduced to LCA. The concept of LCA token is proposed for aggregating the local spatial patterns to form global spatial representation with transformer encoding. And, it is shared by all the samples. In this way, it can involve global sample-level information to leverage generalization ability. The experiments on KITTI dataset verify superiority of our proposition. Furthermore, the proposed approach is more robust to camera data degeneration (e.g., imaging blurring and noise) often faced by the practical applications. Under some challenging test cases, the performance advancement of our method is over $1.9~cm$ /4.1° on translation / rotation error. While our model size (8.77M) is much smaller than existing methods (e.g., LCCNet of 66.75M). The source code will be released at https://github.com/Zaf233/RLCA upon acceptance.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
科研通AI6.4应助你好耀眼采纳,获得10
1秒前
1秒前
FashionBoy应助心事全在脸上采纳,获得10
2秒前
忧虑的羊发布了新的文献求助10
2秒前
彭于晏应助李薇采纳,获得10
2秒前
2秒前
maoamo2024发布了新的文献求助10
2秒前
2秒前
3秒前
3秒前
3秒前
LLLKAIXINGUO完成签到,获得积分10
3秒前
可爱书本发布了新的文献求助10
3秒前
缥缈伟宸发布了新的文献求助10
3秒前
4秒前
哈哈发布了新的文献求助10
4秒前
4秒前
orixero应助CNS999采纳,获得10
5秒前
5秒前
量子星尘发布了新的文献求助10
5秒前
18275412695发布了新的文献求助10
5秒前
MLDBrook发布了新的文献求助10
5秒前
6秒前
MLDBrook发布了新的文献求助10
6秒前
伶俐的储完成签到,获得积分20
6秒前
传奇3应助一只长颈卢采纳,获得10
6秒前
6秒前
鱼乐乐完成签到,获得积分10
6秒前
完美世界应助成就的元槐采纳,获得10
7秒前
青雉发布了新的文献求助50
7秒前
7秒前
7秒前
瑶瑶发布了新的文献求助10
7秒前
lee完成签到,获得积分10
7秒前
8秒前
粱自中发布了新的文献求助10
8秒前
研友_LpQ3rn发布了新的文献求助10
8秒前
伶俐的储发布了新的文献求助20
9秒前
天天发布了新的文献求助10
9秒前
9秒前
高分求助中
Entre Praga y Madrid: los contactos checoslovaco-españoles (1948-1977) 1000
Polymorphism and polytypism in crystals 1000
Signals, Systems, and Signal Processing 610
Discrete-Time Signals and Systems 610
Hope Teacher Rating Scale 600
Death Without End: Korea and the Thanatographics of War 500
Der Gleislage auf der Spur 500
热门求助领域 (近24小时)
化学 材料科学 医学 生物 工程类 纳米技术 有机化学 物理 生物化学 化学工程 计算机科学 复合材料 内科学 催化作用 光电子学 物理化学 电极 冶金 遗传学 细胞生物学
热门帖子
关注 科研通微信公众号,转发送积分 6090051
求助须知:如何正确求助?哪些是违规求助? 7919717
关于积分的说明 16389590
捐赠科研通 5222234
什么是DOI,文献DOI怎么找? 2791752
邀请新用户注册赠送积分活动 1774617
关于科研通互助平台的介绍 1649820