保险丝(电气)
计算机科学
人工智能
雷达
计算机视觉
传感器融合
过程(计算)
分割
目标检测
工程类
电信
电气工程
操作系统
作者
Shanliang Yao,Runwei Guan,Xiaoyu Huang,Zhuoxiao Li,Xiangyu Sha,Yong Yue,Eng Gee Lim,Hyungjoon Seo,Ka Lok Man,Xiaohui Zhu,Yutao Yue
出处
期刊:Cornell University - arXiv
日期:2023-01-01
被引量:2
标识
DOI:10.48550/arxiv.2304.10410
摘要
Driven by deep learning techniques, perception technology in autonomous driving has developed rapidly in recent years, enabling vehicles to accurately detect and interpret surrounding environment for safe and efficient navigation. To achieve accurate and robust perception capabilities, autonomous vehicles are often equipped with multiple sensors, making sensor fusion a crucial part of the perception system. Among these fused sensors, radars and cameras enable a complementary and cost-effective perception of the surrounding environment regardless of lighting and weather conditions. This review aims to provide a comprehensive guideline for radar-camera fusion, particularly concentrating on perception tasks related to object detection and semantic segmentation.Based on the principles of the radar and camera sensors, we delve into the data processing process and representations, followed by an in-depth analysis and summary of radar-camera fusion datasets. In the review of methodologies in radar-camera fusion, we address interrogative questions, including "why to fuse", "what to fuse", "where to fuse", "when to fuse", and "how to fuse", subsequently discussing various challenges and potential research directions within this domain. To ease the retrieval and comparison of datasets and fusion methods, we also provide an interactive website: https://radar-camera-fusion.github.io.
科研通智能强力驱动
Strongly Powered by AbleSci AI