计算机科学
人工智能
计算机视觉
目标检测
跟踪(教育)
视频跟踪
特征(语言学)
嵌入
情态动词
弹道
探测器
利用
对象(语法)
模式识别(心理学)
哲学
物理
电信
化学
高分子化学
语言学
计算机安全
教育学
心理学
天文
作者
Yihan Zeng,Chao Ma,Ming Zhu,Zhiming Fan,Xiaokang Yang
标识
DOI:10.1109/iros51168.2021.9636498
摘要
Detecting and tracking objects in 3D scenes play crucial roles in autonomous driving. Successfully recognizing objects through space and time hinges on a strong detector and a reliable association scheme. Recent 3D detection and tracking approaches widely represent objects as points when associating detection results with trajectories. Despite the demonstrated success, these approaches do not fully exploit the rich appearance information of objects. In this paper, we present a conceptually simple yet effective algorithm, named AlphaTrack, which considers both the location and appearance changes to perform joint 3D object detection and tracking. To achieve this, we propose a cross-modal fusion scheme that fuses camera appearance feature with LiDAR feature to facilitate 3D detection and tracking. We further attach an additional branch to the 3D detector to output instance-aware appearance embedding, which significantly improves tracking performance with our designed association mechanisms. Extensive validations on large-scale autonomous driving dataset demonstrate the effectiveness of the proposed algorithm in comparison with state-of-the-art approaches. Notably, the proposed algorithm ranks first on the nuScenes tracking leaderboard to date.
科研通智能强力驱动
Strongly Powered by AbleSci AI