计算机视觉
计算机科学
人工智能
弹道
跟踪(教育)
对象(语法)
移动机器人
目标检测
激光雷达
航程(航空)
编码(集合论)
信号(编程语言)
视频跟踪
传感器融合
机器人
地理
遥感
分割
工程类
心理学
教育学
物理
集合(抽象数据类型)
天文
航空航天工程
程序设计语言
作者
Alexander Kim,Aljoša Ošep,Laura Leal-Taixé
标识
DOI:10.1109/icra48506.2021.9562072
摘要
Multi-object tracking (MOT) enables mobile robots to perform well-informed motion planning and navigation by localizing surrounding objects in 3D space and time. Existing methods rely on depth sensors (e.g., LiDAR) to detect and track targets in 3D space, but only up to a limited sensing range due to the sparsity of the signal. On the other hand, cameras provide a dense and rich visual signal that helps to localize even distant objects, but only in the image domain. In this paper, we propose EagerMOT, a simple tracking formulation that eagerly integrates all available object observations from both sensor modalities to obtain a well-informed interpretation of the scene dynamics. Using images, we can identify distant incoming objects, while depth estimates allow for precise trajectory localization as soon as objects are within the depth-sensing range. With EagerMOT, we achieve state-of-the-art results across several MOT tasks on the KITTI and NuScenes datasets. Our code is available at https://github.com/aleksandrkim61/EagerMOT
科研通智能强力驱动
Strongly Powered by AbleSci AI