多光谱图像
行人检测
计算机科学
分割
人工智能
水准点(测量)
计算机视觉
目标检测
模式识别(心理学)
传感器融合
特征(语言学)
融合
行人
地图学
地理
语言学
哲学
考古
作者
Dayan Guan,Yanpeng Cao,Jiangxin Yang,Yanlong Cao,Christel-Loïc Tisse
出处
期刊:Applied Optics
[Optica Publishing Group]
日期:2018-04-17
卷期号:57 (18): D108-D108
被引量:32
摘要
Recent research has demonstrated that the fusion of complementary information captured by multi-modal sensors (visible and infrared cameras) enables robust pedestrian detection under various surveillance situations (e.g., daytime and nighttime). In this paper, we investigate a number of fusion architectures in an attempt to identify the optimal way of incorporating multispectral information for joint semantic segmentation and pedestrian detection. We made two important findings: (1) the sum fusion strategy, which computes the sum of two feature maps at the same spatial locations, delivers the best performance of multispectral detection, while the most commonly used concatenation fusion surprisingly performs the worst; and (2) two-stream semantic segmentation without multispectral fusion is the most effective scheme to infuse semantic information as supervision for learning human-related features. Based on these studies, we present a unified multispectral fusion framework for joint training of semantic segmentation and target detection that outperforms state-of-the-art multispectral pedestrian detectors by a large margin on the KAIST benchmark dataset.
科研通智能强力驱动
Strongly Powered by AbleSci AI