计算机科学
分割
人工智能
量化(信号处理)
目标检测
计算机视觉
感知
任务(项目管理)
过程(计算)
机器学习
模式识别(心理学)
工程类
生物
操作系统
神经科学
系统工程
作者
Chi–Chih Chang,Wei–Cheng Lin,Pei–Shuo Wang,Sheng–Feng Yu,Yuchen Lu,Kuan–Cheng Lin,Kai–Chiang Wu
标识
DOI:10.1109/icmew59549.2023.00015
摘要
In this work, we present an efficient and quantization-aware panoptic driving perception model (Q-YOLOP) for object detection, drivable area segmentation, and lane line segmentation, in the context of autonomous driving. Our model employs the Efficient Layer Aggregation Network (ELAN) as its backbone and task-specific heads for each task. We employ a four-stage training process that includes pretraining on the BDD100K dataset, finetuning on both the BDD100K and iVS datasets, and quantization-aware training (QAT) on BDD100K. During the training process, we use powerful data augmentation techniques, such as random perspective and mosaic, and train the model on a combination of the BDD100K and iVS datasets. Both strategies enhance the model's generalization capabilities. The proposed model achieves state-of-the-art performance with an mAP@0.5 of 0.622 for object detection and an mIoU of 0.612 for segmentation, while maintaining low computational and memory requirements.
科研通智能强力驱动
Strongly Powered by AbleSci AI