计算机科学
激光雷达
分割
人工智能
点云
航程(航空)
计算机视觉
代表(政治)
点(几何)
深度学习
遥感
地理
数学
材料科学
几何学
政治
政治学
法学
复合材料
作者
Lingdong Kong,Youquan Liu,Runnan Chen,Yuexin Ma,Xinge Zhu,Yikang Li,Yuenan Hou,Yu Qiao,Ziwei Liu
标识
DOI:10.1109/iccv51070.2023.00028
摘要
LiDAR segmentation is crucial for autonomous driving perception. Recent trends favor point- or voxel-based methods as they often yield better performance than the traditional range view representation. In this work, we unveil several key factors in building powerful range view models. We observe that the "many-to-one" mapping, semantic incoherence, and shape deformation are possible impediments against effective learning from range view projections. We present RangeFormer – a full-cycle framework comprising novel designs across network architecture, data augmentation, and post-processing – that better handles the learning and processing of LiDAR point clouds from the range view. We further introduce a Scalable Training from Range view (STR) strategy that trains on arbitrary low-resolution 2D range images, while still maintaining satisfactory 3D segmentation accuracy. We show that, for the first time, a range view method is able to surpass the point, voxel, and multi-view fusion counterparts in the competing LiDAR semantic and panoptic segmentation benchmarks, i.e., SemanticKITTI, nuScenes, and ScribbleKITTI.
科研通智能强力驱动
Strongly Powered by AbleSci AI