点云
分割
计算机科学
人工智能
棱锥(几何)
特征(语言学)
模式识别(心理学)
背景(考古学)
计算机视觉
体素
比例(比率)
代表(政治)
相似性(几何)
领域(数学)
地理
地图学
图像(数学)
数学
几何学
纯数学
法学
考古
政治学
哲学
语言学
政治
作者
Huchen Li,Haiyan Guan,Lingfei Ma,Xiangda Lei,Yongtao Yu,Hanyun Wang,M. R. Delavar,Jonathan Li
出处
期刊:International journal of applied earth observation and geoinformation
日期:2023-06-19
卷期号:122: 103391-103391
被引量:17
标识
DOI:10.1016/j.jag.2023.103391
摘要
Point cloud semantic segmentation, which contributes to scene understanding at different scales, is crucial for three-dimensional reconstruction and digital twin cities. However, current semantic segmentation methods mostly extract multi-scale features by down-sampling operations, but the feature maps only have a single receptive field at the same scale, resulting in the misclassification of objects with spatial similarity. To effectively capture the geometric features and the semantic information of different receptive fields, a multi-scale voxel-point adaptive fusion network (MVP-Net) is proposed for point cloud semantic segmentation in urban scenes. First, a multi-scale voxel fusion module with gating mechanism is designed to explore the semantic representation ability of different receptive fields. Then, a geometric self-attention module is constructed to deeply fuse fine-grained point features with coarse-grained voxel features. Finally, a pyramid decoder is introduced to aggregate context information at different scales for enhancing feature representation. The proposed MVP-Net was evaluated on three datasets, Toronto3D, WHU-MLS, and SensatUrban, and achieved superior performance in comparison to the state-of-the-art (SOTA) methods. For the public Toronto3D and SensatUrban datasets, our MVP-Net achieved a mIoU of 84.14% and 59.40%, and an overall accuracy of 98.12% and 93.30%, respectively.
科研通智能强力驱动
Strongly Powered by AbleSci AI