高光谱成像
激光雷达
人工智能
计算机科学
模式识别(心理学)
情态动词
特征提取
对偶(语法数字)
特征(语言学)
融合
遥感
地质学
材料科学
艺术
语言学
哲学
文学类
高分子化学
作者
Wuli Wang,C.-H. Li,Peng Ren,Xinchao Lu,Jianbu Wang,Guangbo Ren,Baodi Liu
标识
DOI:10.1109/lgrs.2024.3367171
摘要
The joint classification of hyperspectral image (HSI) and light detection and ranging (LiDAR) data has attracted considerable attention in the field of remote sensing. Integrating the advantages of the two data sources can provide precise data support and analytical decision-making for remote-sensing applications. However, due to the inherent differences in properties and semantic information from heterogeneous data, most existing deep-learning methods suboptimally extract the characteristic features of both data sources while utilizing their interactive information. In this letter, we propose a dual-branch feature fusion network-based cross-modal enhanced CNN and Transformer (DF2NCECT) to make full use of the respective features and interactive information of multisource data. DF2NCECT consists of two main stages. One is the basic feature extraction stage, which builds a hybrid convolution module based on 3DCNN and inception structure to fully extract the joint features of HSI from multiple spatial perspectives. The other is the deep feature fusion stage, where the CNN and Transformer are designed in parallel to fully explore and fuse deep features between HSI and LiDAR. More importantly, to achieve efficacious interactive information between HSI and LiDAR, a cross-modal enhanced CNN and Transformer module (CECT) is designed to deeply enhance the fused interactive features from global/local perspectives. Experiments show that the proposed method is superior and outperforms the comparison methods by an average of 3.06% in OA on Houston2013 and 1.79% on Summer, respectively.
科研通智能强力驱动
Strongly Powered by AbleSci AI