高光谱成像
计算机科学
人工智能
特征(语言学)
遥感
模式识别(心理学)
融合
传感器融合
地质学
语言学
哲学
作者
Wei Gao,Yu Zhang,Youssef Akoudad,Jie Chen
标识
DOI:10.1109/tgrs.2025.3563647
摘要
Hyperspectral unmixing (HU) aims to decompose mixed pixels in remote sensing imagery into material-specific spectra and their respective abundance fractions. Recently, autoencoders have made significant advances in HU due to their strong representational capabilities and ease of implementation. However, relying exclusively on feature extraction from a single-modality hyperspectral image can fail to fully utilize both spatial and spectral information, thereby limiting the ability to distinguish objects in complex scenes. To address these limitations, we propose a multimodal spectral-spatial feature fusion network (MSSF-Net) for enhanced HU. MSSF-Net adopts a dual-stream architecture to extract feature representations from complementary input modalities. Specifically, the hyperspectral subnetwork leverages a convolutional neural network (CNN) to capture spatial information, while the light detection and ranging (LiDAR) subnetwork incorporates an enhanced channel attention mechanism (ECAM) to capture the dynamic changes in spatial information across different channels. Furthermore, we introduce a cross-modal fusion (CMF) module that integrates spectral and spatial information across modalities, leading to more robust feature representations. Experimental results indicate that MSSF-Net significantly outperforms existing traditional and deep learning-based methods in terms of unmixing accuracy. The code is available at https://github.com/Octopus-Squidward/MSSF-Net.
科研通智能强力驱动
Strongly Powered by AbleSci AI