计算机科学
光学(聚焦)
分割
人工智能
特征提取
遥感
图像融合
特征(语言学)
卷积神经网络
小波变换
交叉口(航空)
模式识别(心理学)
代表(政治)
图像分割
哈尔小波转换
计算机视觉
小波
目标检测
空间分析
遥感应用
传感器融合
融合
语义学(计算机科学)
深度学习
特征学习
像素
能见度
作者
Chenxu Ge,Qiangkui Leng,Ting Zhang,Samiullah,Abdallah Namoun,Syed Mudassir Hussain,Hisham Alfuhaid,Muhammad Waqas
标识
DOI:10.1109/jstars.2026.3658488
摘要
Precise semantic segmentation of High-Resolution Remote Sensing(HRRS) images is essential for robust environmental surveillance and detailed land use mapping. Despite substantial advances in deep learning, most conventional approaches focus on the spatial domain. This focus often neglects the rich textural and structural nuances found in the frequency domain, which reduces the representation of comprehensive data. Addressing this issue, we introduce SF-Net. This network synthesizes features across spatial and frequency domains, aiming for seamless and effective integration. The core of SF-Net employs a multiscale Convolutional Grouping Fusion Module (CGFM) to extract spatial features at varying resolutions. Following this, the Haar Wavelet Transform decomposes these features into distinct low-frequency components (structure) and high-frequency components (detail). Subsequently, a Mamba-enhanced Global Spatial Feature Extraction Module (GSFEM) reinforces low-frequency semantic information with global context, while a Spatial-Frequency Fusion Module (S-FFM) applies targeted attention to sharpen high-frequency details. Experimental results on the ISPRS Vaihingen, LoveDA, and Potsdam benchmarks confirm SF-Net's superior performance, achieving state-of-the-art mean Intersection over Union (mIoU) scores of 83.12%, 53.28%, and 83.35%, respectively, validating its effectiveness and superority.
科研通智能强力驱动
Strongly Powered by AbleSci AI