遥感
探测器
计算机科学
计算机视觉
目标检测
基于对象
人工智能
对象(语法)
模式识别(心理学)
地理
电信
作者
Zheng Li,Yongcheng Wang,Hao Feng,Chi Chen,XU Dongdong,Tianqi Zhao,Yunxiao Gao,Zhikang Zhao
标识
DOI:10.1109/tgrs.2025.3525720
摘要
Object detection plays a crucial role in remote sensing due to the urgent demands of various applications, such as urban planning and environmental monitoring. Despite notable progress, current methods still struggle with detecting challenging small objects. At the object level, the limited pixel representation, blurred details, and background interference of small objects impose greater demands on feature extractors. At the network level, resource bias fails to provide adequate learning signals for these objects. In this paper, we propose a Sparse Transformer-based detector (STDet) to tackle these challenges. Specifically, we design a Local-to-Global Transformer network (LGFormer) to explore essential feature representations. The Local Transformer Block establishes correlations between tokens and their surrounding data, while the Global Transformer Block captures long-distance dependencies related to the objects. Meanwhile, we introduce a Scale-Balanced Label Assignment (SBLA) strategy that considers more samples to small objects. SBLA dynamically shifts the learning focus to easily overlooked objects and alleviates the issue of sample imbalance. Extensive experiments on three large-scale remote sensing datasets demonstrate the effectiveness of STDet and its superiority in small object detection.
科研通智能强力驱动
Strongly Powered by AbleSci AI