判别式
计算机科学
人工智能
稳健性(进化)
观点
特征提取
模式识别(心理学)
分类器(UML)
特征学习
无人机
数据挖掘
计算机视觉
生物化学
生物
遗传学
基因
艺术
视觉艺术
化学
作者
Tianrui Shen,Yingmei Wei,Lai Kang,Shanshan Wan,Yee‐Hong Yang
标识
DOI:10.1109/tcsvt.2023.3296074
摘要
The key to crossview geolocalization is to match images of the same target from different viewpoints, e.g., images from drones and satellites. It is a challenging problem due to the changing appearance of objects from variable viewpoints. Most existing methods focus mainly on extracting global features or on segmenting feature maps, causing the loss of information contained in the images. To address the above issues, we propose a new ConvNeXt-based method called MCCG, which stands for Multiple Classifier for Cross-view Geolocalization. The proposed method captures rich discriminative information by cross-dimension interaction and acquires multiple feature representations, realizing a comprehensive feature representation. Additionally, the robustness of the model is improved crediting the multiple feature representations exploiting more contextual information despite position shifting or scale variations. Extensive experiments on the widely used public benchmarks University-1652 and SUES-200 demonstrate that the proposed method achieves state-of-the-art performance in both drone-view target localization and drone navigation applications by over 3% compared to existing methods. Our code and model are available at https://github.com/mode-str/crossview .
科研通智能强力驱动
Strongly Powered by AbleSci AI