计算机科学
判别式
成对比较
人工智能
特征学习
杠杆(统计)
代表(政治)
水准点(测量)
发电机(电路理论)
机器学习
模式识别(心理学)
对比度(视觉)
特征(语言学)
功率(物理)
哲学
语言学
物理
大地测量学
量子力学
政治
政治学
法学
地理
作者
Fangrui Liu,Zihao Liu,Zheng Liu
标识
DOI:10.1007/978-3-030-88004-0_8
摘要
Fine-grained visual classification is challenging due to subtle differences between sub-categories. Current popular methods usually leverage a single image and are designed by two main perspectives: feature representation learning and discriminative parts localization, while a few methods utilize pairwise images as input. However, it is difficult to learn representations discriminatively both across the images and across the categories, as well as to guarantee for accurate location of discriminative parts. In this paper, different from the existing methods, we argue to solve these difficulties from the perspective of contrastive learning and propose a novel Attentive Contrast Learning Network (ACLN). The network aims to attract the representation of positive pairs, which are from the same category, and repulse the representation of negative pairs, which are from different categories. A contrastive learning module, equipped with two contrastive losses, is proposed to achieve this. Specifically, the attention maps, generated by the attention generator, are bounded with the original CNN feature as positive pair, while the attention maps of different images form the negative pairs. Besides, the final classification results are obtained by a synergic learning module, utilizing both the original feature and the attention maps. Comprehensive experiments are conducted on four benchmark datasets, on which our ACLN outperforms all the existing SOTA approaches. For reproducible scientific research https://github.com/mpskex/AttentiveContrastiveLearningNetwork.
科研通智能强力驱动
Strongly Powered by AbleSci AI