计算机科学
班级(哲学)
人工智能
模式识别(心理学)
任务(项目管理)
特征(语言学)
跳跃式监视
注意力网络
代表(政治)
人工神经网络
空间分析
图像(数学)
机器学习
数据挖掘
数学
经济
管理
法学
哲学
统计
政治
语言学
政治学
作者
Adu Asare Baffour,Zhen Qin,Yong J. Wang,Zhiguang Qin,Kim‐Kwang Raymond Choo
标识
DOI:10.1016/j.jvcir.2021.103368
摘要
The underlining task for fine-grained image recognition captures both the inter-class and intra-class discriminate features. Existing methods generally use auxiliary data to guide the network or a complex network comprising multiple sub-networks. They have two significant drawbacks: (1) Using auxiliary data like bounding boxes requires expert knowledge and expensive data annotation. (2) Using multiple sub-networks make network architecture complex and requires complicated training or multiple training steps. We propose an end-to-end Spatial Self-Attention Network (SSANet) comprising a spatial self-attention module (SSA) and a self-attention distillation (Self-AD) technique. The SSA encodes contextual information into local features, improving intra-class representation. Then, the Self-AD distills knowledge from the SSA to a primary feature map, obtaining inter-class representation. By accumulating classification losses from these two modules enables the network to learn both inter-class and intra-class features in one training step. The experiment findings demonstrate that SSANet is effective and achieves competitive performance.
科研通智能强力驱动
Strongly Powered by AbleSci AI