利用
计算机科学
特征(语言学)
人工智能
卷积神经网络
背景(考古学)
关系(数据库)
特征学习
代表(政治)
机器学习
模式识别(心理学)
数据挖掘
古生物学
哲学
语言学
计算机安全
政治
政治学
法学
生物
作者
Zhizheng Zhang,Cuiling Lan,Wenjun Zeng,Xin Jin,Zhibo Chen
摘要
Attention mechanism aims to increase the representation power by focusing on important features and suppressing unnecessary ones. For convolutional neural networks (CNNs), attention is typically learned with local convolutions, which ignores the global information and the hidden relation. How to efficiently exploit the long-range context to globally learn attention is underexplored. In this paper, we propose an effective Relation-Aware Global Attention (RGA) module for CNNs to fully exploit the global correlations to infer the attention. Specifically, when computing the attention at a feature position, in order to grasp information of global scope, we propose to stack the relations, i.e., its pairwise correlations/affinities with all the feature positions, and the feature itself together for learning the attention with convolutional operations. Given an intermediate feature map, we have validated the effectiveness of this design across both the spatial and channel dimensions. When applied to the task of person re-identification, our model achieves the state-of-the-art performance. Extensive ablation studies demonstrate that our RGA can significantly enhance the feature representation power. We further demonstrate the general applicability of RGA to vision tasks by applying it to the scene segmentation and image classification tasks resulting in consistent performance improvement.
科研通智能强力驱动
Strongly Powered by AbleSci AI