计算机科学
多样性(控制论)
背景(考古学)
卷积神经网络
注意力网络
强迫(数学)
人工智能
学习迁移
编码(集合论)
人工神经网络
机器学习
地质学
生物
气候学
古生物学
集合(抽象数据类型)
程序设计语言
作者
Sergey Zagoruyko,Nikos Komodakis
出处
期刊:Cornell University - arXiv
日期:2016-12-12
被引量:1541
标识
DOI:10.48550/arxiv.1612.03928
摘要
Attention plays a critical role in human visual experience. Furthermore, it has recently been demonstrated that attention can also play an important role in the context of applying artificial neural networks to a variety of tasks from fields such as computer vision and NLP. In this work we show that, by properly defining attention for convolutional neural networks, we can actually use this type of information in order to significantly improve the performance of a student CNN network by forcing it to mimic the attention maps of a powerful teacher network. To that end, we propose several novel methods of transferring attention, showing consistent improvement across a variety of datasets and convolutional neural network architectures. Code and models for our experiments are available at https://github.com/szagoruyko/attention-transfer
科研通智能强力驱动
Strongly Powered by AbleSci AI