计算机科学
瓶颈
失败
卷积神经网络
块(置换群论)
推论
卷积(计算机科学)
移动设备
编码(集合论)
计算机工程
树(集合论)
航程(航空)
加速
人工智能
计算机体系结构
并行计算
嵌入式系统
人工神经网络
操作系统
程序设计语言
数学
复合材料
集合(抽象数据类型)
材料科学
数学分析
几何学
作者
Yehui Tang,Kai Han,Jianyuan Guo,Chang Xu,Chao Xu,Yunhe Wang
出处
期刊:Cornell University - arXiv
日期:2022-01-01
被引量:211
标识
DOI:10.48550/arxiv.2211.12905
摘要
Light-weight convolutional neural networks (CNNs) are specially designed for applications on mobile devices with faster inference speed. The convolutional operation can only capture local information in a window region, which prevents performance from being further improved. Introducing self-attention into convolution can capture global information well, but it will largely encumber the actual speed. In this paper, we propose a hardware-friendly attention mechanism (dubbed DFC attention) and then present a new GhostNetV2 architecture for mobile applications. The proposed DFC attention is constructed based on fully-connected layers, which can not only execute fast on common hardware but also capture the dependence between long-range pixels. We further revisit the expressiveness bottleneck in previous GhostNet and propose to enhance expanded features produced by cheap operations with DFC attention, so that a GhostNetV2 block can aggregate local and long-range information simultaneously. Extensive experiments demonstrate the superiority of GhostNetV2 over existing architectures. For example, it achieves 75.3% top-1 accuracy on ImageNet with 167M FLOPs, significantly suppressing GhostNetV1 (74.5%) with a similar computational cost. The source code will be available at https://github.com/huawei-noah/Efficient-AI-Backbones/tree/master/ghostnetv2_pytorch and https://gitee.com/mindspore/models/tree/master/research/cv/ghostnetv2.
科研通智能强力驱动
Strongly Powered by AbleSci AI