卷积神经网络
计算机科学
核(代数)
硬件加速
控制重构
特征(语言学)
计算机硬件
图层(电子)
加速度
深度学习
并行计算
人工智能
现场可编程门阵列
嵌入式系统
数学
物理
材料科学
纳米技术
经典力学
组合数学
哲学
语言学
作者
Rastislav Struharik,Bogdan Vukobratović,Andrea Erdeljan,Damjan Rakanovic
标识
DOI:10.1109/dsd.2018.00070
摘要
In this paper we propose a novel Convolutional Neural Network hardware accelerator, called CoNNA, capable of accelerating pruned, quantized, CNNs. In contrast to most existing solutions, CoNNA offers a complete solution to the full, compressed CNN acceleration, being able to accelerate all layer types commonly found in contemporary CNNs. CoNNA is designed as a coarse-grained reconfigurable architecture, which uses rapid, dynamic reconfiguration during CNN layer processing. Furthermore, by being able to directly process compressed feature and kernel maps, CoNNA is able to achieve higher CNN processing efficiency than some of the previously proposed solutions. Results of the experiments indicate that CoNNA architecture is up to 14.10 times faster than previously proposed MIT's Eyeriss CNN accelerator, up to 6.05 times faster than NullHop CNN accelerator, and up to 4.91 times faster than NVIDIA's Deep Learning Accelerator (NVDLA), while using identical number of computing units and operating at the same clock frequency.
科研通智能强力驱动
Strongly Powered by AbleSci AI