计算机科学
高效能源利用
最上等的
计算机硬件
吞吐量
嵌入式系统
计算机体系结构
计算机工程
电气工程
天文
电信
方位角
无线
物理
工程类
作者
Taha Soliman,Nellie Laleni,Tobias Kirchner,Franz Mȕller,Ashish Shrivastava,Thomas Kämpfe,Andre Guntoro,Norbert Wehn
出处
期刊:ACM Transactions in Embedded Computing Systems
[Association for Computing Machinery]
日期:2022-04-09
卷期号:21 (6): 1-25
被引量:30
摘要
Today, a large number of applications depend on deep neural networks (DNN) to process data and perform complicated tasks at restricted power and latency specifications. Therefore, processing-in-memory (PIM) platforms are actively explored as a promising approach to improve the throughput and the energy efficiency of DNN computing systems. Several PIM architectures adopt resistive non-volatile memories as their main unit to build crossbar-based accelerators for DNN inference. However, these structures suffer from several drawbacks such as reliability, low accuracy, large ADCs/DACs power consumption and area, high write energy, and so on. In this article, we present a new mixed-signal in-memory architecture based on the bit-decomposition of the multiply and accumulate (MAC) operations. Our in-memory inference architecture uses a single FeFET as a non-volatile memory cell. Compared to the prior work, this system architecture provides a high level of parallelism while using only 3-bit ADCs. Also, it eliminates the need for any DAC. In addition, we provide flexibility and a very high utilization efficiency even for varying tasks and loads. Simulations demonstrate that we outperform state-of-the-art efficiencies with 36.5 TOPS/W and can pack 2.05 TOPS with 8-bit activation and 4-bit weight precision in an area of 4.9 mm 2 using 22 nm FDSOI technology. Employing binary operation, we obtain 1169 TOPS/W and over 261 TOPS/W/mm 2 on system level.
科研通智能强力驱动
Strongly Powered by AbleSci AI