计算机科学
瓶颈
高效能源利用
硬件加速
吞吐量
嵌入式系统
宏
计算机硬件
计算机体系结构
操作系统
现场可编程门阵列
电气工程
无线
程序设计语言
工程类
作者
H. Ekin Sumbul,Jae-sun Seo,Daniel H. Morris,Édith Beigné
出处
期刊:IEEE Micro
[Institute of Electrical and Electronics Engineers]
日期:2023-12-05
卷期号:44 (2): 61-70
被引量:1
标识
DOI:10.1109/mm.2023.3338059
摘要
Compute-in-memory (CIM) has emerged as an effective technique to address memory access bottleneck for Deep Neural Networks (DNN). Augmented/Virtual Reality (AR/VR) devices require running high-performance DNN inference at tight power budgets, making CIMs ideal candidates for low-power on-device acceleration. While high energy-efficiencies have been reported at the CIM macro levels, system-on-chip (SoC) level energy-efficiency of CIM based accelerators have been underexplored for realistic system integration considerations. In this work, we present a CIM-accelerator architecture comprised of 16 row-pipelined fully-digital CIM macros, and provide a comprehensive analysis of CIM energy-efficiency benefits at the SoC-level targeting representative AR/VR workloads. Two key results are: (1) Realistic SoC-level CIM-accelerator energy-efficiency may be ∼50% lower than the CIM macro-level peak energy-efficiency when additional logic, memory hierarchies, and NN-dependent sub-optimal compute utilization are considered. (2) The CIM-accelerator still demonstrates up to ∼2.1× energy savings at the SoC-level compared to a systolic-array-based DNN accelerator at iso-peak throughput.
科研通智能强力驱动
Strongly Powered by AbleSci AI