电阻随机存取存储器
宏
GSM演进的增强数据速率
计算机科学
点(几何)
并行计算
电气工程
工程类
人工智能
数学
电压
几何学
程序设计语言
作者
Tai-Hao Wen,Hung-Hsi Hsu,Win-San Khwa,Wei-Hsing Huang,Zhao-En Ke,Yu-Hsiang Chin,Hua-Jin Wen,Yu‐Chen Chang,Wei‐Ting Hsu,Chung‐Chuan Lo,Ren-Shuo Liu,Chih-Cheng Hsieh,Kea‐Tiong Tang,Shih-Hsin Teng,Chung-Cheng Chou,Yu-Der Chih,Tsung-Yung Jonathan Chang,Meng‐Fan Chang
标识
DOI:10.1109/isscc49657.2024.10454468
摘要
AI-edge devices demand high-precision computation (e.g. FP16 and BF16) for accurate inference in practical applications, while maintaining high energy efficiency (EF) and low standby power to prolong battery life. Thus, advanced non-volatile AI-edge processors [1, 2] require non-volatile compute-in-memory (nvCIM) [3–5] with a large non-volatile on-chip memory, to store all of the neural network's parameters (weight data) during power-off, and high-precision high-EF multiply-and-accumulate (MAC) operations during compute, to maximize battery life. Among nvCIMs, ReRAM-nvCIM stands out as a promising candidate due to its lowest cost-per-bit (vs. MRAM, PCM, and eFlash), large on-off ratio, and resilience to magnetic-field interference. However, existing nvCIM macros [3–5] do not support floating-point (FP) computation. Implementing a FP-MAC for nvCIM faces challenges, as shown in Fig. 34.8.1, in (1) balancing the bit width tradeoff for weight pre-alignment between accuracy and storage, (2) addressing long latency and energy consumption in MAC operations due to the high input bit width in FP format, and (3) managing high array current consumption when accessing numerous memory cells (MCs) for FP operations, particularly in the low-resistance-state (LRS) ReRAM cells.
科研通智能强力驱动
Strongly Powered by AbleSci AI