量化(信号处理)
内存占用
计算机科学
乘数(经济学)
算法
人工智能
程序设计语言
经济
宏观经济学
作者
Ruokai Yin,Yuhang Li,Abhishek Moitra,Priyadarshini Panda
标识
DOI:10.1109/asp-dac58780.2024.10473825
摘要
We propose Multiplier-less INTeger (MINT) quantization, a uniform quantization scheme that efficiently compresses weights and membrane potentials in spiking neural networks (SNNs). Unlike previous SNN quantization methods, MINT quantizes memory-intensive membrane potentials to an extremely low precision (2-bit), significantly reducing the memory footprint. MINT also shares the quantization scaling factor between weights and membrane potentials, eliminating the need for multipliers required in conventional uniform quantization. Experimental results show that our method matches the accuracy of full-precision models and other state-of-the-art SNN quantization techniques while surpassing them in memory footprint reduction and hardware cost efficiency at deployment. For example, 2-bit MINT VGG-16 achieves 90.6% accuracy on CIFAR-10, with roughly 93.8% reduction in memory footprint from the full-precision model and 90% reduction in computation energy compared to vanilla uniform quantization at deployment. 1 1 Code is available at https://github.com/Intelligent-Computing-Lab-Yale/MINT-Quantization
科研通智能强力驱动
Strongly Powered by AbleSci AI