横杆开关
记忆电阻器
人工神经网络
计算机科学
培训(气象学)
蒸馏
容错
电子工程
电气工程
人工智能
工程类
电信
分布式计算
物理
气象学
有机化学
化学
作者
Mei Guo,Xingwei Zhang,Gang Dou,Herbert Ho‐Ching Iu
出处
期刊:IEEE Transactions on Circuits and Systems I-regular Papers
[Institute of Electrical and Electronics Engineers]
日期:2025-04-24
卷期号:72 (11): 7120-7131
被引量:1
标识
DOI:10.1109/tcsi.2025.3561803
摘要
Knowledge distillation is widely used as an effective model compression technique to improve the performance of small models. Most of the current researches on knowledge distillation focus on the algorithmic level and ignore the potential benefits of hardware implementation. In this paper, a multi-loss knowledge distillation online training circuit based on memristor crossbar array is designed, which can improve the inference efficiency and reduce the power consumption of deep learning models on edge devices. The circuit is able to process data in real time, and it can be used to handle stuck-at-faults (SAF) caused by factors such as manufacturing defects in the memristor. Moreover, a fault detection scheme with low time cost is proposed in order to address the low efficiency of stuck-at-fault detection in memristor crossbar arrays. The scheme is combined with a self-compensating pruning method and knowledge distillation online training mechanism, which significantly improves the model training and inference capability of the circuit under fault conditions. Experimental results show that the multi-loss knowledge distillation online training improves the accuracy by 4.15% and 63.48% respectively in two models compared with traditional training schemes. The fault-tolerance scheme reduces the power consumption of the memristor crossbar arrays by 41.2% and 72.6% respectively on the two models, demonstrating its potential and advantages in edge computing.
科研通智能强力驱动
Strongly Powered by AbleSci AI