MNIST数据库
计算机科学
卷积神经网络
硬件加速
量化(信号处理)
人工智能
人工神经网络
模式识别(心理学)
随机梯度下降算法
上下文图像分类
现场可编程门阵列
计算机硬件
算法
图像(数学)
标识
DOI:10.1109/cmsda58069.2022.00010
摘要
Convolutional Neural Network (CNN), a feedforward neural network, is widely used in large-scale machine learning tasks such as image recognition, image classification, and object detection. Although CNN has strong performance, its application in daily embedded systems is limited due to the complex network hierarchy and a large amount of data. In this paper, the LeNet-5 network structure is optimized. The recognition accuracy of the optimized network on the MNIST dataset is 98.32%, but the number of weights is reduced by more than ten times compared with the LeNet-5 network, and data quantization is used to convert the weights into 8-bit fixed-point numbers, which is more suitable for deployment in the hardware with limited resources and power consumption. At the same time, different hardware structures and optimization strategies are designed for each layer of the network, and the convolution, pooling and fully connected layers are computed in parallel and optimized using methods such as loop unrolling and pipelining. Finally, the feasibility of the hardware design is verified by simulating each layer using the MNIST dataset and quantized weights.
科研通智能强力驱动
Strongly Powered by AbleSci AI