超参数
人工神经网络
计算机科学
强迫(数学)
机器学习
过程(计算)
期限(时间)
任务(项目管理)
人工智能
时间序列
系列(地层学)
数学
数学分析
古生物学
物理
管理
量子力学
经济
生物
操作系统
作者
Roland Bolboacă,Piroska Haller
出处
期刊:Mathematics
[MDPI AG]
日期:2023-03-15
卷期号:11 (6): 1432-1432
被引量:7
摘要
Long short-term memory neural networks have been proposed as a means of creating accurate models from large time series data originating from various fields. These models can further be utilized for prediction, control, or anomaly-detection algorithms. However, finding the optimal hyperparameters to maximize different performance criteria remains a challenge for both novice and experienced users. Hyperparameter optimization algorithms can often be a resource-intensive and time-consuming task, particularly when the impact of the hyperparameters on the performance of the neural network is not comprehended or known. Teacher forcing denotes a procedure that involves feeding the ground truth output from the previous time-step as input to the current time-step during training, while during testing feeding back the predicted values. This paper presents a comprehensive examination of the impact of hyperparameters on long short-term neural networks, with and without teacher forcing, on prediction performance. The study includes testing long short-term memory neural networks, with two variations of teacher forcing, in two prediction modes, using two configurations (i.e., multi-input single-output and multi-input multi-output) on a well-known chemical process simulation dataset. Furthermore, this paper demonstrates the applicability of a long short-term memory neural network with a modified teacher forcing approach in a process state monitoring system. Over 100,000 experiments were conducted with varying hyperparameters and in multiple neural network operation modes, revealing the direct impact of each tested hyperparameter on the training and testing procedures.
科研通智能强力驱动
Strongly Powered by AbleSci AI