过度拟合
非线性系统
计算机科学
人工神经网络
人工智能
非线性回归
感知器
回归
数学
数学优化
机器学习
回归分析
统计
物理
量子力学
作者
C. Beltran‐Royo,Laura Llopis-Ibor,Juan José Pantrigo,Iván Ramírez
标识
DOI:10.1016/j.knosys.2023.111154
摘要
In this paper, we analyze Difference of Convex Neural Networks in the context of one-dimensional nonlinear regression. Specifically, we show the surprising ability of the Difference of Convex Multilayer Perceptron (DC-MLP) to avoid overfitting in nonlinear regression. Otherwise said, DC-MLPs self-regularize (do not require additional regularization techniques). Thus, DC-MLPs could result very useful for practical purposes based on one-dimensional nonlinear regression. It turns out that shallow MLPs with a convex activation (ReLU, softplus, etc.) fall in the class of DC-MLPs. On the other hand, we call SQ-MLP the shallow MLP with a Squashing activation (logistic, hyperbolic tangent, etc.). In the numerical experiments, we show that DC-MLPs used for nonlinear regression avoid overfitting, in contrast with SQ-MLPs. We also compare DC-MLPs and SQ-MLPs from a theoretical point of view.
科研通智能强力驱动
Strongly Powered by AbleSci AI