可解释性
可微函数
计算机科学
人工神经网络
块(置换群论)
模块化设计
班级(哲学)
前馈神经网络
数学优化
理论计算机科学
人工智能
数学
纯数学
几何学
操作系统
作者
Jian Wang,S.M. Wu,Huaqing Zhang,Bin Yuan,Caili Dai,Nikhil R. Pal
标识
DOI:10.1109/tnnls.2024.3378697
摘要
Approximation ability is one of the most important topics in the field of neural networks (NNs). Feedforward NNs, activated by rectified linear units and some of their specific smoothed versions, provide universal approximators to convex as well as continuous functions. However, most of these networks are investigated empirically, or their characteristics are analyzed based on specific operation rules. Moreover, an adequate level of interpretability of the networks is missing as well. In this work, we propose a class of new network architecture, built with reusable neural modules (functional blocks), to supply differentiable and interpretable approximators for convex and continuous target functions. Specifically, first, we introduce a concrete model construction mechanism with particular blocks based on differentiable programming and the composition essence of the max operator, extending the scope of existing activation functions. Moreover, explicit block diagrams are provided for a clear understanding of the external architecture and the internal processing mechanism. Subsequently, the approximation behavior of the proposed network to convex functions and continuous functions is rigorously proved as well, by virtue of mathematical induction. Finally, plenty of numerical experiments are conducted on a wide variety of problems, which exhibit the effectiveness and the superiority of the proposed model over some existing ones.
科研通智能强力驱动
Strongly Powered by AbleSci AI