激活函数
计算机科学
分段线性函数
人工智能
人工神经网络
分段
树遍历
推论
灵活性(工程)
树(集合论)
功能(生物学)
算法
模式识别(心理学)
数学
数学分析
统计
几何学
进化生物学
生物
作者
Zezhou Zhu,Yucong Zhou,Yuan Dong,Zhao Zhong
标识
DOI:10.1109/tpami.2023.3286109
摘要
The choice of activation functions is crucial to deep neural networks. ReLU is a popular hand-designed activation function. Swish, the automatically searched activation function, outperforms ReLU on many challenging datasets. However, the search method has two main drawbacks. First, the tree-based search space is highly discrete and restricted, which is difficult to search. Second, the sample-based search method is inefficient in finding specialized activation functions for each dataset or neural architecture. To overcome these drawbacks, we propose a new activation function called Piecewise Linear Unit (PWLU), incorporating a carefully designed formulation and learning method. PWLU can learn specialized activation functions for different models, layers, or channels. Besides, we propose a non-uniform version of PWLU, which maintains sufficient flexibility but requires fewer intervals and parameters. Additionally, we generalize PWLU to three-dimensional space to define a piecewise linear surface named 2D-PWLU, which can be treated as a non-linear binary operator. Experimental results show that PWLU achieves SOTA performance on various tasks and models, and 2D-PWLU is better than element-wise addition when aggregating features from different branches. The proposed PWLU and its variation are easy to implement and efficient for inference, which can be widely applied in real-world applications.
科研通智能强力驱动
Strongly Powered by AbleSci AI