循环神经网络
计算机科学
正规化(语言学)
秩(图论)
软件部署
约束(计算机辅助设计)
人工智能
人工神经网络
投影(关系代数)
自适应路由
布线(电子设计自动化)
机器学习
算法
数学
路由协议
计算机网络
静态路由
几何学
操作系统
组合数学
作者
Dongjing Shan,Yong Luo,Xiongwei Zhang,Chao Zhang
标识
DOI:10.1109/tnnls.2021.3105818
摘要
Recurrent neural networks (RNNs) continue to show outstanding performance in sequence learning tasks such as language modeling, but it remains difficult to train RNNs for long sequences. The main challenges lie in the complex dependencies, gradient vanishing or exploding, and low resource requirement in model deployment. In order to address these challenges, we propose dynamic recurrent routing neural networks (DRRNets), which can: 1) shorten the recurrent lengths by allocating recurrent routes dynamically for different dependencies and 2) reduce the number of parameters significantly by imposing low-rank constraints on the fully connected layers. A novel optimization algorithm via low-rank constraint and sparsity projection is developed to train the network. We verify the effectiveness of the proposed method by comparing it with multiple competitive approaches in several popular sequential learning tasks, such as language modeling and speaker recognition. The results in terms of different criteria demonstrate the superiority of our proposed method.
科研通智能强力驱动
Strongly Powered by AbleSci AI