计算机科学
规范化(社会学)
建筑
可微函数
人工神经网络
对偶(语法数字)
人工智能
网络体系结构
计算机工程
数学
计算机网络
社会学
视觉艺术
艺术
数学分析
文学类
人类学
作者
Yu Xue,Xiaolong Han,Zehong Wang
标识
DOI:10.1109/tii.2023.3348843
摘要
Differentiable architecture search is a popular gradient-based method for neural architecture search, and has achieved great success in automating design of neural network architectures. However, it still has some limitations such as performance collapse, which seriously affects network architecture performance. To solve this issue, we propose an algorithm called self-adaptive weight based on dual-attention for differentiable neural architecture search (SWD-NAS) in this article. SWD-NAS utilizes a dual-attention mechanism to measure architectural weights. Specifically, an upper-attention module is used to adaptively select channels based on their weights before inputting into the search space. A lower-attention (LA) module is utilized to calculate architectural weights. In addition, we propose an architectural weight normalization to alleviate the unfair competition among connection edges. Finally, we evaluate the architectures searched on CIFAR-10 and CIFAR-100, achieving test errors of 2.51% and 16.13%, respectively. Furthermore, we transfer the architecture searched on CIFAR-10 to ImageNet, achieving top-one and top-five errors of 24.5% and 7.6%, respectively. This demonstrates the superior performance of the proposed algorithm compared to many gradient-based algorithms.
科研通智能强力驱动
Strongly Powered by AbleSci AI