计算机科学                        
                
                                
                        
                            卷积神经网络                        
                
                                
                        
                            安全性令牌                        
                
                                
                        
                            特征学习                        
                
                                
                        
                            归纳偏置                        
                
                                
                        
                            人工智能                        
                
                                
                        
                            嵌入                        
                
                                
                        
                            桥接(联网)                        
                
                                
                        
                            模式识别(心理学)                        
                
                                
                        
                            频道(广播)                        
                
                                
                        
                            代表(政治)                        
                
                                
                        
                            机器学习                        
                
                                
                        
                            多任务学习                        
                
                                
                        
                            任务(项目管理)                        
                
                                
                        
                            工程类                        
                
                                
                        
                            政治                        
                
                                
                        
                            计算机安全                        
                
                                
                        
                            计算机网络                        
                
                                
                        
                            法学                        
                
                                
                        
                            系统工程                        
                
                                
                        
                            政治学                        
                
                        
                    
            作者
            
                Zhiying Lu,Hongtao Xie,Chuanbin Liu,Yongdong Zhang            
         
                    
            出处
            
                                    期刊:Cornell University - arXiv
                                                                        日期:2022-01-01
                                                                        被引量:24
                                
         
        
    
            
            标识
            
                                    DOI:10.48550/arxiv.2210.05958
                                    
                                
                                 
         
        
                
            摘要
            
            There still remains an extreme performance gap between Vision Transformers (ViTs) and Convolutional Neural Networks (CNNs) when training from scratch on small datasets, which is concluded to the lack of inductive bias. In this paper, we further consider this problem and point out two weaknesses of ViTs in inductive biases, that is, the spatial relevance and diverse channel representation. First, on spatial aspect, objects are locally compact and relevant, thus fine-grained feature needs to be extracted from a token and its neighbors. While the lack of data hinders ViTs to attend the spatial relevance. Second, on channel aspect, representation exhibits diversity on different channels. But the scarce data can not enable ViTs to learn strong enough representation for accurate recognition. To this end, we propose Dynamic Hybrid Vision Transformer (DHVT) as the solution to enhance the two inductive biases. On spatial aspect, we adopt a hybrid structure, in which convolution is integrated into patch embedding and multi-layer perceptron module, forcing the model to capture the token features as well as their neighboring features. On channel aspect, we introduce a dynamic feature aggregation module in MLP and a brand new "head token" design in multi-head self-attention module to help re-calibrate channel representation and make different channel group representation interacts with each other. The fusion of weak channel representation forms a strong enough representation for classification. With this design, we successfully eliminate the performance gap between CNNs and ViTs, and our DHVT achieves a series of state-of-the-art performance with a lightweight model, 85.68% on CIFAR-100 with 22.8M parameters, 82.3% on ImageNet-1K with 24.0M parameters. Code is available at https://github.com/ArieSeirack/DHVT.
         
            
 
                 
                
                    
                    科研通智能强力驱动
Strongly Powered by AbleSci AI