判别式
计算机科学
水准点(测量)
人工智能
词(群论)
特征(语言学)
背景(考古学)
任务(项目管理)
自然语言处理
航程(航空)
模式识别(心理学)
机器学习
数学
古生物学
材料科学
复合材料
语言学
大地测量学
地理
哲学
几何学
经济
管理
生物
作者
Qianli Ma,Jiangyue Yan,Zhenxi Lin,Liuhong Yu,Zipeng Chen
标识
DOI:10.1109/taslp.2021.3067210
摘要
Text classification is an important task in natural language processing. Contextual information is essential for text classification, and different words usually need different sizes of contextual information. However, most existing methods learn contextual features with predefined fixed sizes, which cannot extract the different sizes of contextual features for different words. To this end, we propose a new model named Deformable Self-Attention (DSA) to flexibly learn word-specific contextual features, rather than extracting features of fixed context sizes. Our model is mainly composed of a Deformable Local Attention Weight Generation (DLAWG) module and a Multi-Range Feature Integration (MRFI) module. The DLAWG module can adaptively determine different context sizes for different words within a particular range and then learn word-specific contextual features for each word. DLAWG then employs multiple ranges to capture context dependencies of different ranges. After that, the MRFI module integrates features from different ranges by considering the interactions with features of different ranges, which can delete irrelevant features while enhancing discriminative ones. Experiments on extensive benchmark datasets and visualizations illustrate the effectiveness of our model.
科研通智能强力驱动
Strongly Powered by AbleSci AI