过度拟合
计算机科学
人工智能
词(群论)
卷积神经网络
自然语言处理
特征(语言学)
特征提取
特征向量
人工神经网络
模式识别(心理学)
机器学习
语言学
哲学
作者
Youyao Liu,Haimei Huang,Jialei Gao,Shihao Gai
标识
DOI:10.1109/icnlp58431.2023.00062
摘要
Chinese Text Classification (TC) is the process of mapping text to a pre-given topics category. In recent years, it has been found that TC is mainly based on RNN and BERT, so the development of different novel pre-training applied to Chinese TC is described as based on BERT pre-training. BERT combined with convolutional neural network is proposed to extend the BERT-CNN model for the problem of lack of semantic knowledge of BERT to derive a good classification effect. The second RoBERTa model performs feature extraction and fusion to obtain the feature word vector as the text output vector, which can solve the problem of insufficient BERT extracted features. The BERT-BiGRU model, on the other hand, mainly avoids the increase in the number of texts leading to long training time and overfitting, and uses a simpler GRU bi-word network as the main network to fully extract the contextual information of Chinese texts and finally complete the classification of Chinese texts; at the same time, it makes an outlook and conclusion on the new pre-training model for Chinese TC.
科研通智能强力驱动
Strongly Powered by AbleSci AI