文字2vec
计算机科学
人工智能
情绪分析
文字嵌入
学习迁移
深度学习
自然语言处理
背景(考古学)
孟加拉语
机器学习
嵌入
领域(数学分析)
古生物学
数学分析
数学
生物
作者
Nusrat Jahan Prottasha,Abdullah As Sami,Md. Kowsher,Saydul Akbar Murad,Anupam Kumar Bairagi,Mehedi Masud,Mohammed Baz
出处
期刊:Sensors
[Multidisciplinary Digital Publishing Institute]
日期:2022-05-30
卷期号:22 (11): 4157-4157
被引量:133
摘要
The growth of the Internet has expanded the amount of data expressed by users across multiple platforms. The availability of these different worldviews and individuals' emotions empowers sentiment analysis. However, sentiment analysis becomes even more challenging due to a scarcity of standardized labeled data in the Bangla NLP domain. The majority of the existing Bangla research has relied on models of deep learning that significantly focus on context-independent word embeddings, such as Word2Vec, GloVe, and fastText, in which each word has a fixed representation irrespective of its context. Meanwhile, context-based pre-trained language models such as BERT have recently revolutionized the state of natural language processing. In this work, we utilized BERT's transfer learning ability to a deep integrated model CNN-BiLSTM for enhanced performance of decision-making in sentiment analysis. In addition, we also introduced the ability of transfer learning to classical machine learning algorithms for the performance comparison of CNN-BiLSTM. Additionally, we explore various word embedding techniques, such as Word2Vec, GloVe, and fastText, and compare their performance to the BERT transfer learning strategy. As a result, we have shown a state-of-the-art binary classification performance for Bangla sentiment analysis that significantly outperforms all embedding and algorithms.
科研通智能强力驱动
Strongly Powered by AbleSci AI