计算机科学
命名实体识别
变压器
超参数
预处理器
人工智能
自然语言处理
语言模型
答疑
机器学习
电压
物理
管理
量子力学
经济
任务(项目管理)
作者
Prasanna Kumar R,Bharathi Mohan G,Parthasarathy Srinivasan,R. Venkatakrishnan
标识
DOI:10.1109/icccnt56998.2023.10308039
摘要
NER is a critical problem in natural language processing (NLP) that includes detecting and classifying named entities in text. We give a complete investigation on NER using the CoNLL dataset in this research article. We explore the evolution of NER techniques, from traditional rule-based methods to state-of-the-art transformer-based models. Our experiments focus on evaluating the performance of three popular models: BERT, ALBERT, and XLM-RoBERTa, on the CoNLL dataset. We analyze their accuracies and discuss their strengths and limitations. The results demonstrate the effectiveness of transformer-based models, with BERT achieving an accuracy of 99.50%, ALBERT achieving 96.90% accuracy, and XLM-RoBERTa achieving 87.82% accuracy. We discuss the importance of dataset preprocessing, model fine-tuning, and hyperparameter optimization for achieving optimal performance. This research contributes to the understanding of NER techniques and provides insights into the performance of transformer-based models for NER tasks.
科研通智能强力驱动
Strongly Powered by AbleSci AI