计算机科学
水准点(测量)
人工智能
自然语言处理
语言模型
领域(数学分析)
任务(项目管理)
生物医学
又称作
语言理解
命名实体识别
多样性(控制论)
生物信息学
数学分析
图书馆学
经济
生物
管理
地理
数学
大地测量学
作者
裕二 池谷,Robert Tinn,Hao Cheng,Michael Lucas,Naoto Usuyama,Xiaodong Liu,Tristan Naumann,Jianfeng Gao,Hoifung Poon
出处
期刊:ACM transactions on computing for healthcare
[Association for Computing Machinery]
日期:2021-10-15
卷期号:3 (1): 1-23
被引量:1195
摘要
Pretraining large neural language models, such as BERT, has led to impressive gains on many natural language processing (NLP) tasks. However, most pretraining efforts focus on general domain corpora, such as newswire and Web. A prevailing assumption is that even domain-specific pretraining can benefit by starting from general-domain language models. In this article, we challenge this assumption by showing that for domains with abundant unlabeled text, such as biomedicine, pretraining language models from scratch results in substantial gains over continual pretraining of general-domain language models. To facilitate this investigation, we compile a comprehensive biomedical NLP benchmark from publicly available datasets. Our experiments show that domain-specific pretraining serves as a solid foundation for a wide range of biomedical NLP tasks, leading to new state-of-the-art results across the board. Further, in conducting a thorough evaluation of modeling choices, both for pretraining and task-specific fine-tuning, we discover that some common practices are unnecessary with BERT models, such as using complex tagging schemes in named entity recognition. To help accelerate research in biomedical NLP, we have released our state-of-the-art pretrained and task-specific models for the community, and created a leaderboard featuring our BLURB benchmark (short for Biomedical Language Understanding & Reasoning Benchmark) at https://aka.ms/BLURB .
科研通智能强力驱动
Strongly Powered by AbleSci AI