计算机科学
多样性(控制论)
集合(抽象数据类型)
符号
封面(代数)
人工智能
自然语言
领域(数学)
功能(生物学)
语言模型
弦(物理)
自然语言处理
自然语言理解
程序设计语言
语言学
哲学
机械工程
物理
数学
量子力学
进化生物学
纯数学
工程类
生物
作者
Pengfei Liu,Weizhe Yuan,Jinlan Fu,Zhengbao Jiang,Hiroaki Hayashi,Graham Neubig
出处
期刊:ACM Computing Surveys
[Association for Computing Machinery]
日期:2022-09-14
卷期号:55 (9): 1-35
被引量:2103
摘要
This article surveys and organizes research works in a new paradigm in natural language processing, which we dub “prompt-based learning.” Unlike traditional supervised learning, which trains a model to take in an input x and predict an output y as P ( y|x ), prompt-based learning is based on language models that model the probability of text directly. To use these models to perform prediction tasks, the original input x is modified using a template into a textual string prompt x′ that has some unfilled slots, and then the language model is used to probabilistically fill the unfilled information to obtain a final string x̂ , from which the final output y can be derived. This framework is powerful and attractive for a number of reasons: It allows the language model to be pre-trained on massive amounts of raw text, and by defining a new prompting function the model is able to perform few-shot or even zero-shot learning, adapting to new scenarios with few or no labeled data. In this article, we introduce the basics of this promising paradigm, describe a unified set of mathematical notations that can cover a wide variety of existing work, and organize existing work along several dimensions, e.g., the choice of pre-trained language models, prompts, and tuning strategies. To make the field more accessible to interested beginners, we not only make a systematic review of existing works and a highly structured typology of prompt-based concepts but also release other resources, e.g., a website NLPedia–Pretrain including constantly updated survey and paperlist.
科研通智能强力驱动
Strongly Powered by AbleSci AI