计算机科学
任务(项目管理)
人工智能
自然语言处理
编码(集合论)
语言模型
弹丸
自然语言
可微函数
自然语言理解
程序设计语言
数学分析
经济
集合(抽象数据类型)
有机化学
化学
管理
数学
作者
Ningyu Zhang,Luoqiu Li,Xiang Chen,Shumin Deng,Zhen Bi,Chuanqi Tan,Fei Huang,Huajun Chen
出处
期刊:Cornell University - arXiv
日期:2021-08-30
被引量:2
标识
DOI:10.48550/arxiv.2108.13161
摘要
Large-scale pre-trained language models have contributed significantly to natural language processing by demonstrating remarkable abilities as few-shot learners. However, their effectiveness depends mainly on scaling the model parameters and prompt design, hindering their implementation in most real-world applications. This study proposes a novel pluggable, extensible, and efficient approach named DifferentiAble pRompT (DART), which can convert small language models into better few-shot learners without any prompt engineering. The main principle behind this approach involves reformulating potential natural language processing tasks into the task of a pre-trained language model and differentially optimizing the prompt template as well as the target label with backpropagation. Furthermore, the proposed approach can be: (i) Plugged to any pre-trained language models; (ii) Extended to widespread classification tasks. A comprehensive evaluation of standard NLP tasks demonstrates that the proposed approach achieves a better few-shot performance. Code is available in https://github.com/zjunlp/DART.
科研通智能强力驱动
Strongly Powered by AbleSci AI