概括性
缩放比例
计算机科学
刮擦
学习迁移
幂律
标度律
熵(时间箭头)
人工智能
数学
统计
物理
几何学
操作系统
心理治疗师
量子力学
心理学
作者
Danny Hernandez,Jared Kaplan,Tom Henighan,Sam McCandlish
出处
期刊:Cornell University - arXiv
日期:2021-02-02
被引量:27
摘要
We study empirical scaling laws for transfer learning between distributions in an unsupervised, fine-tuning setting. When we train increasingly large neural networks from-scratch on a fixed-size dataset, they eventually become data-limited and stop improving in performance (cross-entropy loss). When we do the same for models pre-trained on a large language dataset, the slope in performance gains is merely reduced rather than going to zero. We calculate the effective data from pre-training by determining how much data a transformer of the same size would have required to achieve the same loss when training from scratch. In other words, we focus on units of data while holding everything else fixed. We find that the effective data transferred is described well in the low data regime by a power-law of parameter count and fine-tuning dataset size. We believe the exponents in these power-laws correspond to measures of the generality of a model and proximity of distributions (in a directed rather than symmetric sense). We find that pre-training effectively multiplies the fine-tuning dataset size. Transfer, like overall performance, scales predictably in terms of parameters, data, and compute.
科研通智能强力驱动
Strongly Powered by AbleSci AI