判别式
序列(生物学)
发电机(电路理论)
计算机科学
鉴别器
生成模型
强化学习
生成语法
人工智能
机器学习
功率(物理)
电信
遗传学
物理
量子力学
探测器
生物
作者
Lantao Yu,Weinan Zhang,Jun Wang,Yong Yu
出处
期刊:Proceedings of the ... AAAI Conference on Artificial Intelligence
[Association for the Advancement of Artificial Intelligence (AAAI)]
日期:2017-02-13
卷期号:31 (1)
被引量:2245
标识
DOI:10.1609/aaai.v31i1.10804
摘要
As a new way of training generative models, Generative Adversarial Net (GAN) that uses a discriminative model to guide the training of the generative model has enjoyed considerable success in generating real-valued data. However, it has limitations when the goal is for generating sequences of discrete tokens. A major reason lies in that the discrete outputs from the generative model make it difficult to pass the gradient update from the discriminative model to the generative model. Also, the discriminative model can only assess a complete sequence, while for a partially generated sequence, it is non-trivial to balance its current score and the future one once the entire sequence has been generated. In this paper, we propose a sequence generation framework, called SeqGAN, to solve the problems. Modeling the data generator as a stochastic policy in reinforcement learning (RL), SeqGAN bypasses the generator differentiation problem by directly performing gradient policy update. The RL reward signal comes from the GAN discriminator judged on a complete sequence, and is passed back to the intermediate state-action steps using Monte Carlo search. Extensive experiments on synthetic data and real-world tasks demonstrate significant improvements over strong baselines.
科研通智能强力驱动
Strongly Powered by AbleSci AI