半监督学习
机器学习
计算机科学
人工智能
熵(时间箭头)
标记数据
正规化(语言学)
加权
无监督学习
生成语法
稳健性(进化)
监督学习
数学
模式识别(心理学)
人工神经网络
生物化学
量子力学
医学
基因
物理
放射科
化学
作者
Yves Grandvalet,Yoshua Bengio
出处
期刊:Neural Information Processing Systems
日期:2004-12-01
卷期号:17: 529-536
被引量:494
摘要
We consider the semi-supervised learning problem, where a decision rule is to be learned from labeled and unlabeled data. In this framework, we motivate minimum entropy regularization, which enables to incorporate unlabeled data in the standard supervised learning. Our approach includes other approaches to the semi-supervised problem as particular or limiting cases. A series of experiments illustrates that the proposed solution benefits from unlabeled data. The method challenges mixture models when the data are sampled from the distribution class spanned by the generative model. The performances are definitely in favor of minimum entropy regularization when generative models are misspecified, and the weighting of unlabeled data provides robustness to the violation of the cluster assumption. Finally, we also illustrate that the method can also be far superior to manifold learning in high dimension spaces.
科研通智能强力驱动
Strongly Powered by AbleSci AI