过度拟合
计算机科学
卷积神经网络
正规化(语言学)
人工智能
机器学习
稳健性(进化)
编码(集合论)
人工神经网络
训练集
深层神经网络
模式识别(心理学)
基因
生物化学
集合(抽象数据类型)
化学
程序设计语言
作者
Terrance DeVries,Graham W. Taylor
出处
期刊:Cornell University - arXiv
日期:2017-08-15
被引量:822
摘要
Convolutional neural networks are capable of learning powerful representational spaces, which are necessary for tackling complex learning tasks. However, due to the model capacity required to capture such representations, they are often susceptible to overfitting and therefore require proper regularization in order to generalize well. In this paper, we show that the simple regularization technique of randomly masking out square regions of input during training, which we call cutout, can be used to improve the robustness and overall performance of convolutional neural networks. Not only is this method extremely easy to implement, but we also demonstrate that it can be used in conjunction with existing forms of data augmentation and other regularizers to further improve model performance. We evaluate this method by applying it to current state-of-the-art architectures on the CIFAR-10, CIFAR-100, and SVHN datasets, yielding new state-of-the-art results of 2.56%, 15.20%, and 1.30% test error respectively. Code is available at this https URL
科研通智能强力驱动
Strongly Powered by AbleSci AI