学习迁移
卷积神经网络
计算机科学
人工智能
深度学习
特征(语言学)
领域(数学分析)
核(代数)
断层(地质)
人工神经网络
模式识别(心理学)
特征提取
机器学习
数学
组合数学
地质学
数学分析
哲学
语言学
地震学
作者
Jun Zhu,Nan Chen,Changqing Shen
标识
DOI:10.1109/jsen.2019.2936932
摘要
Fault diagnosis is very important for condition based maintenance. Recently, deep learning models are introduced to learn hierarchical representations from raw data instead of using hand-crafted features, which exhibit excellent performance. The success of current deep learning lies in: 1) the training (source domain) and testing (target domain) datasets are from the same feature distribution; 2) Enough labeled data with fault information exist. However, because the machine operates under a non-stationary working condition, the trained model built on the source domain can not be directly applied on the target domain. Moreover, since no sufficient labeled or even unlabeled data are available in target domain, collecting the labeled data and building the model from scratch is time-consuming and expensive. Motivated by transfer learning (TL), we present a new fault diagnosis method, which generalizes convolutional neural network (CNN) to TL scenario. Two layers with regard to task-specific features are adapted in a layer-wise way to regularize the parameters of CNN. What's more, the domain loss is calculated by a linear combination of multiple Gaussian kernels so that the ability of adaptation is enhanced compared to single kernel. Through these two means, the distribution discrepancy is reduced and the transferable features are learned. The proposed method is validated by transfer fault diagnosis experiments. Compared to CNN without domain adaptation and shallow transfer learning methods, the proposed method gets the best performance for fault classification.
科研通智能强力驱动
Strongly Powered by AbleSci AI