可解释性
计算机科学
统一
机器学习
班级(哲学)
人工智能
一致性(知识库)
直觉
统一模型
夏普里值
特征(语言学)
理论计算机科学
数学
博弈论
气象学
数理经济学
程序设计语言
哲学
物理
认识论
语言学
作者
Scott Lundberg,Su‐In Lee
标识
DOI:10.4230/oasics.dx.2024.27
摘要
In an industrial maintenance context, degradation diagnosis is the problem of determining the current level of degradation of operating machines based on measurements. With the emergence of Machine Learning techniques, such a problem can now be solved by training a degradation model offline and by using it online. While such models are more and more accurate and performant, they are often black-box and their decisions are therefore not interpretable for human maintenance operators. On the contrary, interpretable ML models are able to provide explanations for the model’s decisions and consequently improves the confidence of the human operator about the maintenance decision based on these models. This paper proposes a new method to quantitatively measure the interpretability of such models that is agnostic (no assumption about the class of models) and that is applied on degradation models. The proposed method requires that the decision maker sets up some high level parameters in order to measure the interpretability of the models and then can decide whether the obtained models are satisfactory or not. The method is formally defined and is fully illustrated on a decision tree degradation model and a model trained with a recent neural network architecture called Multiclass Neural Additive Model.
科研通智能强力驱动
Strongly Powered by AbleSci AI