可解释性
计算机科学
机器学习
人工智能
公制(单位)
多样性(控制论)
因果关系(物理学)
数据科学
特征(语言学)
因果模型
黑匣子
管理科学
医学
病理
语言学
运营管理
物理
哲学
量子力学
经济
作者
Raha Moraffah,Mansooreh Karami,Ruocheng Guo,Adrienne Raglin,Huan Liu
出处
期刊:SIGKDD explorations
[Association for Computing Machinery]
日期:2020-05-13
卷期号:22 (1): 18-33
被引量:191
标识
DOI:10.1145/3400051.3400058
摘要
Machine learning models have had discernible achievements in a myriad of applications. However, most of these models are black-boxes, and it is obscure how the decisions are made by them. This makes the models unreliable and untrustworthy. To provide insights into the decision making processes of these models, a variety of traditional interpretable models have been proposed. Moreover, to generate more humanfriendly explanations, recent work on interpretability tries to answer questions related to causality such as "Why does this model makes such decisions?" or "Was it a specific feature that caused the decision made by the model?". In this work, models that aim to answer causal questions are referred to as causal interpretable models. The existing surveys have covered concepts and methodologies of traditional interpretability. In this work, we present a comprehensive survey on causal interpretable models from the aspects of the problems and methods. In addition, this survey provides in-depth insights into the existing evaluation metrics for measuring interpretability, which can help practitioners understand for what scenarios each evaluation metric is suitable.
科研通智能强力驱动
Strongly Powered by AbleSci AI