透明度(行为)
自由裁量权
可信赖性
计算机科学
官僚主义
背景(考古学)
决策者
互联网隐私
计算机安全
管理科学
政治学
法学
经济
生物
政治
古生物学
作者
Stephan Grimmelikhuijsen
摘要
Abstract Algorithms based on Artificial Intelligence technologies are slowly transforming street‐level bureaucracies, yet a lack of algorithmic transparency may jeopardize citizen trust. Based on procedural fairness theory, this article hypothesizes that two core elements of algorithmic transparency (accessibility and explainability) are crucial to strengthening the perceived trustworthiness of street‐level decision‐making. This is tested in one experimental scenario with low discretion (a denied visa application) and one scenario with high discretion (a suspicion of welfare fraud). The results show that: (1) explainability has a more pronounced effect on trust than the accessibility of the algorithm; (2) the effect of algorithmic transparency not only pertains to trust in the algorithm itself but also—partially—to trust in the human decision‐maker; (3) the effects of algorithmic transparency are not robust across decision context. These findings imply that transparency‐as‐accessibility is insufficient to foster citizen trust. Algorithmic explainability must be addressed to maintain and foster trustworthiness algorithmic decision‐making.
科研通智能强力驱动
Strongly Powered by AbleSci AI