How model accuracy and explanation fidelity influence user trust

Research output: Working paperPreprintAcademic

7 Downloads (Pure)

Abstract

Machine learning systems have become popular in fields such as marketing, financing, or data mining. While they are highly accurate, complex machine learning systems pose challenges for engineers and users. Their inherent complexity makes it impossible to easily judge their fairness and the correctness of statistically learned relations between variables and classes. Explainable AI aims to solve this challenge by modelling explanations alongside with the classifiers, potentially improving user trust and acceptance. However, users should not be fooled by persuasive, yet untruthful explanations. We therefore conduct a user study in which we investigate the effects of model accuracy and explanation fidelity, i.e. how truthfully the explanation represents the underlying model, on user trust. Our findings show that accuracy is more important for user trust than explainability. Adding an explanation for a classification result can potentially harm trust, e.g. when adding nonsensical explanations. We also found that users cannot be tricked by high-fidelity explanations into having trust for a bad classifier. Furthermore, we found a mismatch between observed (implicit) and self-reported (explicit) trust.
Original languageEnglish
PublisherArXiv.org
DOIs
Publication statusPublished - 26 Jul 2019

Keywords

  • cs.CY
  • cs.AI
  • cs.LG

Fingerprint

Dive into the research topics of 'How model accuracy and explanation fidelity influence user trust'. Together they form a unique fingerprint.

Cite this