Explanation matters: An experimental study on explainable AI

  • Pascal Hamm
  • , Michael Klesel*
  • , Patricia Coberger
  • , H. Felix Wittmann
  • *Corresponding author for this work

Research output: Contribution to journalArticleAcademicpeer-review

42 Citations (Scopus)
144 Downloads (Pure)

Abstract

Explainable artificial intelligence (XAI) is an important advance in the field of machine learning to shed light on black box algorithms and thus a promising approach to improving artificial intelligence (AI) adoption. While previous literature has already addressed the technological benefits of XAI, there has been little research on XAI from the user’s perspective. Building upon the theory of trust, we propose a model that hypothesizes that post hoc explainability (using Shapley Additive Explanations) has a significant impact on use-related variables in this context. To test our model, we designed an experiment using a randomized controlled trial design where participants compare signatures and detect forged signatures. Surprisingly, our study shows that XAI only has a small but significant impact on perceived explainability. Nevertheless, we demonstrate that a high level of perceived explainability has a strong impact on important constructs including trust and perceived usefulness. A post hoc analysis shows that hedonic factors are significantly related to perceived explainability and require more attention in future research. We conclude with important directions for academia and for organizations.

Original languageEnglish
Article number17
JournalElectronic markets
Volume33
Issue number1
Early online date10 May 2023
DOIs
Publication statusPublished - Dec 2023

Keywords

  • AI
  • Artificial intelligence
  • Experiment
  • Perception
  • XAI
  • UT-Hybrid-D

Fingerprint

Dive into the research topics of 'Explanation matters: An experimental study on explainable AI'. Together they form a unique fingerprint.

Cite this