Identifying XAI User Needs

Authors Jenia Kim, Henry Maathuis, Kees van Montfort, Danielle Sent
Published in Proceedings of the Workshops at the Third International Conference on Hybrid Human-Artificial Intelligence
Publication date 10 June 2024
Research groups Artificial Intelligence
Type Lecture

Summary

One aspect of a responsible application of Artificial Intelligence (AI) is ensuring that the operation and outputs of an AI system are understandable for non-technical users, who need to consider its recommendations in their decision making. The importance of explainable AI (XAI) is widely acknowledged; however, its practical implementation is not straightforward. In particular, it is still unclear what the requirements are of non-technical users from explanations, i.e. what makes an explanation meaningful. In this paper, we synthesize insights on meaningful explanations from a literature study and two use cases in the financial sector. We identified 30 components of meaningfulness in XAI literature. In addition, we report three themes associated with explanation needs that were central to the users in our use cases, but are not prominently described in literature: actionability, coherent narratives and context. Our results highlight the importance of narrowing the gap between theoretical and applied responsible AI.

Downloads en links

On this publication contributed

  • Jenia Kim
    • Researcher
    • Research group: Artificial Intelligence
  • Danielle Sent
    • Senior lecturer
    • Research group: Artificial Intelligence

Language English
Published in Proceedings of the Workshops at the Third International Conference on Hybrid Human-Artificial Intelligence
Year and volume 3 3825
Key words Explainable AI, Finance, Human-Centered Evaluation
Page range 221-227