iSee: A case-based reasoning platform for the design of explanation experiences

  • Marta Caro-Martínez
  • , Juan A. Recio-García
  • , Belén Díaz-Agudo
  • , Jesus M. Darias
  • , Nirmalie Wiratunga
  • , Kyle Martin
  • , Anjana Wijekoon
  • , Ikechukwu Nkisi-Orji
  • , David Corsar
  • , Preeja Pradeep
  • , Derek Bridge
  • , Anne Liret

Research output: Contribution to journalArticlepeer-review

Abstract

Explainable Artificial Intelligence (XAI) is an emerging field within Artificial Intelligence (AI) that has provided many methods that enable humans to understand and interpret the outcomes of AI systems. However, deciding on the best explanation approach for a given AI problem is currently a challenging decision-making task. This paper presents the iSee project, which aims to address some of the XAI challenges by providing a unifying platform where personalized explanation experiences are generated using Case-Based Reasoning. An explanation experience includes the proposed solution to a particular explainability problem and its corresponding evaluation, provided by the end user. The ultimate goal is to provide an open catalog of explanation experiences that can be transferred to other scenarios where trustworthy AI is required.

Original languageEnglish
Article number112305
JournalKnowledge-Based Systems
Volume302
DOIs
Publication statusPublished - 25 Oct 2024

Keywords

  • Case-based reasoning
  • eXplainable artificial intelligence
  • Trustworthy AI

Fingerprint

Dive into the research topics of 'iSee: A case-based reasoning platform for the design of explanation experiences'. Together they form a unique fingerprint.

Cite this