• Content Type

Research and analysis item

Four principles of explainable artificial intelligence

Abstract

We introduce four principles for explainable artiļ¬cial intelligence (AI) that comprise fundamental properties for explainable AI systems. We propose that explainable AI systems deliver accompanying evidence or reasons for outcomes and processes; provide explanations that are understandable to individual users; provide explanations that correctly reļ¬‚ect the systemā€™s process for generating the output; and that a system only operates under conditions for which it was designed and when it reaches sufļ¬cient conļ¬dence in its output. We have termed these four principles as explanation, meaningful, explanation accuracy, and knowledge limits, respectively. Through signiļ¬cant stakeholder engagement, these four principles were developed to encompass the multidisciplinary nature of explainable AI, including the ļ¬elds of computer science, engineering, and psychology. Because one-sizeļ¬ts-all explanations do not exist, different users will require different types of explanations. We present ļ¬ve categories of explanation and summarize theories of explainable AI. We give an overview of the algorithms in the ļ¬eld that cover the major classes of explainable algorithms. As a baseline comparison, we assess how well explanations provided by people follow our four principles. This assessment provides insights to the challenges of designing explainable AI systems.

Key Information

Type of organisation: Government

Date published: 25 Apr 2024

Categorisation

Domain: Horizontal
Type: Report

Type of organisation:

Discussion forum

  • Author
    Posts
  • Up
    0
    ::

    Share your thoughts on this item here.

You must be logged in to contribute to the discussion

Login