Explainable ArtificiaI Intelligence
Abstract
Artificial Intelligence (AI) has become one of the central axes of digitization strategies in the European Union (EU). Despite its great benefits, this group of emerging technologies also presents significant ethical risks and challenges that must be addressed. For this reason, the EU has proposed four fundamental ethical principles for AI: fairness, respect for human autonomy, prevention of harm, and explicability. The central concept of the paper is the explicability of AI (XAI). Besides contextualizing the relevance of XAI, this work aims to tackle some of the central aspects around it. Therefore, it will address what explicability is, why explanations are important (in the field of AI), when it is necessary to give explanations, and the different forms of explanations available to make citizens understand the decisions made by AI systems.
Downloads
Eunomía. Revista en Cultura de la Legalidad is a duly registered journal, with EISSN 2253-6655.
The articles published in Eunomía are –unless indicated otherwise– under a Creative Commons Attribution-No Derivative Works 3.0 Spain license. You can copy, distribute and communicate them publicly as long as you cite their author and the journal and institution that publishes them and do not make derivative works with them. The full license can be consulted at: http://creativecommons.org/licenses/by-nd/3.0/es/deed.es
Funding data
-
Comunidad de Madrid
Grant numbers Programa On TRUST–CM H 2019/HUM-5699