Evidence-based Explanation to Promote Fairness in AI Systems


Workshop paper


Juliana Jansen Ferreira, Mateus de Souza Monteiro

Cite

Cite

APA   Click to copy
Ferreira, J. J., & de Souza Monteiro, M. Evidence-based Explanation to Promote Fairness in AI Systems.


Chicago/Turabian   Click to copy
Ferreira, Juliana Jansen, and Mateus de Souza Monteiro. Evidence-Based Explanation to Promote Fairness in AI Systems, n.d.


MLA   Click to copy
Ferreira, Juliana Jansen, and Mateus de Souza Monteiro. Evidence-Based Explanation to Promote Fairness in AI Systems.


BibTeX   Click to copy

@techreport{juliana-a,
  title = {Evidence-based Explanation to Promote Fairness in AI Systems},
  author = {Ferreira, Juliana Jansen and de Souza Monteiro, Mateus}
}

Abstract
As Artificial Intelligence (AI) technology gets more intertwined with every system, people are using AI to make decisions on their everyday activities. In simple contexts, such as Netflix recommendations, or in more complex context like in judicial scenarios, AI is part of people's decisions. People make decisions and usually, they need to explain their decision to others or in some matter. It is particularly critical in contexts where human expertise is central to decision-making. In order to explain their decisions with AI support, people need to understand how AI is part of that decision. When considering the aspect of fairness, the role that AI has on a decision-making process becomes even more sensitive since it affects the fairness and the responsibility of those people making the ultimate decision. We have been exploring an evidence-based explanation design approach to 'tell the story of a decision'. In this position paper, we discuss our approach for AI systems using fairness sensitive cases in the literature.

PDF

Share



Follow this website


You need to create an Owlstown account to follow this website.


Sign up

Already an Owlstown member?

Log in