Back to Search Start Over

Trustworthy and Responsible AI for Human-Centric Autonomous Decision-Making Systems

Authors :
Dehghani, Farzaneh
Dibaji, Mahsa
Anzum, Fahim
Dey, Lily
Basdemir, Alican
Bayat, Sayeh
Boucher, Jean-Christophe
Drew, Steve
Eaton, Sarah Elaine
Frayne, Richard
Ginde, Gouri
Harris, Ashley
Ioannou, Yani
Lebel, Catherine
Lysack, John
Arzuaga, Leslie Salgado
Stanley, Emma
Souza, Roberto
Santos, Ronnie de Souza
Wells, Lana
Williamson, Tyler
Wilms, Matthias
Wahid, Zaman
Ungrin, Mark
Gavrilova, Marina
Bento, Mariana
Publication Year :
2024

Abstract

Artificial Intelligence (AI) has paved the way for revolutionary decision-making processes, which if harnessed appropriately, can contribute to advancements in various sectors, from healthcare to economics. However, its black box nature presents significant ethical challenges related to bias and transparency. AI applications are hugely impacted by biases, presenting inconsistent and unreliable findings, leading to significant costs and consequences, highlighting and perpetuating inequalities and unequal access to resources. Hence, developing safe, reliable, ethical, and Trustworthy AI systems is essential. Our team of researchers working with Trustworthy and Responsible AI, part of the Transdisciplinary Scholarship Initiative within the University of Calgary, conducts research on Trustworthy and Responsible AI, including fairness, bias mitigation, reproducibility, generalization, interpretability, and authenticity. In this paper, we review and discuss the intricacies of AI biases, definitions, methods of detection and mitigation, and metrics for evaluating bias. We also discuss open challenges with regard to the trustworthiness and widespread application of AI across diverse domains of human-centric decision making, as well as guidelines to foster Responsible and Trustworthy AI models.<br />Comment: 44 pages, 2 figures

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2408.15550
Document Type :
Working Paper