Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
The Role of Explainability in Human–AI Co-Decision Making
0
Zitationen
2
Autoren
2026
Jahr
Abstract
The rapid deployment of artificial intelligence (AI) in decision-support systems has transformed the way humans interact with computational models. While modern AI systems often achieve high predictive accuracy, their lack of transparency can undermine user trust and limit effective collaboration. Human–AI co-decision making, where both human judgment and AI recommendations jointly influence outcomes, requires explainability as a foundational capability rather than an optional feature. This paper investigates the role of explainable artificial intelligence (XAI) in improving co-decision quality, trust calibration, and accountability. A comprehensive literature review is presented, followed by identification of key research gaps. We propose an Explainable Co-Decision Framework (ECDF) that integrates predictive modeling, explanation generation, and adaptive human feedback. Using a simulated risk-assessment dataset comprising 5,000 instances, the framework is evaluated across multiple conditions. Experimental results demonstrate that structured explanations improve joint decision accuracy by up to 10%, reduce trust calibration error by more than 60%, and enhance human engagement with AI outputs. The findings highlight that explanation quality—not merely availability—plays a decisive role in human–AI teaming. The paper concludes with design recommendations and future research directions for robust explainable co-decision systems.
Ähnliche Arbeiten
Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization
2017 · 20.792 Zit.
Generative Adversarial Nets
2023 · 19.896 Zit.
Visualizing and Understanding Convolutional Networks
2014 · 15.331 Zit.
"Why Should I Trust You?"
2016 · 14.605 Zit.
Generative adversarial networks
2020 · 13.213 Zit.