Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Analyzing Trustworthiness and Explainability in Artificial Intelligence: A Comprehensive Review
1
Zitationen
5
Autoren
2024
Jahr
Abstract
Artificial intelligence (AI) has become an important driver in the current dynamic technological environment, presenting itself as a revolutionary power capable of reconfiguring various sectors, economies, and social structures. The paper aims to address a wide range of readers, encompassing AI practitioners, academics, and people in general. Its primary objective is to connect the complex technical aspects of AI and the ethical problems inherent in its creation and implementation. In an era marked by the growing integration of AI systems into various aspects of human existence, the book offers fundamental ideas that contribute to cultivating an environment where these systems function with transparency, ethical considerations, and reliability. The paper's comprehensive coverage spans various subjects that contribute to a complete comprehension of the intricate terrain of reliable AI. The analysis is initiated by conducting an indepth examination of the architectural aspects of AI systems, elucidating the progression from the input of data to the generation of decision-making outcomes. The text introduces the core functions of AI, explores its conceptual framework, and emphasizes the significance of data processing modules, computations, Machine Learning models (ML), and integrating software. This foundational framework establishes a basis for subsequent investigation into the pivotal concepts of integrity, trust, and ethics. This paper bravely tackles urgent issues about bias, justice, and the erosion of data privacy while offering practical solutions to increase AI system openness and explainability by 20%. This paper examines various strategies to improve transparency and explainability, recognizing the importance of strengthening user understanding and confidence. Within the realm of healthcare, the paper acquaints readers with the pioneering notion of Federated Deep Learning, which can improve data privacy by up to 30%. This includes a dedicated part that delves into the concept of explainable AI, introducing various methodologies such as LIME and SHAP, which are employed to interpret predictions made by AI models. The paper provides readers with the knowledge to traverse the ever-changing environment of AI safely and ethically. It emphasizes the importance of utilizing AI's transformative potential for improving humanity while maintaining the utmost adherence to ethical principles.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.593 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.483 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 8.003 Zit.
BioBERT: a pre-trained biomedical language representation model for biomedical text mining
2019 · 6.824 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.