Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Ethical and Explainable Use of Large Language Models in Healthcare
0
Zitationen
6
Autoren
2025
Jahr
Abstract
Advent of large language models such as GPT-4, Claude, and Med-PaLM changes healthcare landscape rapidly. These models enable clinical reporting, diagnostics support, knowledge retrieval, and patient education. Impressive qualities of these models are linguistic fluency, probabilistic reasoning, and scalability making them both extremely important but and potentially dangerous. Decision-making in medicine bears irreversible consequences, in life, dignity, and justice, but explainability cannot be left. It becomes an ethical and legal imperative. This chapter discusses accounting for algorithmic bias, risks to data privacy, and misinformation in these approaches. SHAP and LIME help close gap between AI predictions and clinical reasoning processes. Regulations require governance's active participatory and rooted in beneficence, autonomy, and justice. Interdisciplinary diverse validation, open auditing, ethical committees, and patient education is way forward for chapter, setting AI in healthcare up to be powerful, fair, transparent, and trustworthy.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.635 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.543 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 8.051 Zit.
BioBERT: a pre-trained biomedical language representation model for biomedical text mining
2019 · 6.844 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.