Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Navigating Criminal Liability in an Era Of AI-Assisted Medicine
1
Zitationen
2
Autoren
2025
Jahr
Abstract
This paper explores the interplay between artificial intelligence (AI) and criminal liability within the healthcare sector, particularly in the context of medical malpractice. Adopting a multidisciplinary approach, the study evaluates current legal frameworks and their adequacy in addressing liability for errors involving AI-driven medical systems. Through an analysis of legal theory, case studies, and technological integration, the research highlights the complexities of assigning liability when errors arise from AI-assisted medical decision-making. The methodology includes a comparative legal analysis and a detailed examination of a real-world case involving AI-related treatment errors. Findings reveal that while existing legal frameworks are sufficient for holding humans accountable under the standard of care, they struggle with the unique challenges posed by AI's "black-box" nature. The study argues that further refinement of liability models is necessary, especially as AI systems gain greater autonomy. The paper concludes by offering a roadmap for balancing innovation in AI with the imperative to protect patient rights, emphasizing that liability frameworks must evolve in tandem with technological advancements.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.400 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.261 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.695 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.781 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.506 Zit.