Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Comparing the Performances of a 54-Year-Old Computer-Based Consultation to ChatGPT-4o
1
Zitationen
2
Autoren
2025
Jahr
Abstract
This study aimed to evaluate and compare the diagnostic responses generated by two artificial intelligence (AI) models developed 54 years apart, and encourage physicians to explore the use of large language models (LLMs) like GPT-4o in clinical practice.A clinical case of metabolic acidosis was presented to GPT-4o, and the model's diagnostic reasoning, data interpretation, and management recommendations were recorded. These outputs were then compared with the responses from Schwartz's 1970 AI model built with a decision-tree algorithm using Conversational Algebraic Language (CAL). Both models were given the same patient data to ensure a fair comparison.GPT-4o generated an advanced analysis of the patient's acid-base disturbance, correctly identifying likely causes and suggesting relevant diagnostic tests and treatments. It provided a detailed, narrative explanation of the metabolic acidosis. The 1970 CAL model, while correctly recognizing the metabolic acidosis and flagging implausible inputs, was constrained by its rule-based design. CAL offered only basic stepwise guidance and required sequential prompts for each data point, reflecting a limited capacity to handle complex or unanticipated information. GPT-4o, by contrast, integrated the data more holistically, although it occasionally ventured beyond the provided information.This comparison illustrates substantial advances in AI capabilities over five decades. GPT-4o's performance demonstrates the transformative potential of modern LLMs in clinical decision-making, showcasing abilities to synthesize complex data and assist diagnosis without specialized training, yet necessitating further validation, rigorous clinical trials, and adaptation to clinical contexts. Although innovative for its era and offering certain advantages over GPT-4o, the rule-based CAL system had technical limitations. Rather than viewing one as simply "better," this study provides perspective on how far AI in medicine has progressed while acknowledging that current AI tools remain supplements to-not replacements for-physician judgment.
Ähnliche Arbeiten
Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI
2019 · 8.336 Zit.
Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead
2019 · 8.207 Zit.
High-performance medicine: the convergence of human and artificial intelligence
2018 · 7.607 Zit.
Proceedings of the 19th International Joint Conference on Artificial Intelligence
2005 · 5.776 Zit.
Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI)
2018 · 5.476 Zit.