OpenAlex · Aktualisierung stündlich · Letzte Aktualisierung: 29.03.2026, 14:07

Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.

Enhancing decision-making in glioblastoma surgery through an explainable human-AI collaboration: an international multicenter model development and external validation study

2025·0 Zitationen·npj Precision OncologyOpen Access
Volltext beim Verlag öffnen

0

Zitationen

29

Autoren

2025

Jahr

Abstract

Surgical resection improves survival in glioblastoma, yet predicting the extent of resection (EOR) remains highly challenging. We developed and externally validated an explainable AI model to generate personalized EOR estimates in 811 glioblastoma patients undergoing microsurgical resection. EOR was categorized into gross-total (GTR), near-total (NTR), and subtotal resections (STR). An interpretable framework provided model explanations and sensitivity analyses to assess the model's strengths and limitations. To demonstrate clinical impact, we compared the performance of the human expert (gold standard) with our AI model and a combined human-AI approach. External validation confirmed generalizability (AUC 0.78, CI 0.73-0.82). Class-specific AUCs were 0.75 (0.67-0.82) for GTR, 0.59 (0.50-0.69) for NTR, and 0.69 (0.53-0.85) for STR. Key predictors included KPS and NANO scores, age, tumor volume, and unfavorable anatomical locations. A combined human-AI collaboration outperformed human experts, with higher overall accuracies (0.53 to 0.94), F1 scores (0.30 to 0.92), and Cohen's κ (0.41 to 0.84). Enhancing predictive performance through the clinician-AI collaboration, our explainable model supports preoperative planning and highlights the value of integrating machine intelligence into surgical decision-making.

Ähnliche Arbeiten