Dies ist eine Übersichtsseite mit Metadaten zu dieser wissenschaftlichen Arbeit. Der vollständige Artikel ist beim Verlag verfügbar.
Evaluating AI Chatbots in Addressing the Unmet Survivorship Needs of Adolescents and Young Adults with Melanoma (Preprint)
0
Zitationen
9
Autoren
2025
Jahr
Abstract
<sec> <title>BACKGROUND</title> Melanoma, a highly aggressive form of skin cancer, is the second most common type of cancer for adolescent and young adult (AYA, ages 15-39 years) patients. AYA melanoma patients may turn to internet sources, especially AI-chatbots, to manage uncertainty about prognosis and treatment. </sec> <sec> <title>OBJECTIVE</title> To evaluate the quality, empathy, and readability of responses generated by leading AI chatbots when addressing the top unmet needs of AYA melanoma patients receiving treatment. </sec> <sec> <title>METHODS</title> Our research team recently surveyed 152 AYA melanoma patients using the Needs Assessment Service Bridge (NA-SB), a validated instrument that assesses psychosocial needs for AYA cancer patients. The survey identified the top 5 needs for advanced AYA melanoma patients receiving treatment. Each need was reframed into a question and brief clinical history, then entered into each chatbot by five individuals who cleared their pre- and post-question history. Chatbot responses were evaluated to assess information quality (Global Quality Score (GQS) and DISCERN), accessibility and readability (GQS, Flesch Kincaid Grade Level, Flesch Reading Ease), and perceived empathy (Perceived Empathy of Technology (PETS)). </sec> <sec> <title>RESULTS</title> Across 75 chatbot responses, ChatGPT achieved the highest average quality (mean GQS 4.4, mean DISC 3.2) and empathy (PETS-ER 5.4, PETS-UT 6.4), though with greater variability (SD~1.8). Copilot produced the lowest quality and empathy scores, while Gemini responses were consistently midrange. PETS-UT exceeded PETS-ER across all models, suggesting stronger cognitive empathy than emotional responsiveness. Readability analysis showed outputs exceeded the average U.S. reading level (mean FKGL 11.8, FRE 38.6), limiting accessibility. The most readable responses were found in Question 2, which also scored higher in quality and empathy, whereas Question 4 and 5 produced the most complex, difficult to read responses corresponding with lower quality and empathy ratings. </sec> <sec> <title>CONCLUSIONS</title> AI chatbots can provide moderately accurate and supportive responses to AYA melanoma patient needs, but outputs are inconsistent, written above the recommended reading level for health information, and limited in empathy. Question framing strongly influenced chatbot performance, with more emotional prompts drawing greater empathy, and readability aligning with both quality and empathy. Chatbot use in this population should remain adjunctive, with further research needed to standardize quality, improve readability, and enhance empathetic communication. </sec>
Ähnliche Arbeiten
Sedentary Behavior Research Network (SBRN) – Terminology Consensus Project process and outcome
2017 · 3.704 Zit.
Health Literacy
2004 · 3.419 Zit.
Systematic Review: Impact of Health Information Technology on Quality, Efficiency, and Costs of Medical Care
2006 · 3.173 Zit.
Effects of Computerized Clinical Decision Support Systems on Practitioner Performance and Patient Outcomes
2005 · 2.967 Zit.
Interventions for enhancing medication adherence
2014 · 2.959 Zit.