Back to Search Start Over

Evaluation of ChatGPT as a Reliable Source of Medical Information on Prostate Cancer for Patients: Global Comparative Survey of Medical Oncologists and Urologists.

Authors :
Stenzl A
Armstrong AJ
Rogers E
Habr D
Walz J
Gleave M
Sboner A
Ghith J
Serfass L
Schuler KW
Garas S
Chari D
Truman K
Sternberg CN
Source :
Urology practice [Urol Pract] 2024 Nov 07, pp. 101097UPJ0000000000000740. Date of Electronic Publication: 2024 Nov 07.
Publication Year :
2024
Publisher :
Ahead of Print

Abstract

Introduction: No consensus exists on performance standards for evaluation of generative artificial intelligence (AI) to generate medical responses. The purpose of this study was the assessment of Chat Generative Pre-trained Transformer (ChatGPT) to address medical questions in prostate cancer.<br />Methods: A global online survey was conducted from April to June 2023 among > 700 medical oncologists or urologists who treat patients with prostate cancer. Participants were unaware this was a survey evaluating AI. In component 1, responses to 9 questions were written independently by medical writers (MWs; from medical websites) and ChatGPT-4.0 (AI generated from publicly available information). Respondents were randomly exposed and blinded to both AI-generated and MW-curated responses; evaluation criteria and overall preference were recorded. Exploratory component 2 evaluated AI-generated responses to 5 complex questions with nuanced answers in the medical literature. Responses were evaluated on a 5-point Likert scale. Statistical significance was denoted by P < .05.<br />Results: In component 1, respondents (N = 602) consistently preferred the clarity of AI-generated responses over MW-curated responses in 7 of 9 questions ( P < .05). Despite favoring AI-generated responses when blinded to questions/answers, respondents considered medical websites a more credible source (52%-67%) than ChatGPT (14%). Respondents in component 2 (N = 98) also considered medical websites more credible than ChatGPT, but rated AI-generated responses highly for all evaluation criteria, despite nuanced answers in the medical literature.<br />Conclusions: These findings provide insight into how clinicians rate AI-generated and MW-curated responses with evaluation criteria that can be used in future AI validation studies.

Details

Language :
English
ISSN :
2352-0787
Database :
MEDLINE
Journal :
Urology practice
Publication Type :
Academic Journal
Accession number :
39509585
Full Text :
https://doi.org/10.1097/UPJ.0000000000000740