Dokument: Evaluation of the accuracy and readability of ChatGPT-4 and Google Gemini in providing information on retinal detachment: a multicenter expert comparative study

Titel:Evaluation of the accuracy and readability of ChatGPT-4 and Google Gemini in providing information on retinal detachment: a multicenter expert comparative study
URL für Lesezeichen:https://docserv.uni-duesseldorf.de/servlets/DocumentServlet?id=67867
URN (NBN):urn:nbn:de:hbz:061-20241204-123447-8
Kollektion:Publikationen
Sprache:Englisch
Dokumententyp:Wissenschaftliche Texte » Artikel, Aufsatz
Medientyp:Text
Autoren: Strzalkowski, Piotr [Autor]
Strzalkowska, Alicja [Autor]
Chhablani, Jay [Autor]
Pfau, Kristina [Autor]
Errera, Marie-Hélène [Autor]
Roth, Mathias [Autor]
Schaub, Friederike [Autor]
Bechrakis, Nikolaos E. [Autor]
Hoerauf, Hans [Autor]
Reiter, Constantin [Autor]
Dateien:
[Dateien anzeigen]Adobe PDF
[Details]1,31 MB in einer Datei
[ZIP-Datei erzeugen]
Dateien vom 04.12.2024 / geändert 04.12.2024
Stichwörter:Patient education, Google Gemini, Language learning models, ChatGPT-4, Retinal detachment, Artificial intelligence
Beschreibung:Background

Large language models (LLMs) such as ChatGPT-4 and Google Gemini show potential for patient health education, but concerns about their accuracy require careful evaluation. This study evaluates the readability and accuracy of ChatGPT-4 and Google Gemini in answering questions about retinal detachment.
Methods

Comparative study analyzing responses from ChatGPT-4 and Google Gemini to 13 retinal detachment questions, categorized by difficulty levels (D1, D2, D3). Masked responses were reviewed by ten vitreoretinal specialists and rated on correctness, errors, thematic accuracy, coherence, and overall quality grading. Analysis included Flesch Readability Ease Score, word and sentence counts.
Results

Both Artificial Intelligence tools required college-level understanding for all difficulty levels. Google Gemini was easier to understand (p = 0.03), while ChatGPT-4 provided more correct answers for the more difficult questions (p = 0.0005) with fewer serious errors. ChatGPT-4 scored highest on most challenging questions, showing superior thematic accuracy (p = 0.003). ChatGPT-4 outperformed Google Gemini in 8 of 13 questions, with higher overall quality grades in the easiest (p = 0.03) and hardest levels (p = 0.0002), showing a lower grade as question difficulty increased.
Conclusions

ChatGPT-4 and Google Gemini effectively address queries about retinal detachment, offering mostly accurate answers with few critical errors, though patients require higher education for comprehension. The implementation of AI tools may contribute to improving medical care by providing accurate and relevant healthcare information quickly.
Rechtliche Vermerke:Originalveröffentlichung:
Strzalkowski, P., Strzalkowska, A., Chhablani, J., Pfau, K., Errera, M.-H., Roth, M., Schaub, F., Bechrakis, N. E., Hoerauf, H., Reiter, C., Schuster, A. K., Geerling, Gerd , & Guthoff, R. (2024). Evaluation of the accuracy and readability of ChatGPT-4 and Google Gemini in providing information on retinal detachment: a multicenter expert comparative study. International Journal of Retina and Vitreous, 10, Article 61. https://doi.org/10.1186/s40942-024-00579-9
Lizenz:Creative Commons Lizenzvertrag
Dieses Werk ist lizenziert unter einer Creative Commons Namensnennung 4.0 International Lizenz
Fachbereich / Einrichtung:Medizinische Fakultät
Dokument erstellt am:04.12.2024
Dateien geändert am:04.12.2024
english
Benutzer
Status: Gast
Aktionen