Your browser doesn't support javascript.
loading
ChatGPT's Response Consistency: A Study on Repeated Queries of Medical Examination Questions.
Funk, Paul F; Hoch, Cosima C; Knoedler, Samuel; Knoedler, Leonard; Cotofana, Sebastian; Sofo, Giuseppe; Bashiri Dezfouli, Ali; Wollenberg, Barbara; Guntinas-Lichius, Orlando; Alfertshofer, Michael.
Afiliación
  • Funk PF; Department of Otorhinolaryngology, Head and Neck Surgery, University Hospital Jena, Friedrich Schiller University Jena, Am Klinikum 1, 07747 Jena, Germany.
  • Hoch CC; Department of Otolaryngology, Head and Neck Surgery, School of Medicine and Health, Technical University of Munich (TUM), Ismaningerstrasse 22, 81675 Munich, Germany.
  • Knoedler S; Department of Plastic Surgery and Hand Surgery, Klinikum Rechts der Isar, Technical University of Munich (TUM), Ismaningerstrasse 22, 81675 Munich, Germany.
  • Knoedler L; Division of Plastic and Reconstructive Surgery, Massachusetts General Hospital, Harvard Medical School, 55 Fruit Street, Boston, MA 02114, USA.
  • Cotofana S; Department of Dermatology, Erasmus Medical Centre, Dr. Molewaterplein 40, 3015 GD Rotterdam, The Netherlands.
  • Sofo G; Centre for Cutaneous Research, Blizard Institute, Queen Mary University of London, Mile End Road, London E1 4NS, UK.
  • Bashiri Dezfouli A; Department of Plastic and Reconstructive Surgery, Guangdong Second Provincial General Hospital, Guangzhou 510317, China.
  • Wollenberg B; Instituto Ivo Pitanguy, Hospital Santa Casa de Misericórdia Rio de Janeiro, Pontifícia Universidade Católica do Rio de Janeiro, Rio de Janeiro 20020-022, Brazil.
  • Guntinas-Lichius O; Department of Otolaryngology, Head and Neck Surgery, School of Medicine and Health, Technical University of Munich (TUM), Ismaningerstrasse 22, 81675 Munich, Germany.
  • Alfertshofer M; Department of Otolaryngology, Head and Neck Surgery, School of Medicine and Health, Technical University of Munich (TUM), Ismaningerstrasse 22, 81675 Munich, Germany.
Eur J Investig Health Psychol Educ ; 14(3): 657-668, 2024 Mar 08.
Article en En | MEDLINE | ID: mdl-38534904
ABSTRACT
(1)

Background:

As the field of artificial intelligence (AI) evolves, tools like ChatGPT are increasingly integrated into various domains of medicine, including medical education and research. Given the critical nature of medicine, it is of paramount importance that AI tools offer a high degree of reliability in the information they provide. (2)

Methods:

A total of n = 450 medical examination questions were manually entered into ChatGPT thrice, each for ChatGPT 3.5 and ChatGPT 4. The responses were collected, and their accuracy and consistency were statistically analyzed throughout the series of entries. (3)

Results:

ChatGPT 4 displayed a statistically significantly improved accuracy with 85.7% compared to that of 57.7% of ChatGPT 3.5 (p < 0.001). Furthermore, ChatGPT 4 was more consistent, correctly answering 77.8% across all rounds, a significant increase from the 44.9% observed from ChatGPT 3.5 (p < 0.001). (4)

Conclusions:

The findings underscore the increased accuracy and dependability of ChatGPT 4 in the context of medical education and potential clinical decision making. Nonetheless, the research emphasizes the indispensable nature of human-delivered healthcare and the vital role of continuous assessment in leveraging AI in medicine.
Palabras clave

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Idioma: En Revista: Eur J Investig Health Psychol Educ Año: 2024 Tipo del documento: Article País de afiliación: Alemania

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Idioma: En Revista: Eur J Investig Health Psychol Educ Año: 2024 Tipo del documento: Article País de afiliación: Alemania