Your browser doesn't support javascript.
loading
Charting new AI education in gastroenterology: Cross-sectional evaluation of ChatGPT and perplexity AI in medical residency exam.
Gravina, Antonietta Gerarda; Pellegrino, Raffaele; Palladino, Giovanna; Imperio, Giuseppe; Ventura, Andrea; Federico, Alessandro.
Afiliação
  • Gravina AG; Hepatogastroenterology Division, Department of Precision Medicine, University of Campania Luigi Vanvitelli, Via Luigi de Crecchio, 80138, Naples, Italy.
  • Pellegrino R; Hepatogastroenterology Division, Department of Precision Medicine, University of Campania Luigi Vanvitelli, Via Luigi de Crecchio, 80138, Naples, Italy. Electronic address: raffaele.pellegrino@unicampania.it.
  • Palladino G; Hepatogastroenterology Division, Department of Precision Medicine, University of Campania Luigi Vanvitelli, Via Luigi de Crecchio, 80138, Naples, Italy.
  • Imperio G; Hepatogastroenterology Division, Department of Precision Medicine, University of Campania Luigi Vanvitelli, Via Luigi de Crecchio, 80138, Naples, Italy.
  • Ventura A; Hepatogastroenterology Division, Department of Precision Medicine, University of Campania Luigi Vanvitelli, Via Luigi de Crecchio, 80138, Naples, Italy.
  • Federico A; Hepatogastroenterology Division, Department of Precision Medicine, University of Campania Luigi Vanvitelli, Via Luigi de Crecchio, 80138, Naples, Italy.
Dig Liver Dis ; 56(8): 1304-1311, 2024 Aug.
Article em En | MEDLINE | ID: mdl-38503659
ABSTRACT

BACKGROUND:

Conversational chatbots, fueled by large language models, spark debate over their potential in education and medical career exams. There is debate in the literature about the scientific integrity of the outputs produced by these chatbots.

AIMS:

This study evaluates ChatGPT 3.5 and Perplexity AI's cross-sectional performance in responding to questions from the 2023 Italian national residency admission exam (SSM23), comparing results and chatbots' concordance with previous years SSMs.

METHODS:

Gastroenterology-related SSM23 questions were input into ChatGPT 3.5 and Perplexity AI, evaluating their performance in correct responses and total scores. This process was repeated with questions from the three preceding years. Additionally, chatbot concordance was assessed using Cohen's method.

RESULTS:

In SSM23, ChatGPT 3.5 outperforms Perplexity AI with 94.11% correct responses, demonstrating consistency across years. Concordance weakened in 2023 (κ=0.203, P = 0.148), but ChatGPT consistently maintains a high standard compared to Perplexity AI.

CONCLUSION:

ChatGPT 3.5 and Perplexity AI exhibit promise in addressing gastroenterological queries, emphasizing potential educational roles. However, their variable performance mandates cautious use as supplementary tools alongside conventional study methods. Clear guidelines are crucial for educators to balance traditional approaches and innovative systems, enhancing educational standards.
Assuntos
Palavras-chave

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Assunto principal: Gastroenterologia / Internato e Residência Limite: Humans País/Região como assunto: Europa Idioma: En Revista: Dig Liver Dis Assunto da revista: GASTROENTEROLOGIA Ano de publicação: 2024 Tipo de documento: Article País de afiliação: Itália

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Assunto principal: Gastroenterologia / Internato e Residência Limite: Humans País/Região como assunto: Europa Idioma: En Revista: Dig Liver Dis Assunto da revista: GASTROENTEROLOGIA Ano de publicação: 2024 Tipo de documento: Article País de afiliação: Itália