Your browser doesn't support javascript.
loading
Comparative analysis of artificial intelligence chatbot recommendations for urolithiasis management: A study of EAU guideline compliance.
Altintas, Emre; Ozkent, Mehmet Serkan; Gül, Murat; Batur, Ali Furkan; Kaynar, Mehmet; Kiliç, Özcan; Göktas, Serdar.
Afiliação
  • Altintas E; Selcuk University, Faculty of Medicine, Department of Urology, Konya, Turkey. Electronic address: dr.e.altintas@gmail.com.
  • Ozkent MS; Konya City Hospital, Department of Urology, Konya, Turkey.
  • Gül M; Selcuk University, Faculty of Medicine, Department of Urology, Konya, Turkey.
  • Batur AF; Selcuk University, Faculty of Medicine, Department of Urology, Konya, Turkey.
  • Kaynar M; Selcuk University, Faculty of Medicine, Department of Urology, Konya, Turkey.
  • Kiliç Ö; Selcuk University, Faculty of Medicine, Department of Urology, Konya, Turkey.
  • Göktas S; Selcuk University, Faculty of Medicine, Department of Urology, Konya, Turkey.
Fr J Urol ; 34(7-8): 102666, 2024 Jul.
Article em En | MEDLINE | ID: mdl-38849035
ABSTRACT

OBJECTIVES:

Artificial intelligence (AI) applications are increasingly being utilized by both patients and physicians for accessing medical information. This study focused on the urolithiasis section (pertaining to kidney and ureteral stones) of the European Association of Urology (EAU) guideline, a key reference for urologists. MATERIAL AND

METHODS:

We directed inquiries to four distinct AI chatbots to assess their responses in relation to guideline adherence. A total of 115 recommendations were transformed into questions, and responses were evaluated by two urologists with a minimum of 5 years of experience using a 5-point Likert scale (1 - False, 2 - Inadequate, 3 - Sufficient, 4 - Correct, and 5 - Very correct).

RESULTS:

The mean scores for Perplexity and ChatGPT 4.0 were 4.68 (SD 0.80) and 4.80 (SD 0.47), respectively, both significantly differed the scores of Bing and Bard (Bing vs. Perplexity, P<0.001; Bard vs. Perplexity, P<0.001; Bing vs. ChatGPT, P<0.001; Bard vs. ChatGPT, P<0.001). Bing had a mean score of 4.21 (SD 0.96), while Bard scored 3.56 (SD 1.14), with a significant difference (Bing vs. Bard, P<0.001). Bard exhibited the lowest score among all chatbots. Analysis of references revealed that Perplexity and Bing cited the guideline most frequently (47.3% and 30%, respectively).

CONCLUSION:

Our findings demonstrate that ChatGPT 4.0 and, notably, Perplexity align well with EAU guideline recommendations. These continuously evolving applications may play a crucial role in delivering information to physicians in the future, especially for urolithiasis.
Assuntos
Palavras-chave

Texto completo: 1 Base de dados: MEDLINE Assunto principal: Urologia / Inteligência Artificial / Guias de Prática Clínica como Assunto / Fidelidade a Diretrizes / Urolitíase Idioma: En Ano de publicação: 2024 Tipo de documento: Article

Texto completo: 1 Base de dados: MEDLINE Assunto principal: Urologia / Inteligência Artificial / Guias de Prática Clínica como Assunto / Fidelidade a Diretrizes / Urolitíase Idioma: En Ano de publicação: 2024 Tipo de documento: Article