Your browser doesn't support javascript.
loading
Evaluating ChatGPT's Performance in Answering Questions About Allergic Rhinitis and Chronic Rhinosinusitis.
Ye, Fan; Zhang, He; Luo, Xin; Wu, Tong; Yang, Qintai; Shi, Zhaohui.
Affiliation
  • Ye F; Department of Otolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou, China.
  • Zhang H; Department of Allergy, The Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou, China.
  • Luo X; Department of Otolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou, China.
  • Wu T; Department of Allergy, The Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou, China.
  • Yang Q; Department of Otolaryngology-Head and Neck Surgery, The Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou, China.
  • Shi Z; Department of Allergy, The Third Affiliated Hospital of Sun Yat-Sen University, Guangzhou, China.
Otolaryngol Head Neck Surg ; 171(2): 571-577, 2024 Aug.
Article de En | MEDLINE | ID: mdl-38796735
ABSTRACT

OBJECTIVE:

This study aims to evaluate the accuracy of ChatGPT in answering allergic rhinitis (AR) and chronic rhinosinusitis (CRS) related questions. STUDY

DESIGN:

This is a cross-sectional study.

SETTING:

Each question was inputted as a separate, independent prompt.

METHODS:

Responses to AR (n = 189) and CRS (n = 242) related questions, generated by GPT-3.5 and GPT-4, were independently graded for accuracy by 2 senior rhinology professors, with disagreements adjudicated by a third reviewer.

RESULTS:

Overall, ChatGPT demonstrated a satisfactory performance, accurately answering over 80% of questions across all categories. Specifically, GPT-4.0's accuracy in responding to AR-related questions significantly exceeded that of GPT-3.5, but distinction not evident in CRS-related questions. Patient-originated questions had a significantly higher accuracy compared to doctor-originated questions when utilizing GPT-4.0 to respond to AR-related questions. This discrepancy was not observed with GPT-3.5 or in the context of CRS-related questions. Across different types of content, ChatGPT excelled in covering basic knowledge, prevention, and emotion for AR and CRS. However, it experienced challenges when addressing questions about recent advancements, a trend consistent across both GPT-3.5 and GPT-4.0 iterations. Importantly, the accuracy of responses remained unaffected when questions were posed in Chinese.

CONCLUSION:

Our findings suggest ChatGPT's capability to convey accurate information for AR and CRS patients, and offer insights into its performance across various domains, guiding its utilization and improvement.
Sujet(s)
Mots clés

Texte intégral: 1 Collection: 01-internacional Base de données: MEDLINE Sujet principal: Sinusite / Rhinite allergique Limites: Female / Humans / Male Langue: En Journal: Otolaryngol Head Neck Surg / Otolaryngol. head neck surg / Otolaryngology - Head and neck surgery Sujet du journal: OTORRINOLARINGOLOGIA Année: 2024 Type de document: Article Pays d'affiliation: Chine Pays de publication: Royaume-Uni

Texte intégral: 1 Collection: 01-internacional Base de données: MEDLINE Sujet principal: Sinusite / Rhinite allergique Limites: Female / Humans / Male Langue: En Journal: Otolaryngol Head Neck Surg / Otolaryngol. head neck surg / Otolaryngology - Head and neck surgery Sujet du journal: OTORRINOLARINGOLOGIA Année: 2024 Type de document: Article Pays d'affiliation: Chine Pays de publication: Royaume-Uni