Your browser doesn't support javascript.
loading
A cross-sectional comparative study: ChatGPT 3.5 versus diverse levels of medical experts in the diagnosis of ENT diseases.
Makhoul, Mikhael; Melkane, Antoine E; Khoury, Patrick El; Hadi, Christopher El; Matar, Nayla.
Affiliation
  • Makhoul M; Department of Otolaryngology-Head and Neck Surgery, Hotel Dieu de France Hospital, Saint Joseph University, Alfred Naccache Boulevard, Ashrafieh, PO Box: 166830, Beirut, Lebanon. mikhaelmakhoul651@gmail.com.
  • Melkane AE; Department of Otolaryngology-Head and Neck Surgery, Hotel Dieu de France Hospital, Saint Joseph University, Alfred Naccache Boulevard, Ashrafieh, PO Box: 166830, Beirut, Lebanon.
  • Khoury PE; Department of Otolaryngology-Head and Neck Surgery, Hotel Dieu de France Hospital, Saint Joseph University, Alfred Naccache Boulevard, Ashrafieh, PO Box: 166830, Beirut, Lebanon.
  • Hadi CE; Department of Otolaryngology-Head and Neck Surgery, Hotel Dieu de France Hospital, Saint Joseph University, Alfred Naccache Boulevard, Ashrafieh, PO Box: 166830, Beirut, Lebanon.
  • Matar N; Department of Otolaryngology-Head and Neck Surgery, Hotel Dieu de France Hospital, Saint Joseph University, Alfred Naccache Boulevard, Ashrafieh, PO Box: 166830, Beirut, Lebanon.
Eur Arch Otorhinolaryngol ; 281(5): 2717-2721, 2024 May.
Article in En | MEDLINE | ID: mdl-38365990
ABSTRACT

PURPOSE:

With recent advances in artificial intelligence (AI), it has become crucial to thoroughly evaluate its applicability in healthcare. This study aimed to assess the accuracy of ChatGPT in diagnosing ear, nose, and throat (ENT) pathology, and comparing its performance to that of medical experts.

METHODS:

We conducted a cross-sectional comparative study where 32 ENT cases were presented to ChatGPT 3.5, ENT physicians, ENT residents, family medicine (FM) specialists, second-year medical students (Med2), and third-year medical students (Med3). Each participant provided three differential diagnoses. The study analyzed diagnostic accuracy rates and inter-rater agreement within and between participant groups and ChatGPT.

RESULTS:

The accuracy rate of ChatGPT was 70.8%, being not significantly different from ENT physicians or ENT residents. However, a significant difference in correctness rate existed between ChatGPT and FM specialists (49.8%, p < 0.001), and between ChatGPT and medical students (Med2 47.5%, p < 0.001; Med3 47%, p < 0.001). Inter-rater agreement for the differential diagnosis between ChatGPT and each participant group was either poor or fair. In 68.75% of cases, ChatGPT failed to mention the most critical diagnosis.

CONCLUSIONS:

ChatGPT demonstrated accuracy comparable to that of ENT physicians and ENT residents in diagnosing ENT pathology, outperforming FM specialists, Med2 and Med3. However, it showed limitations in identifying the most critical diagnosis.
Subject(s)
Key words

Full text: 1 Collection: 01-internacional Database: MEDLINE Main subject: Artificial Intelligence / Pharyngeal Diseases Limits: Humans Language: En Journal: Eur Arch Otorhinolaryngol Journal subject: OTORRINOLARINGOLOGIA Year: 2024 Document type: Article Affiliation country: Lebanon Country of publication: Germany

Full text: 1 Collection: 01-internacional Database: MEDLINE Main subject: Artificial Intelligence / Pharyngeal Diseases Limits: Humans Language: En Journal: Eur Arch Otorhinolaryngol Journal subject: OTORRINOLARINGOLOGIA Year: 2024 Document type: Article Affiliation country: Lebanon Country of publication: Germany