Your browser doesn't support javascript.
loading
Artificial intelligence in dental education: ChatGPT's performance on the periodontic in-service examination.
Danesh, Arman; Pazouki, Hirad; Danesh, Farzad; Danesh, Arsalan; Vardar-Sengul, Saynur.
Affiliation
  • Danesh A; Schulich School of Medicine and Dentistry, Western University, London, Ontario, Canada.
  • Pazouki H; Faculty of Science, Western University, London, Ontario, Canada.
  • Danesh F; Elgin Mills Endodontic Specialists, Richmond Hill, Ontario, Canada.
  • Danesh A; Department of Periodontology, College of Dental Medicine, Nova Southeastern University, Davie, Florida, USA.
  • Vardar-Sengul S; Department of Periodontology, College of Dental Medicine, Nova Southeastern University, Davie, Florida, USA.
J Periodontol ; 95(7): 682-687, 2024 Jul.
Article in En | MEDLINE | ID: mdl-38197146
ABSTRACT

BACKGROUND:

ChatGPT's (Chat Generative Pre-Trained Transformer) remarkable capacity to generate human-like output makes it an appealing learning tool for healthcare students worldwide. Nevertheless, the chatbot's responses may be subject to inaccuracies, putting forth an intense risk of misinformation. ChatGPT's capabilities should be examined in every corner of healthcare education, including dentistry and its specialties, to understand the potential of misinformation associated with the chatbot's use as a learning tool. Our investigation aims to explore ChatGPT's foundation of knowledge in the field of periodontology by evaluating the chatbot's performance on questions obtained from an in-service examination administered by the American Academy of Periodontology (AAP).

METHODS:

ChatGPT3.5 and ChatGPT4 were evaluated on 311 multiple-choice questions obtained from the 2023 in-service examination administered by the AAP. The dataset of in-service examination questions was accessed through Nova Southeastern University's Department of Periodontology. Our study excluded questions containing an image as ChatGPT does not accept image inputs.

RESULTS:

ChatGPT3.5 and ChatGPT4 answered 57.9% and 73.6% of in-service questions correctly on the 2023 Periodontics In-Service Written Examination, respectively. A two-tailed t test was incorporated to compare independent sample means, and sample proportions were compared using a two-tailed χ2 test. A p value below the threshold of 0.05 was deemed statistically significant.

CONCLUSION:

While ChatGPT4 showed a higher proficiency compared to ChatGPT3.5, both chatbot models leave considerable room for misinformation with their responses relating to periodontology. The findings of the study encourage residents to scrutinize the periodontic information generated by ChatGPT to account for the chatbot's current limitations.
Subject(s)
Key words

Full text: 1 Collection: 01-internacional Database: MEDLINE Main subject: Periodontics / Artificial Intelligence / Education, Dental / Educational Measurement Type of study: Prognostic_studies Limits: Humans Language: En Journal: J Periodontol Year: 2024 Type: Article Affiliation country: Canada

Full text: 1 Collection: 01-internacional Database: MEDLINE Main subject: Periodontics / Artificial Intelligence / Education, Dental / Educational Measurement Type of study: Prognostic_studies Limits: Humans Language: En Journal: J Periodontol Year: 2024 Type: Article Affiliation country: Canada