Analysis of Responses of GPT-4 V to the Japanese National Clinical Engineer Licensing Examination.
J Med Syst
; 48(1): 83, 2024 Sep 11.
Article
in En
| MEDLINE
| ID: mdl-39259341
ABSTRACT
Chat Generative Pretrained Transformer (ChatGPT; OpenAI) is a state-of-the-art large language model that can simulate human-like conversations based on user input. We evaluated the performance of GPT-4 V in the Japanese National Clinical Engineer Licensing Examination using 2,155 questions from 2012 to 2023. The average correct answer rate for all questions was 86.0%. In particular, clinical medicine, basic medicine, medical materials, biological properties, and mechanical engineering achieved a correct response rate of ≥ 90%. Conversely, medical device safety management, electrical and electronic engineering, and extracorporeal circulation obtained low correct answer rates ranging from 64.8% to 76.5%. The correct answer rates for questions that included figures/tables, required numerical calculation, figure/table â© calculation, and knowledge of Japanese Industrial Standards were 55.2%, 85.8%, 64.2% and 31.0%, respectively. The reason for the low correct answer rates is that ChatGPT lacked recognition of the images and knowledge of standards and laws. This study concludes that careful attention is required when using ChatGPT because several of its explanations lack the correct description.
Key words
Full text:
1
Collection:
01-internacional
Database:
MEDLINE
Main subject:
Biomedical Engineering
Country/Region as subject:
Asia
Language:
En
Journal:
J Med Syst
Year:
2024
Document type:
Article
Affiliation country:
Country of publication: