Your browser doesn't support javascript.
loading
Evaluating ChatGPT's moral competence in health care-related ethical problems.
Rashid, Ahmed A; Skelly, Ryan A; Valdes, Carlos A; Patel, Pruthvi P; Solberg, Lauren B; Giordano, Christopher R; Modave, François.
Afiliação
  • Rashid AA; Department of Anesthesiology, University of Florida College of Medicine, Gainesville, FL 32608, United States.
  • Skelly RA; Department of Anesthesiology, University of Florida College of Medicine, Gainesville, FL 32608, United States.
  • Valdes CA; Department of Surgery, University of Florida College of Medicine, Gainesville, FL 32608, United States.
  • Patel PP; Department of Research, Alabama College of Osteopathic Medicine, Dothan, AL 36303, United States.
  • Solberg LB; Department of Community Health and Family Medicine, University of Florida College of Medicine, Gainesville, FL 32608, United States.
  • Giordano CR; Department of Anesthesiology, University of Florida College of Medicine, Gainesville, FL 32608, United States.
  • Modave F; Department of Anesthesiology, University of Florida College of Medicine, Gainesville, FL 32608, United States.
JAMIA Open ; 7(3): ooae065, 2024 Oct.
Article em En | MEDLINE | ID: mdl-38983845
ABSTRACT

Objectives:

Artificial intelligence tools such as Chat Generative Pre-trained Transformer (ChatGPT) have been used for many health care-related applications; however, there is a lack of research on their capabilities for evaluating morally and/or ethically complex medical decisions. The objective of this study was to assess the moral competence of ChatGPT. Materials and

methods:

This cross-sectional study was performed between May 2023 and July 2023 using scenarios from the Moral Competence Test (MCT). Numerical responses were collected from ChatGPT 3.5 and 4.0 to assess individual and overall stage scores, including C-index and overall moral stage preference. Descriptive analysis and 2-sided Student's t-test were used for all continuous data.

Results:

A total of 100 iterations of the MCT were performed and moral preference was found to be higher in the latter Kohlberg-derived arguments. ChatGPT 4.0 was found to have a higher overall moral stage preference (2.325 versus 1.755) when compared to ChatGPT 3.5. ChatGPT 4.0 was also found to have a statistically higher C-index score in comparison to ChatGPT 3.5 (29.03 ± 11.10 versus 19.32 ± 10.95, P =.0000275).

Discussion:

ChatGPT 3.5 and 4.0 trended towards higher moral preference for the latter stages of Kohlberg's theory for both dilemmas with C-indices suggesting medium moral competence. However, both models showed moderate variation in C-index scores indicating inconsistency and further training is recommended.

Conclusion:

ChatGPT demonstrates medium moral competence and can evaluate arguments based on Kohlberg's theory of moral development. These findings suggest that future revisions of ChatGPT and other large language models could assist physicians in the decision-making process when encountering complex ethical scenarios.
Palavras-chave

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Idioma: En Revista: JAMIA Open Ano de publicação: 2024 Tipo de documento: Article

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Idioma: En Revista: JAMIA Open Ano de publicação: 2024 Tipo de documento: Article