Your browser doesn't support javascript.
loading
A Comparison of ChatGPT and Fine-Tuned Open Pre-Trained Transformers (OPT) Against Widely Used Sentiment Analysis Tools: Sentiment Analysis of COVID-19 Survey Data.
Lossio-Ventura, Juan Antonio; Weger, Rachel; Lee, Angela Y; Guinee, Emily P; Chung, Joyce; Atlas, Lauren; Linos, Eleni; Pereira, Francisco.
Afiliação
  • Lossio-Ventura JA; National Institute of Mental Health, National Institutes of Health, Bethesda, MD, United States.
  • Weger R; School of Medicine, University of Pittsburgh, Pittsburgh, PA, United States.
  • Lee AY; Department of Communication, Stanford University, Stanford, CA, United States.
  • Guinee EP; National Institute of Mental Health, National Institutes of Health, Bethesda, MD, United States.
  • Chung J; National Institute of Mental Health, National Institutes of Health, Bethesda, MD, United States.
  • Atlas L; National Center For Complementary and Alternative Medicine, National Institutes of Health, Bethesda, MD, United States.
  • Linos E; School of Medicine, Stanford University, Stanford, CA, United States.
  • Pereira F; National Institute of Mental Health, National Institutes of Health, Bethesda, MD, United States.
JMIR Ment Health ; 11: e50150, 2024 Jan 25.
Article em En | MEDLINE | ID: mdl-38271138
ABSTRACT

BACKGROUND:

Health care providers and health-related researchers face significant challenges when applying sentiment analysis tools to health-related free-text survey data. Most state-of-the-art applications were developed in domains such as social media, and their performance in the health care context remains relatively unknown. Moreover, existing studies indicate that these tools often lack accuracy and produce inconsistent results.

OBJECTIVE:

This study aims to address the lack of comparative analysis on sentiment analysis tools applied to health-related free-text survey data in the context of COVID-19. The objective was to automatically predict sentence sentiment for 2 independent COVID-19 survey data sets from the National Institutes of Health and Stanford University.

METHODS:

Gold standard labels were created for a subset of each data set using a panel of human raters. We compared 8 state-of-the-art sentiment analysis tools on both data sets to evaluate variability and disagreement across tools. In addition, few-shot learning was explored by fine-tuning Open Pre-Trained Transformers (OPT; a large language model [LLM] with publicly available weights) using a small annotated subset and zero-shot learning using ChatGPT (an LLM without available weights).

RESULTS:

The comparison of sentiment analysis tools revealed high variability and disagreement across the evaluated tools when applied to health-related survey data. OPT and ChatGPT demonstrated superior performance, outperforming all other sentiment analysis tools. Moreover, ChatGPT outperformed OPT, exhibited higher accuracy by 6% and higher F-measure by 4% to 7%.

CONCLUSIONS:

This study demonstrates the effectiveness of LLMs, particularly the few-shot learning and zero-shot learning approaches, in the sentiment analysis of health-related survey data. These results have implications for saving human labor and improving efficiency in sentiment analysis tasks, contributing to advancements in the field of automated sentiment analysis.
Assuntos
Palavras-chave

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Assunto principal: COVID-19 / Análise de Sentimentos Tipo de estudo: Prognostic_studies Limite: Humans País/Região como assunto: America do norte Idioma: En Revista: JMIR Ment Health Ano de publicação: 2024 Tipo de documento: Article País de afiliação: Estados Unidos

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Assunto principal: COVID-19 / Análise de Sentimentos Tipo de estudo: Prognostic_studies Limite: Humans País/Região como assunto: America do norte Idioma: En Revista: JMIR Ment Health Ano de publicação: 2024 Tipo de documento: Article País de afiliação: Estados Unidos