Your browser doesn't support javascript.
loading
A Study of Biomedical Relation Extraction Using GPT Models.
Zhang, Jeffrey; Wibert, Maxwell; Zhou, Huixue; Peng, Xueqing; Chen, Qingyu; Keloth, Vipina K; Hu, Yan; Zhang, Rui; Xu, Hua; Raja, Kalpana.
Afiliação
  • Zhang J; Section for Biomedical Informatics and Data Science, School of Medicine, Yale University, New Haven, USA.
  • Wibert M; Section for Biomedical Informatics and Data Science, School of Medicine, Yale University, New Haven, USA.
  • Zhou H; Institute for Health Informatics, University of Minnesota, Twin Cities, USA.
  • Peng X; Section for Biomedical Informatics and Data Science, School of Medicine, Yale University, New Haven, USA.
  • Chen Q; Section for Biomedical Informatics and Data Science, School of Medicine, Yale University, New Haven, USA.
  • Keloth VK; Section for Biomedical Informatics and Data Science, School of Medicine, Yale University, New Haven, USA.
  • Hu Y; School of Biomedical Informatics, University of Texas Health Science at Houston, Houston, USA.
  • Zhang R; Department of Surgery, Minneapolis, School of Medicine, University of Minnesota, Minneapolis, USA.
  • Xu H; Section for Biomedical Informatics and Data Science, School of Medicine, Yale University, New Haven, USA.
  • Raja K; Section for Biomedical Informatics and Data Science, School of Medicine, Yale University, New Haven, USA.
AMIA Jt Summits Transl Sci Proc ; 2024: 391-400, 2024.
Article em En | MEDLINE | ID: mdl-38827097
ABSTRACT
Relation Extraction (RE) is a natural language processing (NLP) task for extracting semantic relations between biomedical entities. Recent developments in pre-trained large language models (LLM) motivated NLP researchers to use them for various NLP tasks. We investigated GPT-3.5-turbo and GPT-4 on extracting the relations from three standard datasets, EU-ADR, Gene Associations Database (GAD), and ChemProt. Unlike the existing approaches using datasets with masked entities, we used three versions for each dataset for our experiment a version with masked entities, a second version with the original entities (unmasked), and a third version with abbreviations replaced with the original terms. We developed the prompts for various versions and used the chat completion model from GPT API. Our approach achieved a F1-score of 0.498 to 0.809 for GPT-3.5-turbo, and a highest F1-score of 0.84 for GPT-4. For certain experiments, the performance of GPT, BioBERT, and PubMedBERT are almost the same.
Palavras-chave

Texto completo: 1 Base de dados: MEDLINE Idioma: En Ano de publicação: 2024 Tipo de documento: Article

Texto completo: 1 Base de dados: MEDLINE Idioma: En Ano de publicação: 2024 Tipo de documento: Article