Your browser doesn't support javascript.
loading
Fine-tuning large neural language models for biomedical natural language processing.
Tinn, Robert; Cheng, Hao; Gu, Yu; Usuyama, Naoto; Liu, Xiaodong; Naumann, Tristan; Gao, Jianfeng; Poon, Hoifung.
Afiliação
  • Tinn R; Microsoft Research, Redmond, WA, USA.
  • Cheng H; Microsoft Research, Redmond, WA, USA.
  • Gu Y; Microsoft Research, Redmond, WA, USA.
  • Usuyama N; Microsoft Research, Redmond, WA, USA.
  • Liu X; Microsoft Research, Redmond, WA, USA.
  • Naumann T; Microsoft Research, Redmond, WA, USA.
  • Gao J; Microsoft Research, Redmond, WA, USA.
  • Poon H; Microsoft Research, Redmond, WA, USA.
Patterns (N Y) ; 4(4): 100729, 2023 Apr 14.
Article em En | MEDLINE | ID: mdl-37123444
ABSTRACT
Large neural language models have transformed modern natural language processing (NLP) applications. However, fine-tuning such models for specific tasks remains challenging as model size increases, especially with small labeled datasets, which are common in biomedical NLP. We conduct a systematic study on fine-tuning stability in biomedical NLP. We show that fine-tuning performance may be sensitive to pretraining settings and conduct an exploration of techniques for addressing fine-tuning instability. We show that these techniques can substantially improve fine-tuning performance for low-resource biomedical NLP applications. Specifically, freezing lower layers is helpful for standard BERT- B A S E models, while layerwise decay is more effective for BERT- L A R G E and ELECTRA models. For low-resource text similarity tasks, such as BIOSSES, reinitializing the top layers is the optimal strategy. Overall, domain-specific vocabulary and pretraining facilitate robust models for fine-tuning. Based on these findings, we establish a new state of the art on a wide range of biomedical NLP applications.
Palavras-chave

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Tipo de estudo: Prognostic_studies Idioma: En Ano de publicação: 2023 Tipo de documento: Article

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Tipo de estudo: Prognostic_studies Idioma: En Ano de publicação: 2023 Tipo de documento: Article