Your browser doesn't support javascript.
loading
Position-Enhanced Multi-Head Self-Attention Based Bidirectional Gated Recurrent Unit for Aspect-Level Sentiment Classification.
Li, Xianyong; Ding, Li; Du, Yajun; Fan, Yongquan; Shen, Fashan.
Afiliação
  • Li X; School of Computer and Software Engineering, Xihua University, Chengdu, China.
  • Ding L; School of Computer and Software Engineering, Xihua University, Chengdu, China.
  • Du Y; School of Computer and Software Engineering, Xihua University, Chengdu, China.
  • Fan Y; School of Computer and Software Engineering, Xihua University, Chengdu, China.
  • Shen F; Sichuan Suitang Science and Technology Co., Ltd., Chengdu, China.
Front Psychol ; 12: 799926, 2021.
Article em En | MEDLINE | ID: mdl-35145460
ABSTRACT
Aspect-level sentiment classification (ASC) is an interesting and challenging research task to identify the sentiment polarities of aspect words in sentences. Previous attention-based methods rarely consider the position information of aspect and contextual words. For an aspect word in a sentence, its adjacent words should be given more attention than the long distant words. Based on this consideration, this article designs a position influence vector to represent the position information between an aspect word and the context. By combining the position influence vector, multi-head self-attention mechanism and bidirectional gated recurrent unit (BiGRU), a position-enhanced multi-head self-attention network based BiGRU (PMHSAT-BiGRU) model is proposed. To verify the effectiveness of the proposed model, this article makes a large number of experiments on SemEval2014 restaurant, SemEval2014 laptop, SemEval2015 restaurant, and SemEval2016 restaurant data sets. The experiment results show that the performance of the proposed PMHSAT-BiGRU model is obviously better than the baselines. Specially, compared with the original LSTM model, the Accuracy values of the proposed PMHSAT-BiGRU model on the four data sets are improved by 5.72, 6.06, 4.52, and 3.15%, respectively.
Palavras-chave

Texto completo: 1 Base de dados: MEDLINE Idioma: En Ano de publicação: 2021 Tipo de documento: Article

Texto completo: 1 Base de dados: MEDLINE Idioma: En Ano de publicação: 2021 Tipo de documento: Article