Your browser doesn't support javascript.
loading
Lightweight Cross-Modal Information Mutual Reinforcement Network for RGB-T Salient Object Detection.
Lv, Chengtao; Wan, Bin; Zhou, Xiaofei; Sun, Yaoqi; Zhang, Jiyong; Yan, Chenggang.
Afiliação
  • Lv C; School of Automation, Hangzhou Dianzi University, Hangzhou 310018, China.
  • Wan B; School of Automation, Hangzhou Dianzi University, Hangzhou 310018, China.
  • Zhou X; School of Automation, Hangzhou Dianzi University, Hangzhou 310018, China.
  • Sun Y; School of Automation, Hangzhou Dianzi University, Hangzhou 310018, China.
  • Zhang J; Lishui Institute, Hangzhou Dianzi University, Lishui 323000, China.
  • Yan C; School of Automation, Hangzhou Dianzi University, Hangzhou 310018, China.
Entropy (Basel) ; 26(2)2024 Jan 31.
Article em En | MEDLINE | ID: mdl-38392385
ABSTRACT
RGB-T salient object detection (SOD) has made significant progress in recent years. However, most existing works are based on heavy models, which are not applicable to mobile devices. Additionally, there is still room for improvement in the design of cross-modal feature fusion and cross-level feature fusion. To address these issues, we propose a lightweight cross-modal information mutual reinforcement network for RGB-T SOD. Our network consists of a lightweight encoder, the cross-modal information mutual reinforcement (CMIMR) module, and the semantic-information-guided fusion (SIGF) module. To reduce the computational cost and the number of parameters, we employ the lightweight module in both the encoder and decoder. Furthermore, to fuse the complementary information between two-modal features, we design the CMIMR module to enhance the two-modal features. This module effectively refines the two-modal features by absorbing previous-level semantic information and inter-modal complementary information. In addition, to fuse the cross-level feature and detect multiscale salient objects, we design the SIGF module, which effectively suppresses the background noisy information in low-level features and extracts multiscale information. We conduct extensive experiments on three RGB-T datasets, and our method achieves competitive performance compared to the other 15 state-of-the-art methods.
Palavras-chave

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Idioma: En Revista: Entropy (Basel) Ano de publicação: 2024 Tipo de documento: Article País de afiliação: China

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Idioma: En Revista: Entropy (Basel) Ano de publicação: 2024 Tipo de documento: Article País de afiliação: China