Your browser doesn't support javascript.
loading
BTMF-GAN: A multi-modal MRI fusion generative adversarial network for brain tumors.
Liu, Xiao; Chen, Hongyi; Yao, Chong; Xiang, Rui; Zhou, Kun; Du, Peng; Liu, Weifan; Liu, Jie; Yu, Zekuan.
  • Liu X; School of Computer and Information Technology, Beijing Jiaotong University, Beijing, 100044, China.
  • Chen H; Academy for Engineering and Technology, Fudan University, Shanghai, 200433, China.
  • Yao C; College of Information and Electrical Engineering, China Agricultural University, Beijing, 100083, China.
  • Xiang R; Academy for Engineering and Technology, Fudan University, Shanghai, 200433, China.
  • Zhou K; Academy for Engineering and Technology, Fudan University, Shanghai, 200433, China.
  • Du P; Department of Radiology, Huashan Hospital, Fudan University, Shanghai, 200040, China.
  • Liu W; College of Science, Beijing Forestry University, Beijing, 100083, China. Electronic address: weifanliu@bjfu.edu.cn.
  • Liu J; School of Computer and Information Technology, Beijing Jiaotong University, Beijing, 100044, China. Electronic address: jieliu@bjtu.edu.cn.
  • Yu Z; Academy for Engineering and Technology, Fudan University, Shanghai, 200433, China.
Comput Biol Med ; 157: 106769, 2023 05.
Article en En | MEDLINE | ID: mdl-36947904
ABSTRACT
Image fusion techniques have been widely used for multi-modal medical image fusion tasks. Most existing methods aim to improve the overall quality of the fused image and do not focus on the more important textural details and contrast between the tissues of the lesion in the regions of interest (ROIs). This can lead to the distortion of important tumor ROIs information and thus limits the applicability of the fused images in clinical practice. To improve the fusion quality of ROIs relevant to medical implications, we propose a multi-modal MRI fusion generative adversarial network (BTMF-GAN) for the task of multi-modal MRI fusion of brain tumors. Unlike existing deep learning approaches which focus on improving the global quality of the fused image, the proposed BTMF-GAN aims to achieve a balance between tissue details and structural contrasts in brain tumor, which is the region of interest crucial to many medical applications. Specifically, we employ a generator with a U-shaped nested structure and residual U-blocks (RSU) to enhance multi-scale feature extraction. To enhance and recalibrate features of the encoder, the multi-perceptual field adaptive transformer feature enhancement module (MRF-ATFE) is used between the encoder and the decoder instead of a skip connection. To increase contrast between tumor tissues of the fused image, a mask-part block is introduced to fragment the source image and the fused image, based on which, we propose a novel salient loss function. Qualitative and quantitative analysis of the results on the public and clinical datasets demonstrate the superiority of the proposed approach to many other commonly used fusion methods.
Asunto(s)
Palabras clave

Texto completo: 1 Banco de datos: MEDLINE Asunto principal: Neoplasias Encefálicas Tipo de estudio: Qualitative_research Límite: Humans Idioma: En Año: 2023 Tipo del documento: Article

Texto completo: 1 Banco de datos: MEDLINE Asunto principal: Neoplasias Encefálicas Tipo de estudio: Qualitative_research Límite: Humans Idioma: En Año: 2023 Tipo del documento: Article