Your browser doesn't support javascript.
loading
Deep convolutional neural networks for multi-modality isointense infant brain image segmentation.
Zhang, Wenlu; Li, Rongjian; Deng, Houtao; Wang, Li; Lin, Weili; Ji, Shuiwang; Shen, Dinggang.
Afiliação
  • Zhang W; Department of Computer Science, Old Dominion University, Norfolk, VA 23529, USA.
  • Li R; Department of Computer Science, Old Dominion University, Norfolk, VA 23529, USA.
  • Deng H; Instacart, San Francisco, CA 94107, USA.
  • Wang L; IDEA Lab, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, NC 27599, USA.
  • Lin W; MRI Lab, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, NC 27599, USA.
  • Ji S; Department of Computer Science, Old Dominion University, Norfolk, VA 23529, USA. Electronic address: sji@cs.odu.edu.
  • Shen D; IDEA Lab, Department of Radiology and BRIC, University of North Carolina at Chapel Hill, NC 27599, USA; Department of Brain and Cognitive Engineering, Korea University, Seoul, Republic of Korea. Electronic address: dgshen@med.unc.edu.
Neuroimage ; 108: 214-24, 2015 Mar.
Article em En | MEDLINE | ID: mdl-25562829
ABSTRACT
The segmentation of infant brain tissue images into white matter (WM), gray matter (GM), and cerebrospinal fluid (CSF) plays an important role in studying early brain development in health and disease. In the isointense stage (approximately 6-8 months of age), WM and GM exhibit similar levels of intensity in both T1 and T2 MR images, making the tissue segmentation very challenging. Only a small number of existing methods have been designed for tissue segmentation in this isointense stage; however, they only used a single T1 or T2 images, or the combination of T1 and T2 images. In this paper, we propose to use deep convolutional neural networks (CNNs) for segmenting isointense stage brain tissues using multi-modality MR images. CNNs are a type of deep models in which trainable filters and local neighborhood pooling operations are applied alternatingly on the raw input images, resulting in a hierarchy of increasingly complex features. Specifically, we used multi-modality information from T1, T2, and fractional anisotropy (FA) images as inputs and then generated the segmentation maps as outputs. The multiple intermediate layers applied convolution, pooling, normalization, and other operations to capture the highly nonlinear mappings between inputs and outputs. We compared the performance of our approach with that of the commonly used segmentation methods on a set of manually segmented isointense stage brain images. Results showed that our proposed model significantly outperformed prior methods on infant brain tissue segmentation. In addition, our results indicated that integration of multi-modality images led to significant performance improvement.
Assuntos
Palavras-chave

Texto completo: 1 Base de dados: MEDLINE Assunto principal: Processamento de Imagem Assistida por Computador / Encéfalo / Mapeamento Encefálico / Imageamento por Ressonância Magnética / Redes Neurais de Computação Limite: Humans / Infant Idioma: En Ano de publicação: 2015 Tipo de documento: Article

Texto completo: 1 Base de dados: MEDLINE Assunto principal: Processamento de Imagem Assistida por Computador / Encéfalo / Mapeamento Encefálico / Imageamento por Ressonância Magnética / Redes Neurais de Computação Limite: Humans / Infant Idioma: En Ano de publicação: 2015 Tipo de documento: Article