Your browser doesn't support javascript.
loading
GMIM: Self-supervised pre-training for 3D medical image segmentation with adaptive and hierarchical masked image modeling.
Qi, Liangce; Jiang, Zhengang; Shi, Weili; Qu, Feng; Feng, Guanyuan.
Affiliation
  • Qi L; Department of Computer Science and Technology, Changchun University of Science and Technology, Changchun, 130022, Jilin, China. Electronic address: 1263591209@qq.com.
  • Jiang Z; Department of Computer Science and Technology, Changchun University of Science and Technology, Changchun, 130022, Jilin, China; Zhongshan Institute of Changchun University of Science and Technology, Zhongshan, 528400, Guangzhou, China. Electronic address: jiangzhengang@cust.edu.cn.
  • Shi W; Department of Computer Science and Technology, Changchun University of Science and Technology, Changchun, 130022, Jilin, China; Zhongshan Institute of Changchun University of Science and Technology, Zhongshan, 528400, Guangzhou, China.
  • Qu F; Department of Computer Science and Technology, Changchun University of Science and Technology, Changchun, 130022, Jilin, China.
  • Feng G; Department of Computer Science and Technology, Changchun University of Science and Technology, Changchun, 130022, Jilin, China.
Comput Biol Med ; 176: 108547, 2024 Jun.
Article in En | MEDLINE | ID: mdl-38728994
ABSTRACT
Self-supervised pre-training and fully supervised fine-tuning paradigms have received much attention to solve the data annotation problem in deep learning fields. Compared with traditional pre-training on large natural image datasets, medical self-supervised learning methods learn rich representations derived from unlabeled data itself thus avoiding the distribution shift between different image domains. However, nowadays state-of-the-art medical pre-training methods were specifically designed for downstream tasks making them less flexible and difficult to apply to new tasks. In this paper, we propose grid mask image modeling, a flexible and general self-supervised method to pre-train medical vision transformers for 3D medical image segmentation. Our goal is to guide networks to learn the correlations between organs and tissues by reconstructing original images based on partial observations. The relationships are consistent within the human body and invariant to disease type or imaging modality. To achieve this, we design a Siamese framework consisting of an online branch and a target branch. An adaptive and hierarchical masking strategy is employed in the online branch to (1) learn the boundaries or small contextual mutation regions within images; (2) to learn high-level semantic representations from deeper layers of the multiscale encoder. In addition, the target branch provides representations for contrastive learning to further reduce representation redundancy. We evaluate our method through segmentation performance on two public datasets. The experimental results demonstrate our method outperforms other self-supervised methods. Codes are available at https//github.com/mobiletomb/Gmim.
Subject(s)
Key words

Full text: 1 Collection: 01-internacional Database: MEDLINE Main subject: Imaging, Three-Dimensional Limits: Humans Language: En Journal: Comput Biol Med Year: 2024 Document type: Article Country of publication:

Full text: 1 Collection: 01-internacional Database: MEDLINE Main subject: Imaging, Three-Dimensional Limits: Humans Language: En Journal: Comput Biol Med Year: 2024 Document type: Article Country of publication: