Your browser doesn't support javascript.
loading
Deep Transfer Learning for the Multilabel Classification of Chest X-ray Images.
Huang, Guan-Hua; Fu, Qi-Jia; Gu, Ming-Zhang; Lu, Nan-Han; Liu, Kuo-Ying; Chen, Tai-Been.
Affiliation
  • Huang GH; Institute of Statistics, National Yang Ming Chiao Tung University, Hsinchu 30010, Taiwan.
  • Fu QJ; Institute of Statistics, National Yang Ming Chiao Tung University, Hsinchu 30010, Taiwan.
  • Gu MZ; Institute of Statistics, National Yang Ming Chiao Tung University, Hsinchu 30010, Taiwan.
  • Lu NH; Department of Pharmacy, Tajen University, Pingtung City 90741, Taiwan.
  • Liu KY; Department of Radiology, E-Da Hospital, I-Shou University, Kaohsiung City 82445, Taiwan.
  • Chen TB; Department of Medical Imaging and Radiological Science, I-Shou University, Kaohsiung City 82445, Taiwan.
Diagnostics (Basel) ; 12(6)2022 Jun 13.
Article in En | MEDLINE | ID: mdl-35741267
ABSTRACT
Chest X-ray (CXR) is widely used to diagnose conditions affecting the chest, its contents, and its nearby structures. In this study, we used a private data set containing 1630 CXR images with disease labels; most of the images were disease-free, but the others contained multiple sites of abnormalities. Here, we used deep convolutional neural network (CNN) models to extract feature representations and to identify possible diseases in these images. We also used transfer learning combined with large open-source image data sets to resolve the problems of insufficient training data and optimize the classification model. The effects of different approaches of reusing pretrained weights (model finetuning and layer transfer), source data sets of different sizes and similarity levels to the target data (ImageNet, ChestX-ray, and CheXpert), methods integrating source data sets into transfer learning (initiating, concatenating, and co-training), and backbone CNN models (ResNet50 and DenseNet121) on transfer learning were also assessed. The results demonstrated that transfer learning applied with the model finetuning approach typically afforded better prediction models. When only one source data set was adopted, ChestX-ray performed better than CheXpert; however, after ImageNet initials were attached, CheXpert performed better. ResNet50 performed better in initiating transfer learning, whereas DenseNet121 performed better in concatenating and co-training transfer learning. Transfer learning with multiple source data sets was preferable to that with a source data set. Overall, transfer learning can further enhance prediction capabilities and reduce computing costs for CXR images.
Key words

Full text: 1 Collection: 01-internacional Database: MEDLINE Type of study: Prognostic_studies Language: En Journal: Diagnostics (Basel) Year: 2022 Document type: Article Affiliation country: Taiwán

Full text: 1 Collection: 01-internacional Database: MEDLINE Type of study: Prognostic_studies Language: En Journal: Diagnostics (Basel) Year: 2022 Document type: Article Affiliation country: Taiwán