Your browser doesn't support javascript.
loading
Optimizing Vision Transformers for Histopathology: Pretraining and Normalization in Breast Cancer Classification.
Baroni, Giulia Lucrezia; Rasotto, Laura; Roitero, Kevin; Tulisso, Angelica; Di Loreto, Carla; Della Mea, Vincenzo.
Afiliação
  • Baroni GL; Department of Mathematics, Computer Science and Physics, University of Udine, 33100 Udine, Italy.
  • Rasotto L; Department of Mathematics, Computer Science and Physics, University of Udine, 33100 Udine, Italy.
  • Roitero K; Department of Mathematics, Computer Science and Physics, University of Udine, 33100 Udine, Italy.
  • Tulisso A; Istituto di Anatomia Patologica, Azienda Sanitaria Universitaria Friuli Centrale, 33100 Udine, Italy.
  • Di Loreto C; Istituto di Anatomia Patologica, Azienda Sanitaria Universitaria Friuli Centrale, 33100 Udine, Italy.
  • Della Mea V; Department of Mathematics, Computer Science and Physics, University of Udine, 33100 Udine, Italy.
J Imaging ; 10(5)2024 Apr 30.
Article em En | MEDLINE | ID: mdl-38786562
ABSTRACT
This paper introduces a self-attention Vision Transformer model specifically developed for classifying breast cancer in histology images. We examine various training strategies and configurations, including pretraining, dimension resizing, data augmentation and color normalization strategies, patch overlap, and patch size configurations, in order to evaluate their impact on the effectiveness of the histology image classification. Additionally, we provide evidence for the increase in effectiveness gathered through geometric and color data augmentation techniques. We primarily utilize the BACH dataset to train and validate our methods and models, but we also test them on two additional datasets, BRACS and AIDPATH, to verify their generalization capabilities. Our model, developed from a transformer pretrained on ImageNet, achieves an accuracy rate of 0.91 on the BACH dataset, 0.74 on the BRACS dataset, and 0.92 on the AIDPATH dataset. Using a model based on the prostate small and prostate medium HistoEncoder models, we achieve accuracy rates of 0.89 and 0.86, respectively. Our results suggest that pretraining on large-scale general datasets like ImageNet is advantageous. We also show the potential benefits of using domain-specific pretraining datasets, such as extensive histopathological image collections as in HistoEncoder, though not yet with clear advantages.
Palavras-chave

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Idioma: En Revista: J Imaging Ano de publicação: 2024 Tipo de documento: Article

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Idioma: En Revista: J Imaging Ano de publicação: 2024 Tipo de documento: Article