Your browser doesn't support javascript.
loading
Transformer-based Deep Neural Network for Breast Cancer Classification on Digital Breast Tomosynthesis Images.
Lee, Weonsuk; Lee, Hyeonsoo; Lee, Hyunjae; Park, Eun Kyung; Nam, Hyeonseob; Kooi, Thijs.
Afiliación
  • Lee W; From Lunit, 5F, 374 Gangnam-daero, Gangnam-gu, Seoul 06241, Republic of Korea.
  • Lee H; From Lunit, 5F, 374 Gangnam-daero, Gangnam-gu, Seoul 06241, Republic of Korea.
  • Lee H; From Lunit, 5F, 374 Gangnam-daero, Gangnam-gu, Seoul 06241, Republic of Korea.
  • Park EK; From Lunit, 5F, 374 Gangnam-daero, Gangnam-gu, Seoul 06241, Republic of Korea.
  • Nam H; From Lunit, 5F, 374 Gangnam-daero, Gangnam-gu, Seoul 06241, Republic of Korea.
  • Kooi T; From Lunit, 5F, 374 Gangnam-daero, Gangnam-gu, Seoul 06241, Republic of Korea.
Radiol Artif Intell ; 5(3): e220159, 2023 May.
Article en En | MEDLINE | ID: mdl-37293346
ABSTRACT

Purpose:

To develop an efficient deep neural network model that incorporates context from neighboring image sections to detect breast cancer on digital breast tomosynthesis (DBT) images. Materials and

Methods:

The authors adopted a transformer architecture that analyzes neighboring sections of the DBT stack. The proposed method was compared with two baselines an architecture based on three-dimensional (3D) convolutions and a two-dimensional model that analyzes each section individually. The models were trained with 5174 four-view DBT studies, validated with 1000 four-view DBT studies, and tested on 655 four-view DBT studies, which were retrospectively collected from nine institutions in the United States through an external entity. Methods were compared using area under the receiver operating characteristic curve (AUC), sensitivity at a fixed specificity, and specificity at a fixed sensitivity.

Results:

On the test set of 655 DBT studies, both 3D models showed higher classification performance than did the per-section baseline model. The proposed transformer-based model showed a significant increase in AUC (0.88 vs 0.91, P = .002), sensitivity (81.0% vs 87.7%, P = .006), and specificity (80.5% vs 86.4%, P < .001) at clinically relevant operating points when compared with the single-DBT-section baseline. The transformer-based model used only 25% of the number of floating-point operations per second used by the 3D convolution model while demonstrating similar classification performance.

Conclusion:

A transformer-based deep neural network using data from neighboring sections improved breast cancer classification performance compared with a per-section baseline model and was more efficient than a model using 3D convolutions.Keywords Breast, Tomosynthesis, Diagnosis, Supervised Learning, Convolutional Neural Network (CNN), Digital Breast Tomosynthesis, Breast Cancer, Deep Neural Networks, Transformers Supplemental material is available for this article. © RSNA, 2023.
Palabras clave

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Tipo de estudio: Prognostic_studies Idioma: En Revista: Radiol Artif Intell Año: 2023 Tipo del documento: Article

Texto completo: 1 Colección: 01-internacional Base de datos: MEDLINE Tipo de estudio: Prognostic_studies Idioma: En Revista: Radiol Artif Intell Año: 2023 Tipo del documento: Article