Your browser doesn't support javascript.
loading
Application and visualization study of an intelligence-assisted classification model for common eye diseases using B-mode ultrasound images.
Zhu, Shaojun; Liu, Xiangjun; Lu, Ying; Zheng, Bo; Wu, Maonian; Yao, Xue; Yang, Weihua; Gong, Yan.
Afiliação
  • Zhu S; School of Information Engineering, Huzhou University, Huzhou, China.
  • Liu X; School of Information Engineering, Huzhou University, Huzhou, China.
  • Lu Y; School of Information Engineering, Huzhou University, Huzhou, China.
  • Zheng B; School of Information Engineering, Huzhou University, Huzhou, China.
  • Wu M; School of Information Engineering, Huzhou University, Huzhou, China.
  • Yao X; Shenzhen Eye Institute, Shenzhen Eye Hospital, Jinan University, Shenzhen, China.
  • Yang W; Shenzhen Eye Institute, Shenzhen Eye Hospital, Jinan University, Shenzhen, China.
  • Gong Y; Department of Ophthalmology, Ningbo Eye Hospital, Wenzhou Medical University, Ningbo, China.
Front Neurosci ; 18: 1339075, 2024.
Article em En | MEDLINE | ID: mdl-38808029
ABSTRACT

Aim:

Conventional approaches to diagnosing common eye diseases using B-mode ultrasonography are labor-intensive and time-consuming, must requiring expert intervention for accuracy. This study aims to address these challenges by proposing an intelligence-assisted analysis five-classification model for diagnosing common eye diseases using B-mode ultrasound images.

Methods:

This research utilizes 2064 B-mode ultrasound images of the eye to train a novel model integrating artificial intelligence technology.

Results:

The ConvNeXt-L model achieved outstanding performance with an accuracy rate of 84.3% and a Kappa value of 80.3%. Across five classifications (no obvious abnormality, vitreous opacity, posterior vitreous detachment, retinal detachment, and choroidal detachment), the model demonstrated sensitivity values of 93.2%, 67.6%, 86.1%, 89.4%, and 81.4%, respectively, and specificity values ranging from 94.6% to 98.1%. F1 scores ranged from 71% to 92%, while AUC values ranged from 89.7% to 97.8%.

Conclusion:

Among various models compared, the ConvNeXt-L model exhibited superior performance. It effectively categorizes and visualizes pathological changes, providing essential assisted information for ophthalmologists and enhancing diagnostic accuracy and efficiency.
Palavras-chave

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Idioma: En Revista: Front Neurosci Ano de publicação: 2024 Tipo de documento: Article País de afiliação: China

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Idioma: En Revista: Front Neurosci Ano de publicação: 2024 Tipo de documento: Article País de afiliação: China