Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 20
Filter
Add more filters










Publication year range
2.
Jpn J Radiol ; 2024 Jun 10.
Article in English | MEDLINE | ID: mdl-38856878

ABSTRACT

Medicine and deep learning-based artificial intelligence (AI) engineering represent two distinct fields each with decades of published history. The current rapid convergence of deep learning and medicine has led to significant advancements, yet it has also introduced ambiguity regarding data set terms common to both fields, potentially leading to miscommunication and methodological discrepancies. This narrative review aims to give historical context for these terms, accentuate the importance of clarity when these terms are used in medical deep learning contexts, and offer solutions to mitigate misunderstandings by readers from either field. Through an examination of historical documents, including articles, writing guidelines, and textbooks, this review traces the divergent evolution of terms for data sets and their impact. Initially, the discordant interpretations of the word 'validation' in medical and AI contexts are explored. We then show that in the medical field as well, terms traditionally used in the deep learning domain are becoming more common, with the data for creating models referred to as the 'training set', the data for tuning of parameters referred to as the 'validation (or tuning) set', and the data for the evaluation of models as the 'test set'. Additionally, the test sets used for model evaluation are classified into internal (random splitting, cross-validation, and leave-one-out) sets and external (temporal and geographic) sets. This review then identifies often misunderstood terms and proposes pragmatic solutions to mitigate terminological confusion in the field of deep learning in medicine. We support the accurate and standardized description of these data sets and the explicit definition of data set splitting terminologies in each publication. These are crucial methods for demonstrating the robustness and generalizability of deep learning applications in medicine. This review aspires to enhance the precision of communication, thereby fostering more effective and transparent research methodologies in this interdisciplinary field.

4.
Diagn Interv Imaging ; 2024 Jun 24.
Article in English | MEDLINE | ID: mdl-38918123

ABSTRACT

The rapid advancement of artificial intelligence (AI) in healthcare has revolutionized the industry, offering significant improvements in diagnostic accuracy, efficiency, and patient outcomes. However, the increasing adoption of AI systems also raises concerns about their environmental impact, particularly in the context of climate change. This review explores the intersection of climate change and AI in healthcare, examining the challenges posed by the energy consumption and carbon footprint of AI systems, as well as the potential solutions to mitigate their environmental impact. The review highlights the energy-intensive nature of AI model training and deployment, the contribution of data centers to greenhouse gas emissions, and the generation of electronic waste. To address these challenges, the development of energy-efficient AI models, the adoption of green computing practices, and the integration of renewable energy sources are discussed as potential solutions. The review also emphasizes the role of AI in optimizing healthcare workflows, reducing resource waste, and facilitating sustainable practices such as telemedicine. Furthermore, the importance of policy and governance frameworks, global initiatives, and collaborative efforts in promoting sustainable AI practices in healthcare is explored. The review concludes by outlining best practices for sustainable AI deployment, including eco-design, lifecycle assessment, responsible data management, and continuous monitoring and improvement. As the healthcare industry continues to embrace AI technologies, prioritizing sustainability and environmental responsibility is crucial to ensure that the benefits of AI are realized while actively contributing to the preservation of our planet.

5.
Clin Neuroradiol ; 2024 May 28.
Article in English | MEDLINE | ID: mdl-38806794

ABSTRACT

PURPOSE: To compare the diagnostic performance among Generative Pre-trained Transformer (GPT)-4-based ChatGPT, GPT­4 with vision (GPT-4V) based ChatGPT, and radiologists in challenging neuroradiology cases. METHODS: We collected 32 consecutive "Freiburg Neuropathology Case Conference" cases from the journal Clinical Neuroradiology between March 2016 and December 2023. We input the medical history and imaging findings into GPT-4-based ChatGPT and the medical history and images into GPT-4V-based ChatGPT, then both generated a diagnosis for each case. Six radiologists (three radiology residents and three board-certified radiologists) independently reviewed all cases and provided diagnoses. ChatGPT and radiologists' diagnostic accuracy rates were evaluated based on the published ground truth. Chi-square tests were performed to compare the diagnostic accuracy of GPT-4-based ChatGPT, GPT-4V-based ChatGPT, and radiologists. RESULTS: GPT­4 and GPT-4V-based ChatGPTs achieved accuracy rates of 22% (7/32) and 16% (5/32), respectively. Radiologists achieved the following accuracy rates: three radiology residents 28% (9/32), 31% (10/32), and 28% (9/32); and three board-certified radiologists 38% (12/32), 47% (15/32), and 44% (14/32). GPT-4-based ChatGPT's diagnostic accuracy was lower than each radiologist, although not significantly (all p > 0.07). GPT-4V-based ChatGPT's diagnostic accuracy was also lower than each radiologist and significantly lower than two board-certified radiologists (p = 0.02 and 0.03) (not significant for radiology residents and one board-certified radiologist [all p > 0.09]). CONCLUSION: While GPT-4-based ChatGPT demonstrated relatively higher diagnostic performance than GPT-4V-based ChatGPT, the diagnostic performance of GPT­4 and GPT-4V-based ChatGPTs did not reach the performance level of either radiology residents or board-certified radiologists in challenging neuroradiology cases.

7.
AJNR Am J Neuroradiol ; 45(6): 826-832, 2024 Jun 07.
Article in English | MEDLINE | ID: mdl-38663993

ABSTRACT

BACKGROUND: Intermodality image-to-image translation is an artificial intelligence technique for generating one technique from another. PURPOSE: This review was designed to systematically identify and quantify biases and quality issues preventing validation and clinical application of artificial intelligence models for intermodality image-to-image translation of brain imaging. DATA SOURCES: PubMed, Scopus, and IEEE Xplore were searched through August 2, 2023, for artificial intelligence-based image translation models of radiologic brain images. STUDY SELECTION: This review collected 102 works published between April 2017 and August 2023. DATA ANALYSIS: Eligible studies were evaluated for quality using the Checklist for Artificial Intelligence in Medical Imaging (CLAIM) and for bias using the Prediction model Risk Of Bias ASsessment Tool (PROBAST). Medically-focused article adherence was compared with that of engineering-focused articles overall with the Mann-Whitney U test and for each criterion using the Fisher exact test. DATA SYNTHESIS: Median adherence to the relevant CLAIM criteria was 69% and 38% for PROBAST questions. CLAIM adherence was lower for engineering-focused articles compared with medically-focused articles (65% versus 73%, P < .001). Engineering-focused studies had higher adherence for model description criteria, and medically-focused studies had higher adherence for data set and evaluation descriptions. LIMITATIONS: Our review is limited by the study design and model heterogeneity. CONCLUSIONS: Nearly all studies revealed critical issues preventing clinical application, with engineering-focused studies showing higher adherence for the technical model description but significantly lower overall adherence than medically-focused studies. The pursuit of clinical application requires collaboration from both fields to improve reporting.


Subject(s)
Neuroimaging , Humans , Neuroimaging/methods , Neuroimaging/standards , Bias , Artificial Intelligence
8.
Sci Rep ; 14(1): 2911, 2024 02 05.
Article in English | MEDLINE | ID: mdl-38316892

ABSTRACT

This study created an image-to-image translation model that synthesizes diffusion tensor images (DTI) from conventional diffusion weighted images, and validated the similarities between the original and synthetic DTI. Thirty-two healthy volunteers were prospectively recruited. DTI and DWI were obtained with six and three directions of the motion probing gradient (MPG), respectively. The identical imaging plane was paired for the image-to-image translation model that synthesized one direction of the MPG from DWI. This process was repeated six times in the respective MPG directions. Regions of interest (ROIs) in the lentiform nucleus, thalamus, posterior limb of the internal capsule, posterior thalamic radiation, and splenium of the corpus callosum were created and applied to maps derived from the original and synthetic DTI. The mean values and signal-to-noise ratio (SNR) of the original and synthetic maps for each ROI were compared. The Bland-Altman plot between the original and synthetic data was evaluated. Although the test dataset showed a larger standard deviation of all values and lower SNR in the synthetic data than in the original data, the Bland-Altman plots showed each plot localizing in a similar distribution. Synthetic DTI could be generated from conventional DWI with an image-to-image translation model.


Subject(s)
Deep Learning , White Matter , Humans , Corpus Callosum/diagnostic imaging , Signal-To-Noise Ratio , Internal Capsule , Diffusion Magnetic Resonance Imaging/methods
9.
Neuroradiology ; 66(1): 73-79, 2024 Jan.
Article in English | MEDLINE | ID: mdl-37994939

ABSTRACT

PURPOSE: The noteworthy performance of Chat Generative Pre-trained Transformer (ChatGPT), an artificial intelligence text generation model based on the GPT-4 architecture, has been demonstrated in various fields; however, its potential applications in neuroradiology remain unexplored. This study aimed to evaluate the diagnostic performance of GPT-4 based ChatGPT in neuroradiology. METHODS: We collected 100 consecutive "Case of the Week" cases from the American Journal of Neuroradiology between October 2021 and September 2023. ChatGPT generated a diagnosis from patient's medical history and imaging findings for each case. Then the diagnostic accuracy rate was determined using the published ground truth. Each case was categorized by anatomical location (brain, spine, and head & neck), and brain cases were further divided into central nervous system (CNS) tumor and non-CNS tumor groups. Fisher's exact test was conducted to compare the accuracy rates among the three anatomical locations, as well as between the CNS tumor and non-CNS tumor groups. RESULTS: ChatGPT achieved a diagnostic accuracy rate of 50% (50/100 cases). There were no significant differences between the accuracy rates of the three anatomical locations (p = 0.89). The accuracy rate was significantly lower for the CNS tumor group compared to the non-CNS tumor group in the brain cases (16% [3/19] vs. 62% [36/58], p < 0.001). CONCLUSION: This study demonstrated the diagnostic performance of ChatGPT in neuroradiology. ChatGPT's diagnostic accuracy varied depending on disease etiologies, and its diagnostic accuracy was significantly lower in CNS tumors compared to non-CNS tumors.


Subject(s)
Artificial Intelligence , Neoplasms , Humans , Head , Brain , Neck
11.
Lancet Healthy Longev ; 4(9): e478-e486, 2023 09.
Article in English | MEDLINE | ID: mdl-37597530

ABSTRACT

BACKGROUND: Chest radiographs are widely available and cost-effective; however, their usefulness as a biomarker of ageing using multi-institutional data remains underexplored. The aim of this study was to develop a biomarker of ageing from chest radiography and examine the correlation between the biomarker and diseases. METHODS: In this retrospective, multi-institutional study, we trained, tuned, and externally tested an artificial intelligence (AI) model to estimate the age of healthy individuals using chest radiographs as a biomarker. For the biomarker modelling phase of the study, we used healthy chest radiographs consecutively collected between May 22, 2008, and Dec 28, 2021, from three institutions in Japan. Data from two institutions were used for training, tuning, and internal testing, and data from the third institution were used for external testing. To evaluate the performance of the AI model in estimating ages, we calculated the correlation coefficient, mean square error, root mean square error, and mean absolute error. The correlation investigation phase of the study included chest radiographs from individuals with a known disease that were consecutively collected between Jan 1, 2018, and Dec 31, 2021, from an additional two institutions in Japan. We investigated the odds ratios (ORs) for various diseases given the difference between the AI-estimated age and chronological age (ie, the difference-age). FINDINGS: We included 101 296 chest radiographs from 70 248 participants across five institutions. In the biomarker modelling phase, the external test dataset from 3467 healthy participants included 8046 radiographs. Between the AI-estimated age and chronological age, the correlation coefficient was 0·95 (99% CI 0·95-0·95), the mean square error was 15·0 years (99% CI 14·0-15·0), the root mean square error was 3·8 years (99% CI 3·8-3·9), and the mean absolute error was 3·0 years (99% CI 3·0-3·1). In the correlation investigation phase, the external test datasets from 34 197 participants with a known disease included 34 197 radiographs. The ORs for difference-age were as follows: 1·04 (99% CI 1·04-1·05) for hypertension; 1·02 (1·01-1·03) for hyperuricaemia; 1·05 (1·03-1·06) for chronic obstructive pulmonary disease; 1·08 (1·06-1·09) for interstitial lung disease; 1·05 (1·03-1·06) for chronic renal failure; 1·04 (1·03-1·06) for atrial fibrillation; 1·03 (1·02-1·04) for osteoporosis; and 1·05 (1·03-1·06) for liver cirrhosis. INTERPRETATION: The AI-estimated age using chest radiographs showed a strong correlation with chronological age in the healthy cohorts. Furthermore, in cohorts of individuals with known diseases, the difference between estimated age and chronological age correlated with various chronic diseases. The use of this biomarker might pave the way for enhanced risk stratification methodologies, individualised therapeutic interventions, and innovative early diagnostic and preventive approaches towards age-associated pathologies. FUNDING: None. TRANSLATION: For the Japanese translation of the abstract see Supplementary Materials section.


Subject(s)
Aging , Artificial Intelligence , Humans , Japan , Retrospective Studies , Biomarkers
12.
Radiology ; 308(2): e223016, 2023 08.
Article in English | MEDLINE | ID: mdl-37526545

ABSTRACT

Background Carbon 11 (11C)-methionine is a useful PET radiotracer for the management of patients with glioma, but radiation exposure and lack of molecular imaging facilities limit its use. Purpose To generate synthetic methionine PET images from contrast-enhanced (CE) MRI through an artificial intelligence (AI)-based image-to-image translation model and to compare its performance for grading and prognosis of gliomas with that of real PET. Materials and Methods An AI-based model to generate synthetic methionine PET images from CE MRI was developed and validated from patients who underwent both methionine PET and CE MRI at a university hospital from January 2007 to December 2018 (institutional data set). Pearson correlation coefficients for the maximum and mean tumor to background ratio (TBRmax and TBRmean, respectively) of methionine uptake and the lesion volume between synthetic and real PET were calculated. Two additional open-source glioma databases of preoperative CE MRI without methionine PET were used as the external test set. Using the TBRs, the area under the receiver operating characteristic curve (AUC) for classifying high-grade and low-grade gliomas and overall survival were evaluated. Results The institutional data set included 362 patients (mean age, 49 years ± 19 [SD]; 195 female, 167 male; training, n = 294; validation, n = 34; test, n = 34). In the internal test set, Pearson correlation coefficients were 0.68 (95% CI: 0.47, 0.81), 0.76 (95% CI: 0.59, 0.86), and 0.92 (95% CI: 0.85, 0.95) for TBRmax, TBRmean, and lesion volume, respectively. The external test set included 344 patients with gliomas (mean age, 53 years ± 15; 192 male, 152 female; high grade, n = 269). The AUC for TBRmax was 0.81 (95% CI: 0.75, 0.86) and the overall survival analysis showed a significant difference between the high (2-year survival rate, 27%) and low (2-year survival rate, 71%; P < .001) TBRmax groups. Conclusion The AI-based model-generated synthetic methionine PET images strongly correlated with real PET images and showed good performance for glioma grading and prognostication. Published under a CC BY 4.0 license. Supplemental material is available for this article.


Subject(s)
Brain Neoplasms , Glioma , Humans , Male , Female , Middle Aged , Methionine , Brain Neoplasms/diagnostic imaging , Brain Neoplasms/pathology , Artificial Intelligence , Positron-Emission Tomography/methods , Neoplasm Grading , Glioma/diagnostic imaging , Glioma/pathology , Magnetic Resonance Imaging/methods , Racemethionine
13.
Lancet Digit Health ; 5(8): e525-e533, 2023 08.
Article in English | MEDLINE | ID: mdl-37422342

ABSTRACT

BACKGROUND: Chest radiography is a common and widely available examination. Although cardiovascular structures-such as cardiac shadows and vessels-are visible on chest radiographs, the ability of these radiographs to estimate cardiac function and valvular disease is poorly understood. Using datasets from multiple institutions, we aimed to develop and validate a deep-learning model to simultaneously detect valvular disease and cardiac functions from chest radiographs. METHODS: In this model development and validation study, we trained, validated, and externally tested a deep learning-based model to classify left ventricular ejection fraction, tricuspid regurgitant velocity, mitral regurgitation, aortic stenosis, aortic regurgitation, mitral stenosis, tricuspid regurgitation, pulmonary regurgitation, and inferior vena cava dilation from chest radiographs. The chest radiographs and associated echocardiograms were collected from four institutions between April 1, 2013, and Dec 31, 2021: we used data from three sites (Osaka Metropolitan University Hospital, Osaka, Japan; Habikino Medical Center, Habikino, Japan; and Morimoto Hospital, Osaka, Japan) for training, validation, and internal testing, and data from one site (Kashiwara Municipal Hospital, Kashiwara, Japan) for external testing. We evaluated the area under the receiver operating characteristic curve (AUC), sensitivity, specificity, and accuracy. FINDINGS: We included 22 551 radiographs associated with 22 551 echocardiograms obtained from 16 946 patients. The external test dataset featured 3311 radiographs from 2617 patients with a mean age of 72 years [SD 15], of whom 49·8% were male and 50·2% were female. The AUCs, accuracy, sensitivity, and specificity for this dataset were 0·92 (95% CI 0·90-0·95), 86% (85-87), 82% (75-87), and 86% (85-88) for classifying the left ventricular ejection fraction at a 40% cutoff, 0·85 (0·83-0·87), 75% (73-76), 83% (80-87), and 73% (71-75) for classifying the tricuspid regurgitant velocity at a 2·8 m/s cutoff, 0·89 (0·86-0·92), 85% (84-86), 82% (76-87), and 85% (84-86) for classifying mitral regurgitation at the none-mild versus moderate-severe cutoff, 0·83 (0·78-0·88), 73% (71-74), 79% (69-87), and 72% (71-74) for classifying aortic stenosis, 0·83 (0·79-0·87), 68% (67-70), 88% (81-92), and 67% (66-69) for classifying aortic regurgitation, 0·86 (0·67-1·00), 90% (89-91), 83% (36-100), and 90% (89-91) for classifying mitral stenosis, 0·92 (0·89-0·94), 83% (82-85), 87% (83-91), and 83% (82-84) for classifying tricuspid regurgitation, 0·86 (0·82-0·90), 69% (68-71), 91% (84-95), and 68% (67-70) for classifying pulmonary regurgitation, and 0·85 (0·81-0·89), 86% (85-88), 73% (65-81), and 87% (86-88) for classifying inferior vena cava dilation. INTERPRETATION: The deep learning-based model can accurately classify cardiac functions and valvular heart diseases using information from digital chest radiographs. This model can classify values typically obtained from echocardiography in a fraction of the time, with low system requirements and the potential to be continuously available in areas where echocardiography specialists are scarce or absent. FUNDING: None.


Subject(s)
Heart Valve Diseases , Mitral Valve Insufficiency , Humans , Male , Female , Aged , Retrospective Studies , Artificial Intelligence , Stroke Volume , Ventricular Function, Left , Heart Valve Diseases/complications , Heart Valve Diseases/diagnosis , Mitral Valve Insufficiency/complications , Mitral Valve Insufficiency/diagnostic imaging , Radiography
15.
Eur Respir Rev ; 32(168)2023 Jun 30.
Article in English | MEDLINE | ID: mdl-37286217

ABSTRACT

BACKGROUND: Deep learning (DL), a subset of artificial intelligence (AI), has been applied to pneumothorax diagnosis to aid physician diagnosis, but no meta-analysis has been performed. METHODS: A search of multiple electronic databases through September 2022 was performed to identify studies that applied DL for pneumothorax diagnosis using imaging. Meta-analysis via a hierarchical model to calculate the summary area under the curve (AUC) and pooled sensitivity and specificity for both DL and physicians was performed. Risk of bias was assessed using a modified Prediction Model Study Risk of Bias Assessment Tool. RESULTS: In 56 of the 63 primary studies, pneumothorax was identified from chest radiography. The total AUC was 0.97 (95% CI 0.96-0.98) for both DL and physicians. The total pooled sensitivity was 84% (95% CI 79-89%) for DL and 85% (95% CI 73-92%) for physicians and the pooled specificity was 96% (95% CI 94-98%) for DL and 98% (95% CI 95-99%) for physicians. More than half of the original studies (57%) had a high risk of bias. CONCLUSIONS: Our review found the diagnostic performance of DL models was similar to that of physicians, although the majority of studies had a high risk of bias. Further pneumothorax AI research is needed.


Subject(s)
Deep Learning , Pneumothorax , Humans , Pneumothorax/diagnostic imaging , Artificial Intelligence , Sensitivity and Specificity , Diagnostic Imaging
16.
Br J Radiol ; 95(1140): 20220058, 2022 Dec 01.
Article in English | MEDLINE | ID: mdl-36193755

ABSTRACT

OBJECTIVES: The purpose of this study was to develop an artificial intelligence-based model to prognosticate COVID-19 patients at admission by combining clinical data and chest radiographs. METHODS: This retrospective study used the Stony Brook University COVID-19 dataset of 1384 inpatients. After exclusions, 1356 patients were randomly divided into training (1083) and test datasets (273). We implemented three artificial intelligence models, which classified mortality, ICU admission, or ventilation risk. Each model had three submodels with different inputs: clinical data, chest radiographs, and both. We showed the importance of the variables using SHapley Additive exPlanations (SHAP) values. RESULTS: The mortality prediction model was best overall with area under the curve, sensitivity, specificity, and accuracy of 0.79 (0.72-0.86), 0.74 (0.68-0.79), 0.77 (0.61-0.88), and 0.74 (0.69-0.79) for the clinical data-based model; 0.77 (0.69-0.85), 0.67 (0.61-0.73), 0.81 (0.67-0.92), 0.70 (0.64-0.75) for the image-based model, and 0.86 (0.81-0.91), 0.76 (0.70-0.81), 0.77 (0.61-0.88), 0.76 (0.70-0.81) for the mixed model. The mixed model had the best performance (p value < 0.05). The radiographs ranked fourth for prognostication overall, and first of the inpatient tests assessed. CONCLUSIONS: These results suggest that prognosis models become more accurate if AI-derived chest radiograph features and clinical data are used together. ADVANCES IN KNOWLEDGE: This AI model evaluates chest radiographs together with clinical data in order to classify patients as having high or low mortality risk. This work shows that chest radiographs taken at admission have significant COVID-19 prognostic information compared to clinical data other than age and sex.


Subject(s)
COVID-19 , Humans , COVID-19/diagnostic imaging , Artificial Intelligence , Retrospective Studies , Radiography , Prognosis
17.
Radiol Artif Intell ; 4(2): e210221, 2022 Mar.
Article in English | MEDLINE | ID: mdl-35391769

ABSTRACT

Purpose: To develop an artificial intelligence-based model to detect mitral regurgitation on chest radiographs. Materials and Methods: This retrospective study included echocardiographs and associated chest radiographs consecutively collected at a single institution between July 2016 and May 2019. Associated radiographs were those obtained within 30 days of echocardiography. These radiographs were labeled as positive or negative for mitral regurgitation on the basis of the echocardiographic reports and were divided into training, validation, and test datasets. An artificial intelligence model was developed by using the training dataset and was tuned by using the validation dataset. To evaluate the model, the area under the curve, sensitivity, specificity, accuracy, positive predictive value, and negative predictive value were assessed by using the test dataset. Results: This study included a total of 10 367 images from 5270 patients. The training dataset included 8240 images (4216 patients), the validation dataset included 1073 images (527 patients), and the test dataset included 1054 images (527 patients). The area under the curve, sensitivity, specificity, accuracy, positive predictive value, and negative predictive value in the test dataset were 0.80 (95% CI: 0.77, 0.82), 71% (95% CI: 67, 75), 74% (95% CI: 70, 77), 73% (95% CI: 70, 75), 68% (95% CI: 64, 72), and 77% (95% CI: 73, 80), respectively. Conclusion: The developed deep learning-based artificial intelligence model may possibly differentiate patients with and without mitral regurgitation by using chest radiographs.Keywords: Computer-aided Diagnosis (CAD), Cardiac, Heart, Valves, Supervised Learning, Convolutional Neural Network (CNN), Deep Learning Algorithms, Machine Learning Algorithms Supplemental material is available for this article. © RSNA, 2022.

18.
Eur Radiol ; 32(9): 5890-5897, 2022 Sep.
Article in English | MEDLINE | ID: mdl-35357542

ABSTRACT

OBJECTIVE: The purpose of this study was to develop an artificial intelligence (AI)-based model to detect features of atrial fibrillation (AF) on chest radiographs. METHODS: This retrospective study included consecutively collected chest radiographs of patients who had echocardiography at our institution from July 2016 to May 2019. Eligible radiographs had been acquired within 30 days of the echocardiography. These radiographs were labeled as AF-positive or AF-negative based on the associated electronic medical records; then, each patient was randomly divided into training, validation, and test datasets in an 8:1:1 ratio. A deep learning-based model to classify radiographs as with or without AF was trained on the training dataset, tuned with the validation dataset, and evaluated with the test dataset. RESULTS: The training dataset included 11,105 images (5637 patients; 3145 male, mean age ± standard deviation, 68 ± 14 years), the validation dataset included 1388 images (704 patients, 397 male, 67 ± 14 years), and the test dataset included 1375 images (706 patients, 395 male, 68 ± 15 years). Applying the model to the validation and test datasets gave a respective area under the curve of 0.81 (95% confidence interval, 0.78-0.85) and 0.80 (0.76-0.84), sensitivity of 0.76 (0.70-0.81) and 0.70 (0.64-0.76), specificity of 0.75 (0.72-0.77) and 0.74 (0.72-0.77), and accuracy of 0.75 (0.72-0.77) and 0.74 (0.71-0.76). CONCLUSION: Our AI can identify AF on chest radiographs, which provides a new way for radiologists to infer AF. KEY POINTS: • A deep learning-based model was trained to detect atrial fibrillation in chest radiographs, showing that there are indicators of atrial fibrillation visible even on static images. • The validation and test datasets each gave a solid performance with area under the curve, sensitivity, and specificity of 0.81, 0.76, and 0.75, respectively, for the validation dataset, and 0.80, 0.70, and 0.74, respectively, for the test dataset. • The saliency maps highlighted anatomical areas consistent with those reported for atrial fibrillation on chest radiographs, such as the atria.


Subject(s)
Artificial Intelligence , Atrial Fibrillation , Deep Learning , Aged , Aged, 80 and over , Atrial Fibrillation/diagnostic imaging , Female , Humans , Male , Middle Aged , Radiography , Radiography, Thoracic/methods , Retrospective Studies
19.
Eur Heart J Digit Health ; 3(1): 20-28, 2022 Mar.
Article in English | MEDLINE | ID: mdl-36713993

ABSTRACT

Aims: We aimed to develop models to detect aortic stenosis (AS) from chest radiographs-one of the most basic imaging tests-with artificial intelligence. Methods and results: We used 10 433 retrospectively collected digital chest radiographs from 5638 patients to train, validate, and test three deep learning models. Chest radiographs were collected from patients who had also undergone echocardiography at a single institution between July 2016 and May 2019. These were labelled from the corresponding echocardiography assessments as AS-positive or AS-negative. The radiographs were separated on a patient basis into training [8327 images from 4512 patients, mean age 65 ± (standard deviation) 15 years], validation (1041 images from 563 patients, mean age 65 ± 14 years), and test (1065 images from 563 patients, mean age 65 ± 14 years) datasets. The soft voting-based ensemble of the three developed models had the best overall performance for predicting AS with an area under the receiver operating characteristic curve, sensitivity, specificity, accuracy, positive predictive value, and negative predictive value of 0.83 (95% confidence interval 0.77-0.88), 0.78 (0.67-0.86), 0.71 (0.68-0.73), 0.71 (0.68-0.74), 0.18 (0.14-0.23), and 0.97 (0.96-0.98), respectively, in the validation dataset and 0.83 (0.78-0.88), 0.83 (0.74-0.90), 0.69 (0.66-0.72), 0.71 (0.68-0.73), 0.23 (0.19-0.28), and 0.97 (0.96-0.98), respectively, in the test dataset. Conclusion: Deep learning models using chest radiographs have the potential to differentiate between radiographs of patients with and without AS. Lay Summary: We created artificial intelligence (AI) models using deep learning to identify aortic stenosis (AS) from chest radiographs. Three AI models were developed and evaluated with 10 433 retrospectively collected radiographs and labelled from echocardiography reports. The ensemble AI model could detect AS in a test dataset with an area under the receiver operating characteristic curve of 0.83 (95% confidence interval 0.78-0.88). Since chest radiography is a cost-effective and widely available imaging test, our model can provide an additive resource for the detection of AS.

20.
Radiology ; 299(3): 675-681, 2021 06.
Article in English | MEDLINE | ID: mdl-33787336

ABSTRACT

Background Digital subtraction angiography (DSA) generates an image by subtracting a mask image from a dynamic angiogram. However, patient movement-caused misregistration artifacts can result in unclear DSA images that interrupt procedures. Purpose To train and to validate a deep learning (DL)-based model to produce DSA-like cerebral angiograms directly from dynamic angiograms and then quantitatively and visually evaluate these angiograms for clinical usefulness. Materials and Methods A retrospective model development and validation study was conducted on dynamic and DSA image pairs consecutively collected from January 2019 through April 2019. Angiograms showing misregistration were first separated per patient by two radiologists and sorted into the misregistration test data set. Nonmisregistration angiograms were divided into development and external test data sets at a ratio of 8:1 per patient. The development data set was divided into training and validation data sets at ratio of 3:1 per patient. The DL model was created by using the training data set, tuned with the validation data set, and then evaluated quantitatively with the external test data set and visually with the misregistration test data set. Quantitative evaluations used the peak signal-to-noise ratio (PSNR) and the structural similarity (SSIM) with mixed liner models. Visual evaluation was conducted by using a numerical rating scale. Results The training, validation, nonmisregistration test, and misregistration test data sets included 10 751, 2784, 1346, and 711 paired images collected from 40 patients (mean age, 62 years ± 11 [standard deviation]; 33 women). In the quantitative evaluation, DL-generated angiograms showed a mean PSNR value of 40.2 dB ± 4.05 and a mean SSIM value of 0.97 ± 0.02, indicating high coincidence with the paired DSA images. In the visual evaluation, the median ratings of the DL-generated angiograms were similar to or better than those of the original DSA images for all 24 sequences. Conclusion The deep learning-based model provided clinically useful cerebral angiograms free from clinically significant artifacts directly from dynamic angiograms. Published under a CC BY 4.0 license. Supplemental material is available for this article.


Subject(s)
Cerebral Angiography , Deep Learning , Image Enhancement/methods , Adult , Aged , Aged, 80 and over , Angiography, Digital Subtraction , Artifacts , Female , Humans , Image Processing, Computer-Assisted/methods , Male , Middle Aged , Retrospective Studies , Signal-To-Noise Ratio
SELECTION OF CITATIONS
SEARCH DETAIL
...