Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 27
Filter
1.
Radiology ; 310(1): e230981, 2024 Jan.
Article in English | MEDLINE | ID: mdl-38193833

ABSTRACT

Background Multiple commercial artificial intelligence (AI) products exist for assessing radiographs; however, comparable performance data for these algorithms are limited. Purpose To perform an independent, stand-alone validation of commercially available AI products for bone age prediction based on hand radiographs and lung nodule detection on chest radiographs. Materials and Methods This retrospective study was carried out as part of Project AIR. Nine of 17 eligible AI products were validated on data from seven Dutch hospitals. For bone age prediction, the root mean square error (RMSE) and Pearson correlation coefficient were computed. The reference standard was set by three to five expert readers. For lung nodule detection, the area under the receiver operating characteristic curve (AUC) was computed. The reference standard was set by a chest radiologist based on CT. Randomized subsets of hand (n = 95) and chest (n = 140) radiographs were read by 14 and 17 human readers, respectively, with varying experience. Results Two bone age prediction algorithms were tested on hand radiographs (from January 2017 to January 2022) in 326 patients (mean age, 10 years ± 4 [SD]; 173 female patients) and correlated strongly with the reference standard (r = 0.99; P < .001 for both). No difference in RMSE was observed between algorithms (0.63 years [95% CI: 0.58, 0.69] and 0.57 years [95% CI: 0.52, 0.61]) and readers (0.68 years [95% CI: 0.64, 0.73]). Seven lung nodule detection algorithms were validated on chest radiographs (from January 2012 to May 2022) in 386 patients (mean age, 64 years ± 11; 223 male patients). Compared with readers (mean AUC, 0.81 [95% CI: 0.77, 0.85]), four algorithms performed better (AUC range, 0.86-0.93; P value range, <.001 to .04). Conclusions Compared with human readers, four AI algorithms for detecting lung nodules on chest radiographs showed improved performance, whereas the remaining algorithms tested showed no evidence of a difference in performance. © RSNA, 2024 Supplemental material is available for this article. See also the editorial by Omoumi and Richiardi in this issue.


Subject(s)
Artificial Intelligence , Software , Humans , Female , Male , Child , Middle Aged , Retrospective Studies , Algorithms , Lung
2.
Eur Radiol ; 34(1): 348-354, 2024 Jan.
Article in English | MEDLINE | ID: mdl-37515632

ABSTRACT

OBJECTIVES: To map the clinical use of CE-marked artificial intelligence (AI)-based software in radiology departments in the Netherlands (n = 69) between 2020 and 2022. MATERIALS AND METHODS: Our AI network (one radiologist or AI representative per Dutch hospital organization) received a questionnaire each spring from 2020 to 2022 about AI product usage, financing, and obstacles to adoption. Products that were not listed on www.AIforRadiology.com by July 2022 were excluded from the analysis. RESULTS: The number of respondents was 43 in 2020, 36 in 2021, and 33 in 2022. The number of departments using AI has been growing steadily (2020: 14, 2021: 19, 2022: 23). The diversity (2020: 7, 2021: 18, 2022: 34) and the number of total implementations (2020: 19, 2021: 38, 2022: 68) has rapidly increased. Seven implementations were discontinued in 2022. Four hospital organizations said to use an AI platform or marketplace for the deployment of AI solutions. AI is mostly used to support chest CT (17), neuro CT (17), and musculoskeletal radiograph (12) analysis. The budget for AI was reserved in 13 of the responding centers in both 2021 and 2022. The most important obstacles to the adoption of AI remained costs and IT integration. Of the respondents, 28% stated that the implemented AI products realized health improvement and 32% assumed both health improvement and cost savings. CONCLUSION: The adoption of AI products in radiology departments in the Netherlands is showing common signs of a developing market. The major obstacles to reaching widespread adoption are a lack of financial resources and IT integration difficulties. CLINICAL RELEVANCE STATEMENT: The clinical impact of AI starts with its adoption in daily clinical practice. Increased transparency around AI products being adopted, implementation obstacles, and impact may inspire increased collaboration and improved decision-making around the implementation and financing of AI products. KEY POINTS: • The adoption of artificial intelligence products for radiology has steadily increased since 2020 to at least a third of the centers using AI in clinical practice in the Netherlands in 2022. • The main areas in which artificial intelligence products are used are lung nodule detection on CT, aided stroke diagnosis, and bone age prediction. • The majority of respondents experienced added value (decreased costs and/or improved outcomes) from using artificial intelligence-based software; however, major obstacles to adoption remain the costs and IT-related difficulties.


Subject(s)
Artificial Intelligence , Radiology , Humans , Netherlands , Radiography , Radiologists
3.
Eur Radiol ; 2024 May 17.
Article in English | MEDLINE | ID: mdl-38758252

ABSTRACT

INTRODUCTION: This study investigates the performance of a commercially available artificial intelligence (AI) system to identify normal chest radiographs and its potential to reduce radiologist workload. METHODS: Retrospective analysis included consecutive chest radiographs from two medical centers between Oct 1, 2016 and Oct 14, 2016. Exclusions comprised follow-up exams within the inclusion period, bedside radiographs, incomplete images, imported radiographs, and pediatric radiographs. Three chest radiologists categorized findings into normal, clinically irrelevant, clinically relevant, urgent, and critical. A commercial AI system processed all radiographs, scoring 10 chest abnormalities on a 0-100 confidence scale. AI system performance was evaluated using the area under the ROC curve (AUC), assessing the detection of normal radiographs. Sensitivity was calculated for the default and a conservative operating point. the detection of negative predictive value (NPV) for urgent and critical findings, as well as the potential workload reduction, was calculated. RESULTS: A total of 2603 radiographs were acquired in 2141 unique patients. Post-exclusion, 1670 radiographs were analyzed. Categories included 479 normal, 332 clinically irrelevant, 339 clinically relevant, 501 urgent, and 19 critical findings. The AI system achieved an AUC of 0.92. Sensitivity for normal radiographs was 92% at default and 53% at the conservative operating point. At the conservative operating point, NPV was 98% for urgent and critical findings, and could result in a 15% workload reduction. CONCLUSION: A commercially available AI system effectively identifies normal chest radiographs and holds the potential to lessen radiologists' workload by omitting half of the normal exams from reporting. CLINICAL RELEVANCE STATEMENT: The AI system is able to detect half of all normal chest radiographs at a clinically acceptable operating point, thereby potentially reducing the workload for the radiologists by 15%. KEY POINTS: The AI system reached an AUC of 0.92 for the detection of normal chest radiographs. Fifty-three percent of normal chest radiographs were identified with a NPV of 98% for urgent findings. AI can reduce the workload of chest radiography reporting by 15%.

4.
Pediatr Radiol ; 52(11): 2120-2130, 2022 10.
Article in English | MEDLINE | ID: mdl-34471961

ABSTRACT

Artificial intelligence (AI) applications for chest radiography and chest CT are among the most developed applications in radiology. More than 40 certified AI products are available for chest radiography or chest CT. These AI products cover a wide range of abnormalities, including pneumonia, pneumothorax and lung cancer. Most applications are aimed at detecting disease, complemented by products that characterize or quantify tissue. At present, none of the thoracic AI products is specifically designed for the pediatric population. However, some products developed to detect tuberculosis in adults are also applicable to children. Software is under development to detect early changes of cystic fibrosis on chest CT, which could be an interesting application for pediatric radiology. In this review, we give an overview of current AI products in thoracic radiology and cover recent literature about AI in chest radiography, with a focus on pediatric radiology. We also discuss possible pediatric applications.


Subject(s)
Artificial Intelligence , Radiology , Adult , Child , Humans , Radiography, Thoracic , Thorax , Tomography, X-Ray Computed
5.
Pediatr Radiol ; 52(11): 2087-2093, 2022 10.
Article in English | MEDLINE | ID: mdl-34117522

ABSTRACT

Since the introduction of artificial intelligence (AI) in radiology, the promise has been that it will improve health care and reduce costs. Has AI been able to fulfill that promise? We describe six clinical objectives that can be supported by AI: a more efficient workflow, shortened reading time, a reduction of dose and contrast agents, earlier detection of disease, improved diagnostic accuracy and more personalized diagnostics. We provide examples of use cases including the available scientific evidence for its impact based on a hierarchical model of efficacy. We conclude that the market is still maturing and little is known about the contribution of AI to clinical practice. More real-world monitoring of AI in clinical practice is expected to aid in determining the value of AI and making informed decisions on development, procurement and reimbursement.


Subject(s)
Artificial Intelligence , Radiology , Contrast Media , Humans , Outcome Assessment, Health Care , Radiography
6.
Radiology ; 298(1): E46-E54, 2021 01.
Article in English | MEDLINE | ID: mdl-32787701

ABSTRACT

Background The prognosis of hospitalized patients with severe coronavirus disease 2019 (COVID-19) is difficult to predict, and the capacity of intensive care units was a limiting factor during the peak of the pandemic and is generally dependent on a country's clinical resources. Purpose To determine the value of chest radiographic findings together with patient history and laboratory markers at admission to predict critical illness in hospitalized patients with COVID-19. Materials and Methods In this retrospective study, which included patients from March 7, 2020, to April 24, 2020, a consecutive cohort of hospitalized patients with real-time reverse transcription polymerase chain reaction-confirmed COVID-19 from two large Dutch community hospitals was identified. After univariable analysis, a risk model to predict critical illness (ie, death and/or intensive care unit admission with invasive ventilation) was developed, using multivariable logistic regression including clinical, chest radiographic, and laboratory findings. Distribution and severity of lung involvement were visually assessed by using an eight-point scale (chest radiography score). Internal validation was performed by using bootstrapping. Performance is presented as an area under the receiver operating characteristic curve. Decision curve analysis was performed, and a risk calculator was derived. Results The cohort included 356 hospitalized patients (mean age, 69 years ± 12 [standard deviation]; 237 men) of whom 168 (47%) developed critical illness. The final risk model's variables included sex, chronic obstructive lung disease, symptom duration, neutrophil count, C-reactive protein level, lactate dehydrogenase level, distribution of lung disease, and chest radiography score at hospital presentation. The area under the receiver operating characteristic curve of the model was 0.77 (95% CI: 0.72, 0.81; P < .001). A risk calculator was derived for individual risk assessment: Dutch COVID-19 risk model. At an example threshold of 0.70, 71 of 356 patients would be predicted to develop critical illness, of which 59 (83%) would be true-positive results. Conclusion A risk model based on chest radiographic and laboratory findings obtained at admission was predictive of critical illness in hospitalized patients with coronavirus disease 2019. This risk calculator might be useful for triage of patients to the limited number of intensive care unit beds or facilities. © RSNA, 2020 Online supplemental material is available for this article.


Subject(s)
COVID-19/diagnostic imaging , Hospitalization , Radiography, Thoracic , Aged , Aged, 80 and over , Cohort Studies , Critical Illness/epidemiology , Female , Humans , Male , Middle Aged , Models, Theoretical , Prognosis , Retrospective Studies
7.
Radiology ; 298(2): E98-E106, 2021 02.
Article in English | MEDLINE | ID: mdl-33201791

ABSTRACT

Background Clinicians need to rapidly and reliably diagnose coronavirus disease 2019 (COVID-19) for proper risk stratification, isolation strategies, and treatment decisions. Purpose To assess the real-life performance of radiologist emergency department chest CT interpretation for diagnosing COVID-19 during the acute phase of the pandemic, using the COVID-19 Reporting and Data System (CO-RADS). Materials and Methods This retrospective multicenter study included consecutive patients who presented to emergency departments in six medical centers between March and April 2020 with moderate to severe upper respiratory symptoms suspicious for COVID-19. As part of clinical practice, chest CT scans were obtained for primary work-up and scored using the five-point CO-RADS scheme for suspicion of COVID-19. CT was compared with severe acute respiratory syndrome coronavirus 2 reverse-transcription polymerase chain reaction (RT-PCR) assay and a clinical reference standard established by a multidisciplinary group of clinicians based on RT-PCR, COVID-19 contact history, oxygen therapy, timing of RT-PCR testing, and likely alternative diagnosis. Performance of CT was estimated using area under the receiver operating characteristic curve (AUC) analysis and diagnostic odds ratios against both reference standards. Subgroup analysis was performed on the basis of symptom duration grouped presentations of less than 48 hours, 48 hours through 7 days, and more than 7 days. Results A total of 1070 patients (median age, 66 years; interquartile range, 54-75 years; 626 men) were included, of whom 536 (50%) had a positive RT-PCR result and 137 (13%) of whom were considered to have a possible or probable COVID-19 diagnosis based on the clinical reference standard. Chest CT yielded an AUC of 0.87 (95% CI: 0.84, 0.89) compared with RT-PCR and 0.87 (95% CI: 0.85, 0.89) compared with the clinical reference standard. A CO-RADS score of 4 or greater yielded an odds ratio of 25.9 (95% CI: 18.7, 35.9) for a COVID-19 diagnosis with RT-PCR and an odds ratio of 30.6 (95% CI: 21.1, 44.4) with the clinical reference standard. For symptom duration of less than 48 hours, the AUC fell to 0.71 (95% CI: 0.62, 0.80; P < .001). Conclusion Chest CT analysis using the coronavirus disease 2019 (COVID-19) Reporting and Data System enables rapid and reliable diagnosis of COVID-19, particularly when symptom duration is greater than 48 hours. © RSNA, 2020 Online supplemental material is available for this article. See also the editorial by Elicker in this issue.


Subject(s)
COVID-19/diagnostic imaging , Emergency Service, Hospital , Lung/diagnostic imaging , Tomography, X-Ray Computed/methods , Aged , Female , Humans , Male , Middle Aged , Netherlands , Retrospective Studies , SARS-CoV-2 , Sensitivity and Specificity
8.
Radiology ; 298(1): E18-E28, 2021 01.
Article in English | MEDLINE | ID: mdl-32729810

ABSTRACT

Background The coronavirus disease 2019 (COVID-19) pandemic has spread across the globe with alarming speed, morbidity, and mortality. Immediate triage of patients with chest infections suspected to be caused by COVID-19 using chest CT may be of assistance when results from definitive viral testing are delayed. Purpose To develop and validate an artificial intelligence (AI) system to score the likelihood and extent of pulmonary COVID-19 on chest CT scans using the COVID-19 Reporting and Data System (CO-RADS) and CT severity scoring systems. Materials and Methods The CO-RADS AI system consists of three deep-learning algorithms that automatically segment the five pulmonary lobes, assign a CO-RADS score for the suspicion of COVID-19, and assign a CT severity score for the degree of parenchymal involvement per lobe. This study retrospectively included patients who underwent a nonenhanced chest CT examination because of clinical suspicion of COVID-19 at two medical centers. The system was trained, validated, and tested with data from one of the centers. Data from the second center served as an external test set. Diagnostic performance and agreement with scores assigned by eight independent observers were measured using receiver operating characteristic analysis, linearly weighted κ values, and classification accuracy. Results A total of 105 patients (mean age, 62 years ± 16 [standard deviation]; 61 men) and 262 patients (mean age, 64 years ± 16; 154 men) were evaluated in the internal and external test sets, respectively. The system discriminated between patients with COVID-19 and those without COVID-19, with areas under the receiver operating characteristic curve of 0.95 (95% CI: 0.91, 0.98) and 0.88 (95% CI: 0.84, 0.93), for the internal and external test sets, respectively. Agreement with the eight human observers was moderate to substantial, with mean linearly weighted κ values of 0.60 ± 0.01 for CO-RADS scores and 0.54 ± 0.01 for CT severity scores. Conclusion With high diagnostic performance, the CO-RADS AI system correctly identified patients with COVID-19 using chest CT scans and assigned standardized CO-RADS and CT severity scores that demonstrated good agreement with findings from eight independent observers and generalized well to external data. © RSNA, 2020 Supplemental material is available for this article.


Subject(s)
Artificial Intelligence , COVID-19/diagnostic imaging , Severity of Illness Index , Thorax/diagnostic imaging , Tomography, X-Ray Computed , Aged , Data Systems , Female , Humans , Male , Middle Aged , Research Design , Retrospective Studies
9.
Eur Radiol ; 31(6): 3797-3804, 2021 Jun.
Article in English | MEDLINE | ID: mdl-33856519

ABSTRACT

OBJECTIVES: Map the current landscape of commercially available artificial intelligence (AI) software for radiology and review the availability of their scientific evidence. METHODS: We created an online overview of CE-marked AI software products for clinical radiology based on vendor-supplied product specifications ( www.aiforradiology.com ). Characteristics such as modality, subspeciality, main task, regulatory information, deployment, and pricing model were retrieved. We conducted an extensive literature search on the available scientific evidence of these products. Articles were classified according to a hierarchical model of efficacy. RESULTS: The overview included 100 CE-marked AI products from 54 different vendors. For 64/100 products, there was no peer-reviewed evidence of its efficacy. We observed a large heterogeneity in deployment methods, pricing models, and regulatory classes. The evidence of the remaining 36/100 products comprised 237 papers that predominantly (65%) focused on diagnostic accuracy (efficacy level 2). From the 100 products, 18 had evidence that regarded level 3 or higher, validating the (potential) impact on diagnostic thinking, patient outcome, or costs. Half of the available evidence (116/237) were independent and not (co-)funded or (co-)authored by the vendor. CONCLUSIONS: Even though the commercial supply of AI software in radiology already holds 100 CE-marked products, we conclude that the sector is still in its infancy. For 64/100 products, peer-reviewed evidence on its efficacy is lacking. Only 18/100 AI products have demonstrated (potential) clinical impact. KEY POINTS: • Artificial intelligence in radiology is still in its infancy even though already 100 CE-marked AI products are commercially available. • Only 36 out of 100 products have peer-reviewed evidence of which most studies demonstrate lower levels of efficacy. • There is a wide variety in deployment strategies, pricing models, and CE marking class of AI products for radiology.


Subject(s)
Artificial Intelligence , Radiology , Humans , Radiography , Software
10.
Radiology ; 296(3): E166-E172, 2020 09.
Article in English | MEDLINE | ID: mdl-32384019

ABSTRACT

Background Chest radiography may play an important role in triage for coronavirus disease 2019 (COVID-19), particularly in low-resource settings. Purpose To evaluate the performance of an artificial intelligence (AI) system for detection of COVID-19 pneumonia on chest radiographs. Materials and Methods An AI system (CAD4COVID-XRay) was trained on 24 678 chest radiographs, including 1540 used only for validation while training. The test set consisted of a set of continuously acquired chest radiographs (n = 454) obtained in patients suspected of having COVID-19 pneumonia between March 4 and April 6, 2020, at one center (223 patients with positive reverse transcription polymerase chain reaction [RT-PCR] results, 231 with negative RT-PCR results). Radiographs were independently analyzed by six readers and by the AI system. Diagnostic performance was analyzed with the receiver operating characteristic curve. Results For the test set, the mean age of patients was 67 years ± 14.4 (standard deviation) (56% male). With RT-PCR test results as the reference standard, the AI system correctly classified chest radiographs as COVID-19 pneumonia with an area under the receiver operating characteristic curve of 0.81. The system significantly outperformed each reader (P < .001 using the McNemar test) at their highest possible sensitivities. At their lowest sensitivities, only one reader significantly outperformed the AI system (P = .04). Conclusion The performance of an artificial intelligence system in the detection of coronavirus disease 2019 on chest radiographs was comparable with that of six independent readers. © RSNA, 2020.


Subject(s)
Artificial Intelligence , Coronavirus Infections/diagnostic imaging , Pneumonia, Viral/diagnostic imaging , Radiographic Image Interpretation, Computer-Assisted/methods , Radiography, Thoracic/methods , Aged , Aged, 80 and over , Betacoronavirus , COVID-19 , Databases, Factual , Female , Humans , Male , Middle Aged , Pandemics , ROC Curve , SARS-CoV-2 , Tomography, X-Ray Computed
11.
Clin Infect Dis ; 68(9): 1463-1471, 2019 04 24.
Article in English | MEDLINE | ID: mdl-30307492

ABSTRACT

BACKGROUND: Triazole resistance is an increasing problem in invasive aspergillosis (IA). Small case series show mortality rates of 50%-100% in patients infected with a triazole-resistant Aspergillus fumigatus, but a direct comparison with triazole-susceptible IA is lacking. METHODS: A 5-year retrospective cohort study (2011-2015) was conducted to compare mortality in patients with voriconazole-susceptible and voriconazole-resistant IA. Aspergillus fumigatus culture-positive patients were investigated to identify patients with proven, probable, and putative IA. Clinical characteristics, day 42 and day 90 mortality, triazole-resistance profiles, and antifungal treatments were investigated. RESULTS: Of 196 patients with IA, 37 (19%) harbored a voriconazole-resistant infection. Hematological malignancy was the underlying disease in 103 (53%) patients, and 154 (79%) patients were started on voriconazole. Compared with voriconazole-susceptible cases, voriconazole resistance was associated with an increase in overall mortality of 21% on day 42 (49% vs 28%; P = .017) and 25% on day 90 (62% vs 37%; P = .0038). In non-intensive care unit patients, a 19% lower survival rate was observed in voriconazole-resistant cases at day 42 (P = .045). The mortality in patients who received appropriate initial voriconazole therapy was 24% compared with 47% in those who received inappropriate therapy (P = .016), despite switching to appropriate antifungal therapy after a median of 10 days. CONCLUSIONS: Voriconazole resistance was associated with an excess overall mortality of 21% at day 42 and 25% at day 90 in patients with IA. A delay in the initiation of appropriate antifungal therapy was associated with increased overall mortality.


Subject(s)
Aspergillus fumigatus/genetics , Autoimmune Diseases/drug therapy , Drug Resistance, Fungal/genetics , Hematologic Neoplasms/drug therapy , Invasive Pulmonary Aspergillosis/drug therapy , Voriconazole/therapeutic use , Adolescent , Adult , Aged , Aged, 80 and over , Antifungal Agents/therapeutic use , Aspergillus fumigatus/drug effects , Aspergillus fumigatus/pathogenicity , Autoimmune Diseases/complications , Autoimmune Diseases/microbiology , Autoimmune Diseases/mortality , Child , Child, Preschool , Female , Hematologic Neoplasms/complications , Hematologic Neoplasms/microbiology , Hematologic Neoplasms/mortality , Humans , Invasive Pulmonary Aspergillosis/complications , Invasive Pulmonary Aspergillosis/microbiology , Invasive Pulmonary Aspergillosis/mortality , Male , Microbial Sensitivity Tests , Middle Aged , Retrospective Studies , Survival Analysis , Time Factors
12.
Radiology ; 272(1): 252-61, 2014 Jul.
Article in English | MEDLINE | ID: mdl-24635675

ABSTRACT

PURPOSE: To evaluate the added value of computer-aided detection (CAD) for lung nodules on chest radiographs when radiologists have bone-suppressed images (BSIs) available. MATERIALS AND METHODS: Written informed consent was waived by the institutional review board. Selection of study images and study setup was reviewed and approved by the institutional review boards. Three hundred posteroanterior (PA) and lateral chest radiographs (189 radiographs with negative findings and 111 radiographs with a solitary nodule) in 300 subjects were selected from image archives at four institutions. PA images were processed by using a commercially available CAD, and PA BSIs were generated. Five radiologists and three residents evaluated the radiographs with BSIs available, first, without CAD and, second, after inspection of the CAD marks. Readers marked locations suspicious for a nodule and provided a confidence score for that location to be a nodule. Location-based receiver operating characteristic analysis was performed by using jackknife alternative free-response receiver operating characteristic analysis. Area under the curve (AUC) functioned as figure of merit, and P values were computed with the Dorfman-Berbaum-Metz method. RESULTS: Average nodule size was 16.2 mm. Stand-alone CAD reached a sensitivity of 74% at 1.0 false-positive mark per image. Without CAD, average AUC for observers was 0.812. With CAD, performance significantly improved to an AUC of 0.841 (P = .0001). CAD detected 127 of 239 nodules that were missed after evaluation of the radiographs together with BSIs pooled over all observers. Only 57 of these detections were eventually marked by the observers after review of CAD candidates. CONCLUSION: CAD improved radiologists' performance for the detection of lung nodules on chest radiographs, even when baseline performance was optimized by providing lateral radiographs and BSIs. Still, most of the true-positive CAD candidates are dismissed by observers.


Subject(s)
Diagnosis, Computer-Assisted/methods , Radiography, Thoracic/methods , Solitary Pulmonary Nodule/diagnostic imaging , Tomography, X-Ray Computed/methods , Adult , Aged , Aged, 80 and over , Female , Humans , Male , Middle Aged , Radiographic Image Interpretation, Computer-Assisted , Retrospective Studies , Sensitivity and Specificity
13.
Insights Imaging ; 15(1): 51, 2024 Feb 16.
Article in English | MEDLINE | ID: mdl-38366046

ABSTRACT

BACKGROUND: Intended use statements (IUSs) are mandatory to obtain regulatory clearance for artificial intelligence (AI)-based medical devices in the European Union. In order to guide the safe use of AI-based medical devices, IUSs need to contain comprehensive and understandable information. This study analyzes the IUSs of CE-marked AI products listed on AIforRadiology.com for ambiguity and completeness. METHODS: We retrieved 157 IUSs of CE-marked AI products listed on AIforRadiology.com in September 2022. Duplicate products (n = 1), discontinued products (n = 3), and duplicate statements (n = 14) were excluded. The resulting IUSs were assessed for the presence of 6 items: medical indication, part of the body, patient population, user profile, use environment, and operating principle. Disclaimers, defined as contra-indications or warnings in the IUS, were identified and compared with claims. RESULTS: Of 139 AI products, the majority (n = 78) of IUSs mentioned 3 or less items. IUSs of only 7 products mentioned all 6 items. The intended body part (n = 115) and the operating principle (n = 116) were the most frequently mentioned components, while the intended use environment (n = 24) and intended patient population (n = 29) were mentioned less frequently. Fifty-six statements contained disclaimers that conflicted with the claims in 13 cases. CONCLUSION: The majority of IUSs of CE-marked AI-based medical devices lack substantial information and, in few cases, contradict the claims of the product. CRITICAL RELEVANCE STATEMENT: To ensure correct usage and to avoid off-label use or foreseeable misuse of AI-based medical devices in radiology, manufacturers are encouraged to provide more comprehensive and less ambiguous intended use statements. KEY POINTS: • Radiologists must know AI products' intended use to avoid off-label use or misuse. • Ninety-five percent (n = 132/139) of the intended use statements analyzed were incomplete. • Nine percent (n = 13) of the intended use statements held disclaimers contradicting the claim of the AI product. • Manufacturers and regulatory bodies must ensure that intended use statements are comprehensive.

14.
IEEE Trans Med Imaging ; PP2024 Mar 26.
Article in English | MEDLINE | ID: mdl-38530714

ABSTRACT

Pulmonary nodules may be an early manifestation of lung cancer, the leading cause of cancer-related deaths among both men and women. Numerous studies have established that deep learning methods can yield high-performance levels in the detection of lung nodules in chest X-rays. However, the lack of gold-standard public datasets slows down the progression of the research and prevents benchmarking of methods for this task. To address this, we organized a public research challenge, NODE21, aimed at the detection and generation of lung nodules in chest X-rays. While the detection track assesses state-of-the-art nodule detection systems, the generation track determines the utility of nodule generation algorithms to augment training data and hence improve the performance of the detection systems. This paper summarizes the results of the NODE21 challenge and performs extensive additional experiments to examine the impact of the synthetically generated nodule training images on the detection algorithm performance.

15.
Sci Rep ; 13(1): 12551, 2023 08 02.
Article in English | MEDLINE | ID: mdl-37532773

ABSTRACT

The Automation Platform (AP) is a software platform to support the workflow of radiologists and includes a stroke CT package with integrated artificial intelligence (AI) based tools. The aim of this study was to evaluate the diagnostic performance of the AP for the detection of intracranial large vessel occlusions (LVO) on conventional CT angiography (CTA), and the duration of CT processing in a cohort of acute stroke patients. The diagnostic performance for intracranial LVO detection on CTA by the AP was evaluated in a retrospective cohort of 100 acute stroke patients and compared to the diagnostic performance of five radiologists with different levels of experience. The reference standard was set by an independent neuroradiologist, with access to the readings of the different radiologists, clinical data, and follow-up. The data processing time of the AP for ICH detection on non-contrast CT, LVO detection on CTA, and the processing of CTP maps was assessed in a subset 60 patients of the retrospective cohort. This was compared to 13 radiologists, who were prospectively timed for the processing and reading of 21 stroke CTs. The AP showed shorter processing time of CTA (mean 60 versus 395 s) and CTP (mean 196 versus 243-349 s) as compared to radiologists, but showed lower sensitivity for LVO detection (sensitivity 77% of the AP vs mean sensitivity 87% of radiologists). If the AP would have been used as a stand-alone system, 1 ICA occlusion, 2 M1 occlusions and 8 M2 occlusions would have been missed, which would be eligible for mechanical thrombectomy. In conclusion, the AP showed shorter processing time of CTA and CTP as compared with radiologists, which illustrates the potential of the AP to speed-up the diagnostic work-up. However, its performance for LVO detection was lower as compared with radiologists, especially for M2 vessel occlusions.


Subject(s)
Brain Ischemia , Stroke , Humans , Artificial Intelligence , Retrospective Studies , Workflow , Cerebral Angiography , Stroke/diagnostic imaging , Computed Tomography Angiography
16.
Sci Rep ; 13(1): 19692, 2023 11 11.
Article in English | MEDLINE | ID: mdl-37952026

ABSTRACT

Artificial intelligence (AI) systems for detection of COVID-19 using chest X-Ray (CXR) imaging and point-of-care blood tests were applied to data from four low resource African settings. The performance of these systems to detect COVID-19 using various input data was analysed and compared with antigen-based rapid diagnostic tests. Participants were tested using the gold standard of RT-PCR test (nasopharyngeal swab) to determine whether they were infected with SARS-CoV-2. A total of 3737 (260 RT-PCR positive) participants were included. In our cohort, AI for CXR images was a poor predictor of COVID-19 (AUC = 0.60), since the majority of positive cases had mild symptoms and no visible pneumonia in the lungs. AI systems using differential white blood cell counts (WBC), or a combination of WBC and C-Reactive Protein (CRP) both achieved an AUC of 0.74 with a suggested optimal cut-off point at 83% sensitivity and 63% specificity. The antigen-RDT tests in this trial obtained 65% sensitivity at 98% specificity. This study is the first to validate AI tools for COVID-19 detection in an African setting. It demonstrates that screening for COVID-19 using AI with point-of-care blood tests is feasible and can operate at a higher sensitivity level than antigen testing.


Subject(s)
COVID-19 , Humans , COVID-19/diagnostic imaging , SARS-CoV-2 , Artificial Intelligence , Point-of-Care Systems , Sensitivity and Specificity , Leukocyte Count
17.
Commun Med (Lond) ; 3(1): 156, 2023 Oct 27.
Article in English | MEDLINE | ID: mdl-37891360

ABSTRACT

BACKGROUND: Outside a screening program, early-stage lung cancer is generally diagnosed after the detection of incidental nodules in clinically ordered chest CT scans. Despite the advances in artificial intelligence (AI) systems for lung cancer detection, clinical validation of these systems is lacking in a non-screening setting. METHOD: We developed a deep learning-based AI system and assessed its performance for the detection of actionable benign nodules (requiring follow-up), small lung cancers, and pulmonary metastases in CT scans acquired in two Dutch hospitals (internal and external validation). A panel of five thoracic radiologists labeled all nodules, and two additional radiologists verified the nodule malignancy status and searched for any missed cancers using data from the national Netherlands Cancer Registry. The detection performance was evaluated by measuring the sensitivity at predefined false positive rates on a free receiver operating characteristic curve and was compared with the panel of radiologists. RESULTS: On the external test set (100 scans from 100 patients), the sensitivity of the AI system for detecting benign nodules, primary lung cancers, and metastases is respectively 94.3% (82/87, 95% CI: 88.1-98.8%), 96.9% (31/32, 95% CI: 91.7-100%), and 92.0% (104/113, 95% CI: 88.5-95.5%) at a clinically acceptable operating point of 1 false positive per scan (FP/s). These sensitivities are comparable to or higher than the radiologists, albeit with a slightly higher FP/s (average difference of 0.6). CONCLUSIONS: The AI system reliably detects benign and malignant pulmonary nodules in clinically indicated CT scans and can potentially assist radiologists in this setting.


Early-stage lung cancer can be diagnosed after identifying an abnormal spot on a chest CT scan ordered for other medical reasons. These spots or lung nodules can be overlooked by radiologists, as they are not necessarily the focus of an examination and can be as small as a few millimeters. Software using Artificial Intelligence (AI) technology has proven to be successful for aiding radiologists in this task, but its performance is understudied outside a lung cancer screening setting. We therefore developed and validated AI software for the detection of cancerous nodules or non-cancerous nodules that would need attention. We show that the software can reliably detect these nodules in a non-screening setting and could potentially aid radiologists in daily clinical practice.

18.
PLoS One ; 17(7): e0267539, 2022.
Article in English | MEDLINE | ID: mdl-35900979

ABSTRACT

We propose a deep learning system to automatically detect four explainable emphysema signs on frontal and lateral chest radiographs. Frontal and lateral chest radiographs from 3000 studies were retrospectively collected. Two radiologists annotated these with 4 radiological signs of pulmonary emphysema identified from the literature. A patient with ≥2 of these signs present is considered emphysema positive. Using separate deep learning systems for frontal and lateral images we predict the presence of each of the four visual signs and use these to determine emphysema positivity. The ROC and AUC results on a set of 422 held-out cases, labeled by both radiologists, are reported. Comparison with a black-box model which predicts emphysema without the use of explainable visual features is made on the annotations from both radiologists, as well as the subset that they agreed on. DeLong's test is used to compare with the black-box model ROC and McNemar's test to compare with radiologist performance. In 422 test cases, emphysema positivity was predicted with AUCs of 0.924 and 0.946 using the reference standard from each radiologist separately. Setting model sensitivity equivalent to that of the second radiologist, our model has a comparable specificity (p = 0.880 and p = 0.143 for each radiologist respectively). Our method is comparable with the black-box model with AUCs of 0.915 (p = 0.407) and 0.935 (p = 0.291), respectively. On the 370 cases where both radiologists agreed (53 positives), our model achieves an AUC of 0.981, again comparable to the black-box model AUC of 0.972 (p = 0.289). Our proposed method can predict emphysema positivity on chest radiographs as well as a radiologist or a comparable black-box method. It additionally produces labels for four visual signs to ensure the explainability of the result. The dataset is publicly available at https://doi.org/10.5281/zenodo.6373392.


Subject(s)
Deep Learning , Emphysema , Pulmonary Emphysema , Emphysema/diagnostic imaging , Humans , Pulmonary Emphysema/diagnostic imaging , Radiography , Radiography, Thoracic/methods , Radiologists , Retrospective Studies
19.
J Fungi (Basel) ; 8(9)2022 Aug 31.
Article in English | MEDLINE | ID: mdl-36135650

ABSTRACT

OBJECTIVES: Patients receiving remission induction therapy for acute myeloid leukaemia (AML) are at high risk of developing invasive fungal disease (IFD). Newer therapies with targeted antileukemic agents and the emergence of azole resistance pose a challenge to the strategy of primary antifungal prophylaxis. We report the experience of a diagnostic-driven care pathway (DCP) for the management of IFD in these patients, using only culture-directed mould inactive prophylaxis. METHODS: Retrospectively, we used a single-centre study of consecutive patients receiving intensive chemotherapy for myeloid malignancies between 2014 and 2021. DCP consisted of serial cultures and serum galactomannan (sGM) screening, CT imaging, and bronchoscopy to direct targeted antifungal treatment. IFD was classified according to the 2020 EORTC/MSGERC criteria. RESULTS: A total of 192 patients with myeloid malignancies received 300 courses of intensive chemotherapy. There were 14 cases of invasive yeast infections and 18 of probable/proven invasive mould disease (IMD). The incidence of probable/proven IMD during the first cycle of remission-induction chemotherapy was 4.6% (n = 9). sGM remained negative in all cases of invasive aspergillosis (IA), with positive mycology findings in bronchoalveolar lavage. All-cause mortality was 9.4% (n = 18) 100 days after starting chemotherapy and was comparable between patients with or without IFD. The fungal-related mortality was 1% (n = 2). CONCLUSION: Diagnostic-driven based management without universal mould active prophylaxis is a feasible strategy in the management of IFD and limits unnecessary antimould treatment during intensive chemotherapy. The poor performance of serial serum galactomannan screening in detecting IA warrants further investigation.

20.
Insights Imaging ; 12(1): 133, 2021 Sep 25.
Article in English | MEDLINE | ID: mdl-34564764

ABSTRACT

BACKGROUND: Limited evidence is available on the clinical impact of artificial intelligence (AI) in radiology. Early health technology assessment (HTA) is a methodology to assess the potential value of an innovation at an early stage. We use early HTA to evaluate the potential value of AI software in radiology. As a use-case, we evaluate the cost-effectiveness of AI software aiding the detection of intracranial large vessel occlusions (LVO) in stroke in comparison to standard care. We used a Markov based model from a societal perspective of the United Kingdom predominantly using stroke registry data complemented with pooled outcome data from large, randomized trials. Different scenarios were explored by varying missed diagnoses of LVOs, AI costs and AI performance. Other input parameters were varied to demonstrate model robustness. Results were reported in expected incremental costs (IC) and effects (IE) expressed in quality adjusted life years (QALYs). RESULTS: Applying the base case assumptions (6% missed diagnoses of LVOs by clinicians, $40 per AI analysis, 50% reduction of missed LVOs by AI), resulted in cost-savings and incremental QALYs over the projected lifetime (IC: - $156, - 0.23%; IE: + 0.01 QALYs, + 0.07%) per suspected ischemic stroke patient. For each yearly cohort of patients in the UK this translates to a total cost saving of $11 million. CONCLUSIONS: AI tools for LVO detection in emergency care have the potential to improve healthcare outcomes and save costs. We demonstrate how early HTA may be applied for the evaluation of clinically applied AI software for radiology.

SELECTION OF CITATIONS
SEARCH DETAIL