Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 8 de 8
Filtrar
Mais filtros

Bases de dados
País/Região como assunto
Tipo de documento
Intervalo de ano de publicação
1.
EClinicalMedicine ; 70: 102479, 2024 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-38685924

RESUMO

Background: Artificial intelligence (AI) has repeatedly been shown to encode historical inequities in healthcare. We aimed to develop a framework to quantitatively assess the performance equity of health AI technologies and to illustrate its utility via a case study. Methods: Here, we propose a methodology to assess whether health AI technologies prioritise performance for patient populations experiencing worse outcomes, that is complementary to existing fairness metrics. We developed the Health Equity Assessment of machine Learning performance (HEAL) framework designed to quantitatively assess the performance equity of health AI technologies via a four-step interdisciplinary process to understand and quantify domain-specific criteria, and the resulting HEAL metric. As an illustrative case study (analysis conducted between October 2022 and January 2023), we applied the HEAL framework to a dermatology AI model. A set of 5420 teledermatology cases (store-and-forward cases from patients of 20 years or older, submitted from primary care providers in the USA and skin cancer clinics in Australia), enriched for diversity in age, sex and race/ethnicity, was used to retrospectively evaluate the AI model's HEAL metric, defined as the likelihood that the AI model performs better for subpopulations with worse average health outcomes as compared to others. The likelihood that AI performance was anticorrelated to pre-existing health outcomes was estimated using bootstrap methods as the probability that the negated Spearman's rank correlation coefficient (i.e., "R") was greater than zero. Positive values of R suggest that subpopulations with poorer health outcomes have better AI model performance. Thus, the HEAL metric, defined as p (R >0), measures how likely the AI technology is to prioritise performance for subpopulations with worse average health outcomes as compared to others (presented as a percentage below). Health outcomes were quantified as disability-adjusted life years (DALYs) when grouping by sex and age, and years of life lost (YLLs) when grouping by race/ethnicity. AI performance was measured as top-3 agreement with the reference diagnosis from a panel of 3 dermatologists per case. Findings: Across all dermatologic conditions, the HEAL metric was 80.5% for prioritizing AI performance of racial/ethnic subpopulations based on YLLs, and 92.1% and 0.0% respectively for prioritizing AI performance of sex and age subpopulations based on DALYs. Certain dermatologic conditions were significantly associated with greater AI model performance compared to a reference category of less common conditions. For skin cancer conditions, the HEAL metric was 73.8% for prioritizing AI performance of age subpopulations based on DALYs. Interpretation: Analysis using the proposed HEAL framework showed that the dermatology AI model prioritised performance for race/ethnicity, sex (all conditions) and age (cancer conditions) subpopulations with respect to pre-existing health disparities. More work is needed to investigate ways of promoting equitable AI performance across age for non-cancer conditions and to better understand how AI models can contribute towards improving equity in health outcomes. Funding: Google LLC.

2.
Transl Vis Sci Technol ; 12(12): 11, 2023 12 01.
Artigo em Inglês | MEDLINE | ID: mdl-38079169

RESUMO

Purpose: Real-world evaluation of a deep learning model that prioritizes patients based on risk of progression to moderate or worse (MOD+) diabetic retinopathy (DR). Methods: This nonrandomized, single-arm, prospective, interventional study included patients attending DR screening at four centers across Thailand from September 2019 to January 2020, with mild or no DR. Fundus photographs were input into the model, and patients were scheduled for their subsequent screening from September 2020 to January 2021 in order of predicted risk. Evaluation focused on model sensitivity, defined as correctly ranking patients that developed MOD+ within the first 50% of subsequent screens. Results: We analyzed 1,757 patients, of which 52 (3.0%) developed MOD+. Using the model-proposed order, the model's sensitivity was 90.4%. Both the model-proposed order and mild/no DR plus HbA1c had significantly higher sensitivity than the random order (P < 0.001). Excluding one major (rural) site that had practical implementation challenges, the remaining sites included 567 patients and 15 (2.6%) developed MOD+. Here, the model-proposed order achieved 86.7% versus 73.3% for the ranking that used DR grade and hemoglobin A1c. Conclusions: The model can help prioritize follow-up visits for the largest subgroups of DR patients (those with no or mild DR). Further research is needed to evaluate the impact on clinical management and outcomes. Translational Relevance: Deep learning demonstrated potential for risk stratification in DR screening. However, real-world practicalities must be resolved to fully realize the benefit.


Assuntos
Aprendizado Profundo , Diabetes Mellitus , Retinopatia Diabética , Humanos , Retinopatia Diabética/diagnóstico , Retinopatia Diabética/epidemiologia , Estudos Prospectivos , Hemoglobinas Glicadas , Medição de Risco
3.
Lancet Digit Health ; 4(4): e235-e244, 2022 04.
Artigo em Inglês | MEDLINE | ID: mdl-35272972

RESUMO

BACKGROUND: Diabetic retinopathy is a leading cause of preventable blindness, especially in low-income and middle-income countries (LMICs). Deep-learning systems have the potential to enhance diabetic retinopathy screenings in these settings, yet prospective studies assessing their usability and performance are scarce. METHODS: We did a prospective interventional cohort study to evaluate the real-world performance and feasibility of deploying a deep-learning system into the health-care system of Thailand. Patients with diabetes and listed on the national diabetes registry, aged 18 years or older, able to have their fundus photograph taken for at least one eye, and due for screening as per the Thai Ministry of Public Health guidelines were eligible for inclusion. Eligible patients were screened with the deep-learning system at nine primary care sites under Thailand's national diabetic retinopathy screening programme. Patients with a previous diagnosis of diabetic macular oedema, severe non-proliferative diabetic retinopathy, or proliferative diabetic retinopathy; previous laser treatment of the retina or retinal surgery; other non-diabetic retinopathy eye disease requiring referral to an ophthalmologist; or inability to have fundus photograph taken of both eyes for any reason were excluded. Deep-learning system-based interpretations of patient fundus images and referral recommendations were provided in real time. As a safety mechanism, regional retina specialists over-read each image. Performance of the deep-learning system (accuracy, sensitivity, specificity, positive predictive value [PPV], and negative predictive value [NPV]) were measured against an adjudicated reference standard, provided by fellowship-trained retina specialists. This study is registered with the Thai national clinical trials registry, TCRT20190902002. FINDINGS: Between Dec 12, 2018, and March 29, 2020, 7940 patients were screened for inclusion. 7651 (96·3%) patients were eligible for study analysis, and 2412 (31·5%) patients were referred for diabetic retinopathy, diabetic macular oedema, ungradable images, or low visual acuity. For vision-threatening diabetic retinopathy, the deep-learning system had an accuracy of 94·7% (95% CI 93·0-96·2), sensitivity of 91·4% (87·1-95·0), and specificity of 95·4% (94·1-96·7). The retina specialist over-readers had an accuracy of 93·5 (91·7-95·0; p=0·17), a sensitivity of 84·8% (79·4-90·0; p=0·024), and specificity of 95·5% (94·1-96·7; p=0·98). The PPV for the deep-learning system was 79·2 (95% CI 73·8-84·3) compared with 75·6 (69·8-81·1) for the over-readers. The NPV for the deep-learning system was 95·5 (92·8-97·9) compared with 92·4 (89·3-95·5) for the over-readers. INTERPRETATION: A deep-learning system can deliver real-time diabetic retinopathy detection capability similar to retina specialists in community-based screening settings. Socioenvironmental factors and workflows must be taken into consideration when implementing a deep-learning system within a large-scale screening programme in LMICs. FUNDING: Google and Rajavithi Hospital, Bangkok, Thailand. TRANSLATION: For the Thai translation of the abstract see Supplementary Materials section.


Assuntos
Aprendizado Profundo , Diabetes Mellitus , Retinopatia Diabética , Edema Macular , Estudos de Coortes , Retinopatia Diabética/diagnóstico , Humanos , Edema Macular/diagnóstico , Estudos Prospectivos , Tailândia
4.
Lancet Digit Health ; 3(1): e10-e19, 2021 01.
Artigo em Inglês | MEDLINE | ID: mdl-33735063

RESUMO

BACKGROUND: Diabetic retinopathy screening is instrumental to preventing blindness, but scaling up screening is challenging because of the increasing number of patients with all forms of diabetes. We aimed to create a deep-learning system to predict the risk of patients with diabetes developing diabetic retinopathy within 2 years. METHODS: We created and validated two versions of a deep-learning system to predict the development of diabetic retinopathy in patients with diabetes who had had teleretinal diabetic retinopathy screening in a primary care setting. The input for the two versions was either a set of three-field or one-field colour fundus photographs. Of the 575 431 eyes in the development set 28 899 had known outcomes, with the remaining 546 532 eyes used to augment the training process via multitask learning. Validation was done on one eye (selected at random) per patient from two datasets: an internal validation (from EyePACS, a teleretinal screening service in the USA) set of 3678 eyes with known outcomes and an external validation (from Thailand) set of 2345 eyes with known outcomes. FINDINGS: The three-field deep-learning system had an area under the receiver operating characteristic curve (AUC) of 0·79 (95% CI 0·77-0·81) in the internal validation set. Assessment of the external validation set-which contained only one-field colour fundus photographs-with the one-field deep-learning system gave an AUC of 0·70 (0·67-0·74). In the internal validation set, the AUC of available risk factors was 0·72 (0·68-0·76), which improved to 0·81 (0·77-0·84) after combining the deep-learning system with these risk factors (p<0·0001). In the external validation set, the corresponding AUC improved from 0·62 (0·58-0·66) to 0·71 (0·68-0·75; p<0·0001) following the addition of the deep-learning system to available risk factors. INTERPRETATION: The deep-learning systems predicted diabetic retinopathy development using colour fundus photographs, and the systems were independent of and more informative than available risk factors. Such a risk stratification tool might help to optimise screening intervals to reduce costs while improving vision-related outcomes. FUNDING: Google.


Assuntos
Aprendizado Profundo , Retinopatia Diabética/diagnóstico , Idoso , Área Sob a Curva , Técnicas de Diagnóstico Oftalmológico , Feminino , Humanos , Estimativa de Kaplan-Meier , Masculino , Pessoa de Meia-Idade , Fotografação , Prognóstico , Curva ROC , Reprodutibilidade dos Testes , Medição de Risco/métodos
5.
Abdom Imaging ; 34(3): 359-64, 2009.
Artigo em Inglês | MEDLINE | ID: mdl-18343970

RESUMO

BACKGROUND: The purpose of this study was to evaluate whether an electronic-colonic-cleansing (ECC) algorithm is beneficial for the diagnostic performance compared to a CT colonography (CTC) evaluation without electronic cleansing in tagged datasets. METHODS: Two blinded readers evaluated CTC datasets from 79 patients with 153 colorectal polyps confirmed by optical colonoscopy. Cases were read in a randomized order with and without the use of electronic colon-cleansing software. Per-polyp sensitivity, per-polyp/per-patient specificity and reading times (with and without ECC) have been calculated and reported. RESULTS: Per-polyp sensitivity for polyps >6 mm without using ECC was 60.4% (Reader 1: 59.7%, Reader 2: 61.1%), while polyps >10 mm were detected with a sensitivity of 58.3% (Reader 1: 66.7%, Reader 2: 50%). On electronically cleansed datasets, the sensitivity was 73.6% (Reader 1: 76.4%; Reader 2: 70.8%) for polyps >6 mm and 83.3% (Reader 1: 83.3%; Reader 2: 83.3%), respectively. Per-patient specificity was 75% without using cleansing (Reader 1: 68%, Reader 2: 82%) and 81.5% using ECC (Reader 1: 86%, Reader 2: 77%). CONCLUSION: Reading CTC cases using ECC software improves sensitivity in detecting clinically relevant colorectal polyps.


Assuntos
Algoritmos , Colonografia Tomográfica Computadorizada/métodos , Processamento de Imagem Assistida por Computador/métodos , Intensificação de Imagem Radiográfica/métodos , Técnica de Subtração/estatística & dados numéricos , Bário , Bisacodil/administração & dosagem , Catárticos/administração & dosagem , Colo/diagnóstico por imagem , Pólipos do Colo/diagnóstico por imagem , Meios de Contraste , Diatrizoato , Diatrizoato de Meglumina , Humanos , Imageamento Tridimensional/métodos , Variações Dependentes do Observador , Fosfatos/administração & dosagem , Sensibilidade e Especificidade , Fatores de Tempo
6.
Eye (Lond) ; 33(1): 97-109, 2019 01.
Artigo em Inglês | MEDLINE | ID: mdl-30401899

RESUMO

Remarkable advances in biomedical research have led to the generation of large amounts of data. Using artificial intelligence, it has become possible to extract meaningful information from large volumes of data, in a shorter frame of time, with very less human interference. In effect, convolutional neural networks (a deep learning method) have been taught to recognize pathological lesions from images. Diabetes has high morbidity, with millions of people who need to be screened for diabetic retinopathy (DR). Deep neural networks offer a great advantage of screening for DR from retinal images, in improved identification of DR lesions and risk factors for diseases, with high accuracy and reliability. This review aims to compare the current evidences on various deep learning models for diagnosis of diabetic retinopathy (DR).


Assuntos
Algoritmos , Aprendizado Profundo , Retinopatia Diabética/diagnóstico , Técnicas de Diagnóstico Oftalmológico , Redes Neurais de Computação , Fotografação/métodos , Fundo de Olho , Humanos , Curva ROC
7.
Cerebrovasc Dis ; 26(6): 600-5, 2008.
Artigo em Inglês | MEDLINE | ID: mdl-18946215

RESUMO

PURPOSE: To assess whether blood-brain barrier permeability (BBBP) values, extracted with the Patlak model from the second perfusion CT (PCT) contrast bolus, are significantly lower than the values extracted from the first bolus in the same patient. MATERIALS AND METHODS: 125 consecutive patients (29 with acute hemispheric stroke and 96 without stroke) who underwent a PCT study using a prolonged acquisition time up to 3 min were retrospectively identified. The Patlak model was applied to calculate the rate of contrast leakage out of the vascular compartment. Patlak plots were created from the arterial and parenchymal time enhancement curves obtained in multiple regions of interest drawn in ischemic brain tissue and in nonischemic brain tissue. The slope of a regression line fit to the Patlak plot was used as an indicator of BBBP. Square roots of the mean squared errors and correlation coefficients were used to describe the quality of the linear regression model. This was performed separately for the first and the second PCT bolus. Results from the first and the second bolus were compared in terms of BBBP values and the quality of the linear model fitted to the Patlak plot, using generalized estimating equations with robust variance estimation. RESULTS: BBBP values from the second bolus were not lower than BBBP values from the first bolus in either nonischemic brain tissue [estimated mean with 95% confidence interval: 1.42 (1.10-1.82) ml x 100 g(-1) x min(-1) for the first bolus versus 1.64 (1.31-2.05) ml x 100 g(-1) x min(-1) for the second bolus, p = 1.00] or in ischemic tissue [1.04 (0.97-1.12) ml x 100 g(-1) x min(-1) for the first bolus versus 1.19 (1.11-1.28) ml x 100 g(-1)min(-1) for the second bolus, p = 0.79]. Compared to regression models from the first bolus, the Patlak regression models obtained from the second bolus were of similar or slightly better quality. This was true both in nonischemic and ischemic brain tissue. CONCLUSION: The contrast material from the first bolus of contrast for PCT does not negatively influence measurements of BBBP values from the second bolus. The second bolus can thus be used to increase anatomical coverage of BBBP assessment using PCT.


Assuntos
Barreira Hematoencefálica , Meios de Contraste/administração & dosagem , Iohexol/administração & dosagem , Modelos Biológicos , Acidente Vascular Cerebral/fisiopatologia , Tomografia Computadorizada por Raios X , Doença Aguda , Adolescente , Adulto , Idoso , Idoso de 80 Anos ou mais , Isquemia Encefálica/diagnóstico por imagem , Isquemia Encefálica/fisiopatologia , Infarto Cerebral/diagnóstico por imagem , Infarto Cerebral/fisiopatologia , Meios de Contraste/farmacocinética , Feminino , Humanos , Injeções Intravenosas , Injeções a Jato , Iohexol/farmacocinética , Modelos Lineares , Masculino , Pessoa de Meia-Idade , Estudos Retrospectivos , Acidente Vascular Cerebral/diagnóstico por imagem , Adulto Jovem
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA