RESUMO
In low- and middle-income countries, the substantial costs associated with traditional data collection pose an obstacle to facilitating decision-making in the field of public health. Satellite imagery offers a potential solution, but the image extraction and analysis can be costly and requires specialized expertise. We introduce SatelliteBench, a scalable framework for satellite image extraction and vector embeddings generation. We also propose a novel multimodal fusion pipeline that utilizes a series of satellite imagery and metadata. The framework was evaluated generating a dataset with a collection of 12,636 images and embeddings accompanied by comprehensive metadata, from 81 municipalities in Colombia between 2016 and 2018. The dataset was then evaluated in 3 tasks: including dengue case prediction, poverty assessment, and access to education. The performance showcases the versatility and practicality of SatelliteBench, offering a reproducible, accessible and open tool to enhance decision-making in public health.
Assuntos
Dengue , Saúde Pública , Imagens de Satélites , Colômbia , Humanos , MetadadosRESUMO
In the wake of emergent natural and anthropogenic disasters, telehealth presents opportunities to improve access to healthcare when physical access is not possible. Yet, since the beginning of the COVID pandemic, lessons learned reveal that various populations in the United States do not or cannot adopt telehealth due to inequitable access. We explored the Digital Determinants of Health (DDoHs) for telehealth, characterizing the role of accessibility, broadband connectivity and electrical grids, and patient intersectionality. In addition to its role as an existing Social Determinant of Health, Policies and Laws directly and indirectly affect these DDoHs, making access more complex for marginalized populations. Digital systems lack the flexibility, accessibility, and usability to inclusively provide the essential services patients need in telehealth. We propose the following recommendations: (1) design technology and systems using accessibility and value sensitive design principles; (2) support a range of technologies and settings; (3) support multiple and diverse users; and (4) support clear paths for repair when technical systems fail to meet users' needs. Addressing these requires change not only from providers but also from the institutions providing these systems.
RESUMO
Healthcare has long struggled to improve services through technology without further widening health disparities. With the significant expansion of digital health, a group of healthcare professionals and scholars from across the globe are proposing the official usage of the term "Digital Determinants of Health" (DDOH) to explicitly call out the relationship between technology, healthcare, and equity. This is the final paper in a series published in PLOS Digital Health that seeks to understand and summarize current knowledge of the strategies and solutions that help to mitigate the negative effects of DDOH for underinvested communities. Through a search of English-language Medline, Scopus, and Google Scholar articles published since 2010, 345 articles were identified that discussed the application of digital health technology among underinvested communities. A group of 8 reviewers assessed 132 articles selected at random for the mention of solutions that minimize differences in DDOH. Solutions were then organized by categories of policy; design and development; implementation and adoption; and evaluation and ongoing monitoring. The data were then assessed by category and the findings summarized. The reviewers also looked for common themes across the solutions and evidence of effectiveness. From this limited scoping review, the authors found numerous solutions mentioned across the papers for addressing DDOH and many common themes emerged regardless of the specific community or digital health technology under review. There was notably less information on solutions regarding ongoing evaluation and monitoring which corresponded with a lack of research evidence regarding effectiveness. The findings directionally suggest that universal strategies and solutions can be developed to address DDOH independent of the specific community under focus. With the need for the further development of DDOH measures, we also provide a framework for DDOH assessment.
RESUMO
PURPOSE: Chronic obstructive pulmonary disease (COPD) is one of the most common chronic illnesses in the world. Unfortunately, COPD is often difficult to diagnose early when interventions can alter the disease course, and it is underdiagnosed or only diagnosed too late for effective treatment. Currently, spirometry is the gold standard for diagnosing COPD but it can be challenging to obtain, especially in resource-poor countries. Chest X-rays (CXRs), however, are readily available and may have the potential as a screening tool to identify patients with COPD who should undergo further testing or intervention. In this study, we used three CXR datasets alongside their respective electronic health records (EHR) to develop and externally validate our models. METHOD: To leverage the performance of convolutional neural network models, we proposed two fusion schemes: (1) model-level fusion, using Bootstrap aggregating to aggregate predictions from two models, (2) data-level fusion, using CXR image data from different institutions or multi-modal data, CXR image data, and EHR data for model training. Fairness analysis was then performed to evaluate the models across different demographic groups. RESULTS: Our results demonstrate that DL models can detect COPD using CXRs with an area under the curve of over 0.75, which could facilitate patient screening for COPD, especially in low-resource regions where CXRs are more accessible than spirometry. CONCLUSIONS: By using a ubiquitous test, future research could build on this work to detect COPD in patients early who would not otherwise have been diagnosed or treated, altering the course of this highly morbid disease.
RESUMO
Critical care data contain information about the most physiologically fragile patients in the hospital, who require a significant level of monitoring. However, medical devices used for patient monitoring suffer from measurement biases that have been largely underreported. This article explores sources of bias in commonly used clinical devices, including pulse oximeters, thermometers, and sphygmomanometers. Further, it provides a framework for mitigating these biases and key principles to achieve more equitable health care delivery.
Assuntos
Cuidados Críticos , Humanos , ViésRESUMO
The adoption of artificial intelligence (AI) algorithms is rapidly increasing in healthcare. Such algorithms may be shaped by various factors such as social determinants of health that can influence health outcomes. While AI algorithms have been proposed as a tool to expand the reach of quality healthcare to underserved communities and improve health equity, recent literature has raised concerns about the propagation of biases and healthcare disparities through implementation of these algorithms. Thus, it is critical to understand the sources of bias inherent in AI-based algorithms. This review aims to highlight the potential sources of bias within each step of developing AI algorithms in healthcare, starting from framing the problem, data collection, preprocessing, development, and validation, as well as their full implementation. For each of these steps, we also discuss strategies to mitigate the bias and disparities. A checklist was developed with recommendations for reducing bias during the development and implementation stages. It is important for developers and users of AI-based algorithms to keep these important considerations in mind to advance health equity for all populations.
RESUMO
BACKGROUND: Whether intubation should be initiated early in the clinical course of critically ill patients remains a matter of debate. Results from prior observational studies are difficult to interpret because of avoidable flaws including immortal time bias, inappropriate eligibility criteria, and unrealistic treatment strategies. RESEARCH QUESTION: Do treatment strategies that intubate patients early in the critical care admission improve 30-day survival compared with strategies that delay intubation? STUDY DESIGN AND METHODS: We estimated the effect of strategies that require early intubation of critically ill patients compared with those that delay intubation. With data extracted from the Medical Information Mart for Intensive Care-IV database, we emulated three target trials, varying the flexibility of the treatment strategies and the baseline eligibility criteria. RESULTS: Under unrealistically strict treatment strategies with broad eligibility criteria, the 30-day mortality risk was 7.1 percentage points higher for intubating early compared with delaying intubation (95% CI, 6.2-7.9). Risk differences were 0.4 (95% CI, -0.1 to 0.9) and -0.9 (95% CI, -2.5 to 0.7) percentage points in subsequent target trial emulations that included more realistic treatment strategies and eligibility criteria. INTERPRETATION: When realistic treatment strategies and eligibility criteria are used, strategies that delay intubation result in similar 30-day mortality risks compared with those that intubate early. Delaying intubation ultimately avoids intubation in most patients.
Assuntos
Estado Terminal , Ventilação não Invasiva , Humanos , Estado Terminal/terapia , Respiração Artificial , Ventilação não Invasiva/métodos , Intubação Intratraqueal , Cuidados CríticosRESUMO
As the health-care industry emerges into a new era of digital health driven by cloud data storage, distributed computing, and machine learning, health-care data have become a premium commodity with value for private and public entities. Current frameworks of health data collection and distribution, whether from industry, academia, or government institutions, are imperfect and do not allow researchers to leverage the full potential of downstream analytical efforts. In this Health Policy paper, we review the current landscape of commercial health data vendors, with special emphasis on the sources of their data, challenges associated with data reproducibility and generalisability, and ethical considerations for data vending. We argue for sustainable approaches to curating open-source health data to enable global populations to be included in the biomedical research community. However, to fully implement these approaches, key stakeholders should come together to make health-care datasets increasingly accessible, inclusive, and representative, while balancing the privacy and rights of individuals whose data are being collected.
Assuntos
Algoritmos , Pesquisa Biomédica , Conjuntos de Dados como Assunto , Humanos , Privacidade , Reprodutibilidade dos Testes , Conjuntos de Dados como Assunto/economia , Conjuntos de Dados como Assunto/ética , Conjuntos de Dados como Assunto/tendências , Informação de Saúde ao Consumidor/economia , Informação de Saúde ao Consumidor/éticaRESUMO
The field of medicine is undergoing rapid digital transformation. Pathologists are now striving to digitize their data, workflows, and interpretations, assisted by the enabling development of whole-slide imaging. Going digital means that the analog process of human diagnosis can be augmented or even replaced by rapidly evolving AI approaches, which are just now entering into clinical practice. But with such progress comes challenges that reflect a variety of stressors, including the impact of unrepresentative training data with accompanying implicit bias, data privacy concerns, and fragility of algorithm performance. Beyond such core digital aspects, considerations arise related to difficulties presented by changing disease presentations, diagnostic approaches, and therapeutic options. While some tools such as data federation can help with broadening data diversity while preserving expertise and local control, they may not be the full answer to some of these issues. The impact of AI in pathology on the field's human practitioners is still very much unknown: installation of unconscious bias and deference to AI guidance need to be understood and addressed. If AI is widely adopted, it may remove many inefficiencies in daily practice and compensate for staff shortages. It may also cause practitioner deskilling, dethrilling, and burnout. We discuss the technological, clinical, legal, and sociological factors that will influence the adoption of AI in pathology, and its eventual impact for good or ill.
Assuntos
Algoritmos , Patologistas , Humanos , Inteligência ArtificialRESUMO
This paper aims to identify uncommon cardiothoracic diseases and patterns on chest X-ray images. Training a machine learning model to classify rare diseases with multi-label indications is challenging without sufficient labeled training samples. Our model leverages the information from common diseases and adapts to perform on less common mentions. We propose to use multi-label few-shot learning (FSL) schemes including neighborhood component analysis loss, generating additional samples using distribution calibration and fine-tuning based on multi-label classification loss. We utilize the fact that the widely adopted nearest neighbor-based FSL schemes like ProtoNet are Voronoi diagrams in feature space. In our method, the Voronoi diagrams in the features space generated from multi-label schemes are combined into our geometric DeepVoro Multi-label ensemble. The improved performance in multi-label few-shot classification using the multi-label ensemble is demonstrated in our experiments (The code is publicly available at https://github.com/Saurabh7/Few-shot-learning-multilabel-cxray).
RESUMO
Humans are exposed to numerous compounds daily, some of which have adverse effects on health. Computational approaches for modeling toxicological data in conjunction with machine learning algorithms have gained popularity over the last few years. Machine learning approaches have been used to predict toxicity-related biological activities using chemical structure descriptors. However, toxicity-related proteomic features have not been fully investigated. In this study, we construct a computational pipeline using machine learning models for predicting the most important protein features responsible for the toxicity of compounds taken from the Tox21 dataset that is implemented within the multiscale Computational Analysis of Novel Drug Opportunities (CANDO) therapeutic discovery platform. Tox21 is a highly imbalanced dataset consisting of twelve in vitro assays, seven from the nuclear receptor (NR) signaling pathway and five from the stress response (SR) pathway, for more than 10,000 compounds. For the machine learning model, we employed a random forest with the combination of Synthetic Minority Oversampling Technique (SMOTE) and the Edited Nearest Neighbor (ENN) method (SMOTE+ENN), which is a resampling method to balance the activity class distribution. Within the NR and SR pathways, the activity of the aryl hydrocarbon receptor (NR-AhR) and the mitochondrial membrane potential (SR-MMP) were two of the top-performing twelve toxicity endpoints with AUCROCs of 0.90 and 0.92, respectively. The top extracted features for evaluating compound toxicity were analyzed for enrichment to highlight the implicated biological pathways and proteins. We validated our enrichment results for the activity of the AhR using a thorough literature search. Our case study showed that the selected enriched pathways and proteins from our computational pipeline are not only correlated with AhR toxicity but also form a cascading upstream/downstream arrangement. Our work elucidates significant relationships between protein and compound interactions computed using CANDO and the associated biological pathways to which the proteins belong for twelve toxicity endpoints. This novel study uses machine learning not only to predict and understand toxicity but also elucidates therapeutic mechanisms at a proteomic level for a variety of toxicity endpoints.
Assuntos
Aprendizado de Máquina , Proteômica , Algoritmos , Descoberta de Drogas/métodos , Humanos , ProteínasRESUMO
BACKGROUND: While artificial intelligence (AI) offers possibilities of advanced clinical prediction and decision-making in healthcare, models trained on relatively homogeneous datasets, and populations poorly-representative of underlying diversity, limits generalisability and risks biased AI-based decisions. Here, we describe the landscape of AI in clinical medicine to delineate population and data-source disparities. METHODS: We performed a scoping review of clinical papers published in PubMed in 2019 using AI techniques. We assessed differences in dataset country source, clinical specialty, and author nationality, sex, and expertise. A manually tagged subsample of PubMed articles was used to train a model, leveraging transfer-learning techniques (building upon an existing BioBERT model) to predict eligibility for inclusion (original, human, clinical AI literature). Of all eligible articles, database country source and clinical specialty were manually labelled. A BioBERT-based model predicted first/last author expertise. Author nationality was determined using corresponding affiliated institution information using Entrez Direct. And first/last author sex was evaluated using the Gendarize.io API. RESULTS: Our search yielded 30,576 articles, of which 7,314 (23.9%) were eligible for further analysis. Most databases came from the US (40.8%) and China (13.7%). Radiology was the most represented clinical specialty (40.4%), followed by pathology (9.1%). Authors were primarily from either China (24.0%) or the US (18.4%). First and last authors were predominately data experts (i.e., statisticians) (59.6% and 53.9% respectively) rather than clinicians. And the majority of first/last authors were male (74.1%). INTERPRETATION: U.S. and Chinese datasets and authors were disproportionately overrepresented in clinical AI, and almost all of the top 10 databases and author nationalities were from high income countries (HICs). AI techniques were most commonly employed for image-rich specialties, and authors were predominantly male, with non-clinical backgrounds. Development of technological infrastructure in data-poor regions, and diligence in external validation and model re-calibration prior to clinical implementation in the short-term, are crucial in ensuring clinical AI is meaningful for broader populations, and to avoid perpetuating global health inequity.