RESUMEN
OBJECTIVE: Automated methods for prostate segmentation on MRI are typically developed under ideal scanning and anatomical conditions. This study evaluates three different prostate segmentation AI algorithms in a challenging population of patients with prior treatments, variable anatomic characteristics, complex clinical history, or atypical MRI acquisition parameters. MATERIALS AND METHODS: A single institution retrospective database was queried for the following conditions at prostate MRI: prior prostate-specific oncologic treatment, transurethral resection of the prostate (TURP), abdominal perineal resection (APR), hip prosthesis (HP), diversity of prostate volumes (large ≥ 150 cc, small ≤ 25 cc), whole gland tumor burden, magnet strength, noted poor quality, and various scanners (outside/vendors). Final inclusion criteria required availability of axial T2-weighted (T2W) sequence and corresponding prostate organ segmentation from an expert radiologist. Three previously developed algorithms were evaluated: (1) deep learning (DL)-based model, (2) commercially available shape-based model, and (3) federated DL-based model. Dice Similarity Coefficient (DSC) was calculated compared to expert. DSC by model and scan factors were evaluated with Wilcox signed-rank test and linear mixed effects (LMER) model. RESULTS: 683 scans (651 patients) met inclusion criteria (mean prostate volume 60.1 cc [9.05-329 cc]). Overall DSC scores for models 1, 2, and 3 were 0.916 (0.707-0.971), 0.873 (0-0.997), and 0.894 (0.025-0.961), respectively, with DL-based models demonstrating significantly higher performance (p < 0.01). In sub-group analysis by factors, Model 1 outperformed Model 2 (all p < 0.05) and Model 3 (all p < 0.001). Performance of all models was negatively impacted by prostate volume and poor signal quality (p < 0.01). Shape-based factors influenced DL models (p < 0.001) while signal factors influenced all (p < 0.001). CONCLUSION: Factors affecting anatomical and signal conditions of the prostate gland can adversely impact both DL and non-deep learning-based segmentation models.
Asunto(s)
Algoritmos , Inteligencia Artificial , Imagen por Resonancia Magnética , Neoplasias de la Próstata , Humanos , Masculino , Estudios Retrospectivos , Imagen por Resonancia Magnética/métodos , Neoplasias de la Próstata/diagnóstico por imagen , Neoplasias de la Próstata/cirugía , Neoplasias de la Próstata/patología , Interpretación de Imagen Asistida por Computador/métodos , Persona de Mediana Edad , Anciano , Próstata/diagnóstico por imagen , Aprendizaje ProfundoRESUMEN
INTRODUCTION: Clinical documentation is an essential component of the provision of medical care, enabling continuity of information across provider and site handoffs. This is particularly important in the combat casualty care setting when a single casualty may be treated by four or more or five completely disparate teams across the roles of care. The Battlefield Assisted Trauma Distributed Observation Kit (BATDOK) is a digital battlefield clinical documentation system developed by the Air Force Research Laboratory to address this need. To support the deployment of this tool, we integrated BATDOK into a commercially available virtual reality (VR) medical simulation platform used by the U.S. Air Force and Defense Health Agency personnel in order to provide an immersive simulation training experience which included battlefield documentation. METHODS: A multidisciplinary team consisting of medical educators, VR simulation engineers, emergency physicians and pararescuemen, and BATDOK developers first developed a specification for a virtual BATDOK capability, including a detailed listing of learning objectives, critical interfaces and task plans, and sensor integrations. These specifications were then implemented into the commercially available Virtual Advancement of Learning for Operational Readiness VR Medical Simulation System and underwent developmental testing and evaluation during pararescueman training exercises at the Air Force Special Operations Command Special Operations Center for Medical Integration and Development. RESULTS AND CONCLUSIONS: The BATDOK capability was successfully implemented within the VR Medical Simulation System. The capability consisted of a virtual tablet with replicated interfaces and capabilities based on the developed specifications. These capabilities included integrated point-of-care ultrasound capability, multi-patient management, vitals sign monitoring with sensor pairing and continuous monitoring, mechanism of injury documentation (including injury pattern documentation), intervention logging (including tourniquets, dressing, airways, lines, tubes and drains, splints, fluids, and medications), and event logging. The capability was found to be operational and in alignment with learning objectives and user acceptance goals.
Asunto(s)
Entrenamiento Simulado , Realidad Virtual , Humanos , Simulación por Computador , TorniquetesRESUMEN
PURPOSE: Developing large-scale datasets with research-quality annotations is challenging due to the high cost of refining clinically generated markup into high precision annotations. We evaluated the direct use of a large dataset with only clinically generated annotations in development of high-performance segmentation models for small research-quality challenge datasets. MATERIALS AND METHODS: We used a large retrospective dataset from our institution comprised of 1,620 clinically generated segmentations, and two challenge datasets (PROMISE12: 50 patients, ProstateX-2: 99 patients). We trained a 3D U-Net convolutional neural network (CNN) segmentation model using our entire dataset, and used that model as a template to train models on the challenge datasets. We also trained versions of the template model using ablated proportions of our dataset, and evaluated the relative benefit of those templates for the final models. Finally, we trained a version of the template model using an out-of-domain brain cancer dataset, and evaluated the relevant benefit of that template for the final models. We used five-fold cross-validation (CV) for all training and evaluation across our entire dataset. RESULTS: Our model achieves state-of-the-art performance on our large dataset (mean overall Dice 0.916, average Hausdorff distance 0.135 across CV folds). Using this model as a pre-trained template for refining on two external datasets significantly enhanced performance (30% and 49% enhancement in Dice scores respectively). Mean overall Dice and mean average Hausdorff distance were 0.912 and 0.15 for the ProstateX-2 dataset, and 0.852 and 0.581 for the PROMISE12 dataset. Using even small quantities of data to train the template enhanced performance, with significant improvements using 5% or more of the data. CONCLUSION: We trained a state-of-the-art model using unrefined clinical prostate annotations and found that its use as a template model significantly improved performance in other prostate segmentation tasks, even when trained with only 5% of the original dataset.
Asunto(s)
Curaduría de Datos , Bases de Datos Factuales , Aprendizaje Profundo , Próstata/diagnóstico por imagen , Tomografía Computarizada por Rayos X , Humanos , Masculino , Estudios RetrospectivosRESUMEN
PURPOSE: The appropriate number of systematic biopsy cores to retrieve during magnetic resonance imaging (MRI)-targeted prostate biopsy is not well defined. We aimed to demonstrate a biopsy sampling approach that reduces required core count while maintaining diagnostic performance. MATERIALS AND METHODS: We collected data from a cohort of 971 men who underwent MRI-ultrasound fusion targeted biopsy for suspected prostate cancer. A regional targeted biopsy (RTB) was evaluated retrospectively; only cores within 2 cm of the margin of a radiologist-defined region of interest were considered part of the RTB. We compared detection rates for clinically significant prostate cancer (csPCa) and cancer upgrading rate on final whole mount pathology after prostatectomy between RTB, combined, MRI-targeted, and systematic biopsy. RESULTS: A total of 16,459 total cores from 971 men were included in the study data sets, of which 1,535 (9%) contained csPCa. The csPCa detection rates for systematic, MRI-targeted, combined, and RTB were 27.0% (262/971), 38.3% (372/971), 44.8% (435/971), and 44.0% (427/971), respectively. Combined biopsy detected significantly more csPCa than systematic and MRI-targeted biopsy (p <0.001 and p=0.004, respectively) but was similar to RTB (p=0.71), which used on average 3.8 (22%) fewer cores per patient. In 102 patients who underwent prostatectomy, there was no significant difference in upgrading rates between RTB and combined biopsy (p=0.84). CONCLUSIONS: A RTB approach can maintain state-of-the-art detection rates while requiring fewer retrieved cores. This result informs decision making about biopsy site selection and total retrieved core count.
Asunto(s)
Imagen Multimodal/métodos , Próstata/patología , Prostatectomía/estadística & datos numéricos , Neoplasias de la Próstata/diagnóstico , Anciano , Biopsia con Aguja Gruesa/métodos , Biopsia con Aguja Gruesa/estadística & datos numéricos , Conjuntos de Datos como Asunto , Estudios de Factibilidad , Humanos , Biopsia Guiada por Imagen/métodos , Biopsia Guiada por Imagen/estadística & datos numéricos , Imagen por Resonancia Magnética Intervencional/métodos , Imagen por Resonancia Magnética Intervencional/estadística & datos numéricos , Masculino , Persona de Mediana Edad , Imagen Multimodal/estadística & datos numéricos , Imágenes de Resonancia Magnética Multiparamétrica/estadística & datos numéricos , Clasificación del Tumor , Próstata/diagnóstico por imagen , Próstata/cirugía , Neoplasias de la Próstata/patología , Neoplasias de la Próstata/cirugía , Estudios Retrospectivos , Análisis Espacial , Ultrasonografía Intervencional/estadística & datos numéricosRESUMEN
OBJECTIVE: To demonstrate enabling multi-institutional training without centralizing or sharing the underlying physical data via federated learning (FL). MATERIALS AND METHODS: Deep learning models were trained at each participating institution using local clinical data, and an additional model was trained using FL across all of the institutions. RESULTS: We found that the FL model exhibited superior performance and generalizability to the models trained at single institutions, with an overall performance level that was significantly better than that of any of the institutional models alone when evaluated on held-out test sets from each institution and an outside challenge dataset. DISCUSSION: The power of FL was successfully demonstrated across 3 academic institutions while avoiding the privacy risk associated with the transfer and pooling of patient data. CONCLUSION: Federated learning is an effective methodology that merits further study to enable accelerated development of models across institutions, enabling greater generalizability in clinical use.
Asunto(s)
Aprendizaje Profundo , Difusión de la Información , Humanos , PrivacidadRESUMEN
Information technologies enable programmers and engineers to design and synthesize systems of startling complexity that nonetheless behave as intended. This mastery of complexity is made possible by a hierarchy of formal abstractions that span from high-level programming languages down to low-level implementation specifications, with rigorous connections between the levels. DNA nanotechnology presents us with a new molecular information technology whose potential has not yet been fully unlocked in this way. Developing an effective hierarchy of abstractions may be critical for increasing the complexity of programmable DNA systems. Here, we build on prior practice to provide a new formalization of 'domain-level' representations of DNA strand displacement systems that has a natural connection to nucleic acid biophysics while still being suitable for formal analysis. Enumeration of unimolecular and bimolecular reactions provides a semantics for programmable molecular interactions, with kinetics given by an approximate biophysical model. Reaction condensation provides a tractable simplification of the detailed reactions that respects overall kinetic properties. The applicability and accuracy of the model is evaluated across a wide range of engineered DNA strand displacement systems. Thus, our work can serve as an interface between lower-level DNA models that operate at the nucleotide sequence level, and high-level chemical reaction network models that operate at the level of interactions between abstract species.
Asunto(s)
ADN , Nanotecnología , Fenómenos Biofísicos , Cinética , Lenguajes de ProgramaciónRESUMEN
Purpose: Prostate cancer (PCa) is the most common solid organ cancer and second leading cause of death in men. Multiparametric magnetic resonance imaging (mpMRI) enables detection of the most aggressive, clinically significant PCa (csPCa) tumors that require further treatment. A suspicious region of interest (ROI) detected on mpMRI is now assigned a Prostate Imaging-Reporting and Data System (PIRADS) score to standardize interpretation of mpMRI for PCa detection. However, there is significant inter-reader variability among radiologists in PIRADS score assignment and a minimal input semi-automated artificial intelligence (AI) system is proposed to harmonize PIRADS scores with mpMRI data. Approach: The proposed deep learning model (the seed point model) uses a simulated single-click seed point as input to annotate the lesion on mpMRI. This approach is in contrast to typical medical AI-based approaches that require annotation of the complete lesion. The mpMRI data from 617 patients used in this study were prospectively collected at a major tertiary U.S. medical center. The model was trained and validated to classify whether an mpMRI image had a lesion with a PIRADS score greater than or equal to PIRADS 4. Results: The model yielded an average receiver-operator characteristic (ROC) area under the curve (ROC-AUC) of 0.704 over a 10-fold cross-validation, which is significantly higher than the previously published benchmark. Conclusions: The proposed model could aid in PIRADS scoring of mpMRI, providing second reads to promote quality as well as offering expertise in environments that lack a radiologist with training in prostate mpMRI interpretation. The model could help identify tumors with a higher PIRADS for better clinical management and treatment of PCa patients at an early stage.
RESUMEN
Social media presents a rich opportunity to gather health information with limited intervention through the analysis of completely unstructured and unlabeled microposts. We sought to estimate the health-related quality of life (HRQOL) of Twitter users using automated semantic processing methods. We collected tweets from 878 Twitter users recruited through online solicitation and in-person contact with patients. All participants completed the four-item Centers for Disease Control Healthy Days Questionnaire at the time of enrollment and 30 days later to measure "ground truth" HRQOL. We used a combination of document frequency analysis, sentiment analysis, topic analysis, and concept mapping to extract features from tweets, which we then used to estimate dichotomized HRQOL ("high" vs. "low") using logistic regression. Binary HRQOL status was estimated with moderate performance (AUC = 0.64). This result indicates that free-range social media data only offers a window into HRQOL, but does not afford direct access to current health status.
Asunto(s)
Medios de Comunicación Sociales , Recolección de Datos , Estado de Salud , Humanos , Calidad de Vida , SemánticaRESUMEN
Predicting infarct volume from magnetic resonance perfusion-weighted imaging can provide helpful information to clinicians in deciding how aggressively to treat acute stroke patients. Models have been developed to predict tissue fate, yet these models are mostly built using hand-crafted features (e.g., time-to-maximum) derived from perfusion images, which are sensitive to deconvolution methods. We demonstrate the application of deep convolution neural networks (CNNs) on predicting final stroke infarct volume using only the source perfusion images. We propose a deep CNN architecture that improves feature learning and achieves an area under the curve of 0.871 ± 0.024 , outperforming existing tissue fate models. We further validate the proposed deep CNN with existing 2-D and 3-D deep CNNs for images/video classification, showing the importance of the proposed architecture. Our work leverages deep learning techniques in stroke tissue outcome prediction, advancing magnetic resonance imaging perfusion analysis one step closer to an operational decision support tool for stroke treatment guidance.
RESUMEN
Prostate cancer is the most common and second most deadly form of cancer in men in the United States. The classification of prostate cancers based on Gleason grading using histological images is important in risk assessment and treatment planning for patients. Here, we demonstrate a new region-based convolutional neural network framework for multi-task prediction using an epithelial network head and a grading network head. Compared with a single-task model, our multi-task model can provide complementary contextual information, which contributes to better performance. Our model is achieved a state-of-the-art performance in epithelial cells detection and Gleason grading tasks simultaneously. Using fivefold cross-validation, our model is achieved an epithelial cells detection accuracy of 99.07% with an average area under the curve of 0.998. As for Gleason grading, our model is obtained a mean intersection over union of 79.56% and an overall pixel accuracy of 89.40%.
Asunto(s)
Interpretación de Imagen Asistida por Computador/métodos , Clasificación del Tumor/métodos , Redes Neurales de la Computación , Próstata/diagnóstico por imagen , Neoplasias de la Próstata/diagnóstico por imagen , Histocitoquímica , Humanos , MasculinoRESUMEN
Automated Gleason grading is an important preliminary step for quantitative histopathological feature extraction. Different from the traditional task of classifying small pre-selected homogeneous regions, semantic segmentation provides pixel-wise Gleason predictions across an entire slide. Deep learning-based segmentation models can automatically learn visual semantics from data, which alleviates the need for feature engineering. However, performance of deep learning models is limited by the scarcity of large-scale fully annotated datasets, which can be both expensive and time-consuming to create. One way to address this problem is to leverage external weakly labeled datasets to augment models trained on the limited data. In this paper, we developed an expectation maximization-based approach constrained by an approximated prior distribution in order to extract useful representations from a large number of weakly labeled images generated from low-magnification annotations. This method was utilized to improve the performance of a model trained on a limited fully annotated dataset. Our semi-supervised approach trained with 135 fully annotated and 1800 weakly annotated tiles achieved a mean Jaccard Index of 49.5% on an independent test set, which was 14% higher than the initial model trained only on the fully annotated dataset.
Asunto(s)
Aprendizaje Profundo , Procesamiento de Imagen Asistido por Computador/métodos , Próstata/diagnóstico por imagen , Próstata/patología , Prostatectomía , Aprendizaje Automático Supervisado , Algoritmos , Humanos , Masculino , Neoplasias de la PróstataRESUMEN
OBJECTIVE: It is crucial for clinicians to stay up to date on current literature in order to apply recent evidence to clinical decision making. Automatic summarization systems can help clinicians quickly view an aggregated summary of literature on a topic. Casama, a representation and summarization system based on "contextualized semantic maps," captures the findings of biomedical studies as well as the contexts associated with patient population and study design. This paper presents a user-oriented evaluation of Casama in comparison to a context-free representation, SemRep. MATERIALS AND METHODS: The effectiveness of the representation was evaluated by presenting users with manually annotated Casama and SemRep summaries of ten articles on driver mutations in cancer. Automatic annotations were evaluated on a collection of articles on EGFR mutation in lung cancer. Seven users completed a questionnaire rating the summarization quality for various topics and applications. RESULTS: Casama had higher median scores than SemRep for the majority of the topics (p≤ 0.00032), all of the applications (p≤ 0.00089), and in overall summarization quality (p≤ 1.5e-05). Casama's manual annotations outperformed Casama's automatic annotations (p = 0.00061). DISCUSSION: Casama performed particularly well in the representation of strength of evidence, which was highly rated both quantitatively and qualitatively. Users noted that Casama's less granular, more targeted representation improved usability compared to SemRep. CONCLUSION: This evaluation demonstrated the benefits of a contextualized representation for summarizing biomedical literature on cancer. Iteration on specific areas of Casama's representation, further development of its algorithms, and a clinically-oriented evaluation are warranted.
Asunto(s)
Curaduría de Datos/métodos , Toma de Decisiones Asistida por Computador , Semántica , Biología Computacional , Humanos , Neoplasias Pulmonares/genética , Neoplasias Pulmonares/terapia , Mutación/genéticaRESUMEN
Gleason grading of histological images is important in risk assessment and treatment planning for prostate cancer patients. Much research has been done in classifying small homogeneous cancer regions within histological images. However, semi-supervised methods published to date depend on pre-selected regions and cannot be easily extended to an image of heterogeneous tissue composition. In this paper, we propose a multi-scale U-Net model to classify images at the pixel-level using 224 histological image tiles from radical prostatectomies of 20 patients. Our model was evaluated by a patient-based 10-fold cross validation, and achieved a mean Jaccard index of 65.8% across 4 classes (stroma, Gleason 3, Gleason 4 and benign glands), and 75.5% for 3 classes (stroma, benign glands, prostate cancer), outperforming other methods.
Asunto(s)
Diagnóstico por Computador , Aprendizaje Automático , Clasificación del Tumor , Prostatectomía , Neoplasias de la Próstata/patología , Aprendizaje Profundo , Humanos , Masculino , Próstata/patología , Neoplasias de la Próstata/cirugía , Medición de Riesgo , Semántica , Máquina de Vectores de SoporteRESUMEN
Despite the HIV "test-and-treat" strategy's promise, questions about its clinical rationale, operational feasibility, and ethical appropriateness have led to vigorous debate in the global HIV community. We performed a systematic review of the literature published between January 2009 and May 2012 using PubMed, SCOPUS, Global Health, Web of Science, BIOSIS, Cochrane CENTRAL, EBSCO Africa-Wide Information, and EBSCO CINAHL Plus databases to summarize clinical uncertainties, health service challenges, and ethical complexities that may affect the test-and-treat strategy's success. A thoughtful approach to research and implementation to address clinical and health service questions and meaningful community engagement regarding ethical complexities may bring us closer to safe, feasible, and effective test-and-treat implementation.