Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 55
Filtrar
1.
J Immunother Cancer ; 12(5)2024 05 15.
Artículo en Inglés | MEDLINE | ID: mdl-38749538

RESUMEN

BACKGROUND: Only a subset of patients with gastric cancer experience long-term benefits from immune checkpoint inhibitors (ICIs). Currently, there is a deficiency in precise predictive biomarkers for ICI efficacy. The aim of this study was to develop and validate a pathomics-driven ensemble model for predicting the response to ICIs in gastric cancer, using H&E-stained whole slide images (WSI). METHODS: This multicenter study retrospectively collected and analyzed H&E-stained WSIs and clinical data from 584 patients with gastric cancer. An ensemble model, integrating four classifiers: least absolute shrinkage and selection operator, k-nearest neighbors, decision trees, and random forests, was developed and validated using pathomics features, with the objective of predicting the therapeutic efficacy of immune checkpoint inhibition. Model performance was evaluated using metrics including the area under the curve (AUC), sensitivity, and specificity. Additionally, SHAP (SHapley Additive exPlanations) analysis was used to explain the model's predicted values as the sum of the attribution values for each input feature. Pathogenomics analysis was employed to explain the molecular mechanisms underlying the model's predictions. RESULTS: Our pathomics-driven ensemble model effectively stratified the response to ICIs in training cohort (AUC 0.985 (95% CI 0.971 to 0.999)), which was further validated in internal validation cohort (AUC 0.921 (95% CI 0.839 to 0.999)), as well as in external validation cohort 1 (AUC 0.914 (95% CI 0.837 to 0.990)), and external validation cohort 2 (0.927 (95% CI 0.802 to 0.999)). The univariate Cox regression analysis revealed that the prediction signature of pathomics-driven ensemble model was a prognostic factor for progression-free survival in patients with gastric cancer who underwent immunotherapy (p<0.001, HR 0.35 (95% CI 0.24 to 0.50)), and remained an independent predictor after multivariable Cox regression adjusted for clinicopathological variables, (including sex, age, carcinoembryonic antigen, carbohydrate antigen 19-9, therapy regime, line of therapy, differentiation, location and programmed death ligand 1 (PD-L1) expression in all patients (p<0.001, HR 0.34 (95% CI 0.24 to 0.50)). Pathogenomics analysis suggested that the ensemble model is driven by molecular-level immune, cancer, metabolism-related pathways, and was correlated with the immune-related characteristics, including immune score, Estimation of STromal and Immune cells in MAlignant Tumor tissues using Expression data score, and tumor purity. CONCLUSIONS: Our pathomics-driven ensemble model exhibited high accuracy and robustness in predicting the response to ICIs using WSIs. Therefore, it could serve as a novel and valuable tool to facilitate precision immunotherapy.


Asunto(s)
Inmunoterapia , Neoplasias Gástricas , Humanos , Neoplasias Gástricas/tratamiento farmacológico , Neoplasias Gástricas/inmunología , Neoplasias Gástricas/patología , Neoplasias Gástricas/terapia , Masculino , Femenino , Inmunoterapia/métodos , Estudios Retrospectivos , Persona de Mediana Edad , Inhibidores de Puntos de Control Inmunológico/uso terapéutico , Inhibidores de Puntos de Control Inmunológico/farmacología , Anciano
2.
Molecules ; 29(5)2024 Mar 03.
Artículo en Inglés | MEDLINE | ID: mdl-38474647

RESUMEN

A chemical study of Aesculus wilsonii Rehd. (also called Suo Luo Zi) and the in vitro anti-inflammatory effects of the obtained compounds was conducted. Retrieving results through SciFinder showed that there were four unreported compounds, aeswilosides I-IV (1-4), along with fourteen known isolates (5-18). Their structures were elucidated by extensive spectroscopic methods such as UV, IR, NMR, [α]D, and MS spectra, as well as acid hydrolysis. Among the known ones, compounds 5, 6, 8-10, and 12-16 were obtained from the Aesculus genus for the first time; compounds 7, 11, 17, and 18 were first identified from this plant. The NMR data of 5 and 18 were reported first. The effects of 1-18 on the release of nitric oxide (NO) from lipopolysaccharide (LPS)-induced RAW264.7 cells were determined. The results showed that at concentrations of 10, 25, and 50 µM, the novel compounds, aeswilosides I (1) and IV (4), along with the known ones, 1-(2-methylbutyryl)phloroglucinyl-glucopyranoside (10) and pisuminic acid (15), displayed significant inhibitory effects on NO production in a concentration-dependent manner. It is worth mentioning that compound 10 showed the best NO inhibitory effect with a relative NO production of 88.1%, which was close to that of the positive drug dexamethasone. The Elisa experiment suggested that compounds 1, 4, 10, and 15 suppressed the release of TNF-α and IL-1ß as well. In conclusion, this study enriches the spectra of compounds with potential anti-inflammatory effects in A. wilsonii and provides new references for the discovery of anti-inflammatory lead compounds, but further mechanistic research is still needed.


Asunto(s)
Aesculus , Ratones , Animales , Aesculus/química , Antiinflamatorios/farmacología , Células RAW 264.7 , Factor de Necrosis Tumoral alfa , Semillas/química , Lipopolisacáridos/farmacología , Óxido Nítrico/análisis
3.
IEEE Trans Med Imaging ; PP2024 Feb 12.
Artículo en Inglés | MEDLINE | ID: mdl-38345948

RESUMEN

Upon remarkable progress in cardiac image segmentation, contemporary studies dedicate to further upgrading model functionality toward perfection, through progressively exploring the sequentially delivered datasets over time by domain incremental learning. Existing works mainly concentrated on addressing the heterogeneous style variations, but overlooked the critical shape variations across domains hidden behind the sub-disease composition discrepancy. In case the updated model catastrophically forgets the sub-diseases that were learned in past domains but are no longer present in the subsequent domains, we proposed a dual enrichment synergistic strategy to incrementally broaden model competence for a growing number of sub-diseases. The data-enriched scheme aims to diversify the shape composition of current training data via displacement-aware shape encoding and decoding, to gradually build up the robustness against cross-domain shape variations. Meanwhile, the model-enriched scheme intends to strengthen model capabilities by progressively appending and consolidating the latest expertise into a dynamically-expanded multi-expert network, to gradually cultivate the generalization ability over style-variated domains. The above two schemes work in synergy to collaboratively upgrade model capabilities in two-pronged manners. We have extensively evaluated our network with the ACDC and M&Ms datasets in single-domain and compound-domain incremental learning settings. Our approach outperformed other competing methods and achieved comparable results to the upper bound.

4.
IEEE J Biomed Health Inform ; 28(4): 2115-2125, 2024 Apr.
Artículo en Inglés | MEDLINE | ID: mdl-38289846

RESUMEN

Masked image modeling (MIM) with transformer backbones has recently been exploited as a powerful self-supervised pre-training technique. The existing MIM methods adopt the strategy to mask random patches of the image and reconstruct the missing pixels, which only considers semantic information at a lower level, and causes a long pre-training time. This paper presents HybridMIM, a novel hybrid self-supervised learning method based on masked image modeling for 3D medical image segmentation. Specifically, we design a two-level masking hierarchy to specify which and how patches in sub-volumes are masked, effectively providing the constraints of higher level semantic information. Then we learn the semantic information of medical images at three levels, including: 1) partial region prediction to reconstruct key contents of the 3D image, which largely reduces the pre-training time burden (pixel-level); 2) patch-masking perception to learn the spatial relationship between the patches in each sub-volume (region-level); and 3) drop-out-based contrastive learning between samples within a mini-batch, which further improves the generalization ability of the framework (sample-level). The proposed framework is versatile to support both CNN and transformer as encoder backbones, and also enables to pre-train decoders for image segmentation. We conduct comprehensive experiments on five widely-used public medical image segmentation datasets, including BraTS2020, BTCV, MSD Liver, MSD Spleen, and BraTS2023. The experimental results show the clear superiority of HybridMIM against competing supervised methods, masked pre-training approaches, and other self-supervised methods, in terms of quantitative metrics, speed performance and qualitative observations.


Asunto(s)
Benchmarking , Automanejo , Humanos , Suministros de Energía Eléctrica , Hígado , Semántica , Procesamiento de Imagen Asistido por Computador
5.
J Clin Invest ; 134(6)2024 Jan 25.
Artículo en Inglés | MEDLINE | ID: mdl-38271117

RESUMEN

BACKGROUNDThe tumor immune microenvironment can provide prognostic and therapeutic information. We aimed to develop noninvasive imaging biomarkers from computed tomography (CT) for comprehensive evaluation of immune context and investigate their associations with prognosis and immunotherapy response in gastric cancer (GC).METHODSThis study involved 2,600 patients with GC from 9 independent cohorts. We developed and validated 2 CT imaging biomarkers (lymphoid radiomics score [LRS] and myeloid radiomics score [MRS]) for evaluating the IHC-derived lymphoid and myeloid immune context respectively, and integrated them into a combined imaging biomarker [LRS/MRS: low(-) or high(+)] with 4 radiomics immune subtypes: 1 (-/-), 2 (+/-), 3 (-/+), and 4 (+/+). We further evaluated the imaging biomarkers' predictive values on prognosis and immunotherapy response.RESULTSThe developed imaging biomarkers (LRS and MRS) had a high accuracy in predicting lymphoid (AUC range: 0.765-0.773) and myeloid (AUC range: 0.736-0.750) immune context. Further, similar to the IHC-derived immune context, 2 imaging biomarkers (HR range: 0.240-0.761 for LRS; 1.301-4.012 for MRS) and the combined biomarker were independent predictors for disease-free and overall survival in the training and all validation cohorts (all P < 0.05). Additionally, patients with high LRS or low MRS may benefit more from immunotherapy (P < 0.001). Further, a highly heterogeneous outcome on objective response ​rate was observed in 4 imaging subtypes: 1 (-/-) with 27.3%, 2 (+/-) with 53.3%, 3 (-/+) with 10.2%, and 4 (+/+) with 30.0% (P < 0.0001).CONCLUSIONThe noninvasive imaging biomarkers could accurately evaluate the immune context and provide information regarding prognosis and immunotherapy for GC.


Asunto(s)
Neoplasias Gástricas , Humanos , Neoplasias Gástricas/diagnóstico por imagen , Neoplasias Gástricas/terapia , Radiómica , Inmunoterapia , Tomografía Computarizada por Rayos X , Microambiente Tumoral , Biomarcadores , Pronóstico
6.
IEEE Trans Med Imaging ; 43(1): 149-161, 2024 Jan.
Artículo en Inglés | MEDLINE | ID: mdl-37436855

RESUMEN

Automatic universal lesion segmentation (ULS) from Computed Tomography (CT) images can ease the burden of radiologists and provide a more accurate assessment than the current Response Evaluation Criteria In Solid Tumors (RECIST) guideline measurement. However, this task is underdeveloped due to the absence of large-scale pixel-wise labeled data. This paper presents a weakly-supervised learning framework to utilize the large-scale existing lesion databases in hospital Picture Archiving and Communication Systems (PACS) for ULS. Unlike previous methods to construct pseudo surrogate masks for fully supervised training through shallow interactive segmentation techniques, we propose to unearth the implicit information from RECIST annotations and thus design a unified RECIST-induced reliable learning (RiRL) framework. Particularly, we introduce a novel label generation procedure and an on-the-fly soft label propagation strategy to avoid noisy training and poor generalization problems. The former, named RECIST-induced geometric labeling, uses clinical characteristics of RECIST to preliminarily and reliably propagate the label. With the labeling process, a trimap divides the lesion slices into three regions, including certain foreground, background, and unclear regions, which consequently enables a strong and reliable supervision signal on a wide region. A topological knowledge-driven graph is built to conduct the on-the-fly label propagation for the optimal segmentation boundary to further optimize the segmentation boundary. Experimental results on a public benchmark dataset demonstrate that the proposed method surpasses the SOTA RECIST-based ULS methods by a large margin. Our approach surpasses SOTA approaches over 2.0%, 1.5%, 1.4%, and 1.6% Dice with ResNet101, ResNet50, HRNet, and ResNest50 backbones.


Asunto(s)
Radiólogos , Columna Vertebral , Humanos , Criterios de Evaluación de Respuesta en Tumores Sólidos , Bases de Datos Factuales , Tomografía Computarizada por Rayos X , Procesamiento de Imagen Asistido por Computador , Aprendizaje Automático Supervisado
7.
Bioeng Transl Med ; 8(6): e10587, 2023 Nov.
Artículo en Inglés | MEDLINE | ID: mdl-38023695

RESUMEN

A novel recursive cascaded full-resolution residual network (RCFRR-Net) for abdominal four-dimensional computed tomography (4D-CT) image registration was proposed. The entire network was end-to-end and trained in the unsupervised approach, which meant that the deformation vector field, which presented the ground truth, was not needed during training. The network was designed by cascading three full-resolution residual subnetworks with different architectures. The training loss consisted of the image similarity loss and the deformation vector field regularization loss, which were calculated based on the final warped image and the fixed image, allowing all cascades to be trained jointly and perform the progressive registration cooperatively. Extensive network testing was conducted using diverse datasets, including an internal 4D-CT dataset, a public DIRLAB 4D-CT dataset, and a 4D cone-beam CT (4D-CBCT) dataset. Compared with the iteration-based demon method and two deep learning-based methods (VoxelMorph and recursive cascaded network), the RCFRR-Net achieved consistent and significant gains, which demonstrated that the proposed method had superior performance and generalization capability in medical image registration. The proposed RCFRR-Net was a promising tool for various clinical applications.

9.
IEEE Trans Image Process ; 32: 4036-4045, 2023.
Artículo en Inglés | MEDLINE | ID: mdl-37440404

RESUMEN

Transformer, the model of choice for natural language processing, has drawn scant attention from the medical imaging community. Given the ability to exploit long-term dependencies, transformers are promising to help atypical convolutional neural networks to learn more contextualized visual representations. However, most of recently proposed transformer-based segmentation approaches simply treated transformers as assisted modules to help encode global context into convolutional representations. To address this issue, we introduce nnFormer (i.e., not-another transFormer), a 3D transformer for volumetric medical image segmentation. nnFormer not only exploits the combination of interleaved convolution and self-attention operations, but also introduces local and global volume-based self-attention mechanism to learn volume representations. Moreover, nnFormer proposes to use skip attention to replace the traditional concatenation/summation operations in skip connections in U-Net like architecture. Experiments show that nnFormer significantly outperforms previous transformer-based counterparts by large margins on three public datasets. Compared to nnUNet, the most widely recognized convnet-based 3D medical segmentation model, nnFormer produces significantly lower HD95 and is much more computationally efficient. Furthermore, we show that nnFormer and nnUNet are highly complementary to each other in model ensembling. Codes and models of nnFormer are available at https://git.io/JSf3i.

10.
Radiol Artif Intell ; 5(3): e220082, 2023 May.
Artículo en Inglés | MEDLINE | ID: mdl-37293342

RESUMEN

Purpose: To investigate the correlation between differences in data distributions and federated deep learning (Fed-DL) algorithm performance in tumor segmentation on CT and MR images. Materials and Methods: Two Fed-DL datasets were retrospectively collected (from November 2020 to December 2021): one dataset of liver tumor CT images (Federated Imaging in Liver Tumor Segmentation [or, FILTS]; three sites, 692 scans) and one publicly available dataset of brain tumor MR images (Federated Tumor Segmentation [or, FeTS]; 23 sites, 1251 scans). Scans from both datasets were grouped according to site, tumor type, tumor size, dataset size, and tumor intensity. To quantify differences in data distributions, the following four distance metrics were calculated: earth mover's distance (EMD), Bhattacharyya distance (BD), χ2 distance (CSD), and Kolmogorov-Smirnov distance (KSD). Both federated and centralized nnU-Net models were trained by using the same grouped datasets. Fed-DL model performance was evaluated by using the ratio of Dice coefficients, θ, between federated and centralized models trained and tested on the same 80:20 split datasets. Results: The Dice coefficient ratio (θ) between federated and centralized models was strongly negatively correlated with the distances between data distributions, with correlation coefficients of -0.920 for EMD, -0.893 for BD, and -0.899 for CSD. However, KSD was weakly correlated with θ, with a correlation coefficient of -0.479. Conclusion: Performance of Fed-DL models in tumor segmentation on CT and MRI datasets was strongly negatively correlated with the distances between data distributions.Keywords: CT, Abdomen/GI, Liver, Comparative Studies, MR Imaging, Brain/Brain Stem, Convolutional Neural Network (CNN), Federated Deep Learning, Tumor Segmentation, Data Distribution Supplemental material is available for this article. © RSNA, 2023See also the commentary by Kwak and Bai in this issue.

11.
Int J Surg ; 109(7): 2010-2024, 2023 Jul 01.
Artículo en Inglés | MEDLINE | ID: mdl-37300884

RESUMEN

BACKGROUND: Peritoneal recurrence (PR) is the predominant pattern of relapse after curative-intent surgery in gastric cancer (GC) and indicates a dismal prognosis. Accurate prediction of PR is crucial for patient management and treatment. The authors aimed to develop a noninvasive imaging biomarker from computed tomography (CT) for PR evaluation, and investigate its associations with prognosis and chemotherapy benefit. METHODS: In this multicenter study including five independent cohorts of 2005 GC patients, the authors extracted 584 quantitative features from the intratumoral and peritumoral regions on contrast-enhanced CT images. The artificial intelligence algorithms were used to select significant PR-related features, and then integrated into a radiomic imaging signature. And improvements of diagnostic accuracy for PR by clinicians with the signature assistance were quantified. Using Shapley values, the authors determined the most relevant features and provided explanations to prediction. The authors further evaluated its predictive performance in prognosis and chemotherapy response. RESULTS: The developed radiomics signature had a consistently high accuracy in predicting PR in the training cohort (area under the curve: 0.732) and internal and Sun Yat-sen University Cancer Center validation cohorts (0.721 and 0.728). The radiomics signature was the most important feature in Shapley interpretation. The diagnostic accuracy of PR with the radiomics signature assistance was improved by 10.13-18.86% for clinicians ( P <0.001). Furthermore, it was also applicable in the survival prediction. In multivariable analysis, the radiomics signature remained an independent predictor for PR and prognosis ( P <0.001 for all). Importantly, patients with predicting high risk of PR from radiomics signature could gain survival benefit from adjuvant chemotherapy. By contrast, chemotherapy had no impact on survival for patients with a predicted low risk of PR. CONCLUSION: The noninvasive and explainable model developed from preoperative CT images could accurately predict PR and chemotherapy benefit in patients with GC, which will allow the optimization of individual decision-making.


Asunto(s)
Neoplasias Peritoneales , Neoplasias Gástricas , Humanos , Neoplasias Gástricas/diagnóstico por imagen , Neoplasias Gástricas/tratamiento farmacológico , Neoplasias Gástricas/cirugía , Inteligencia Artificial , Neoplasias Peritoneales/diagnóstico por imagen , Neoplasias Peritoneales/tratamiento farmacológico , Estudios Retrospectivos , Recurrencia Local de Neoplasia/diagnóstico por imagen , Gastrectomía
12.
IEEE Trans Med Imaging ; 42(11): 3374-3383, 2023 11.
Artículo en Inglés | MEDLINE | ID: mdl-37335798

RESUMEN

The fusion of multi-modal medical data is essential to assist medical experts to make treatment decisions for precision medicine. For example, combining the whole slide histopathological images (WSIs) and tabular clinical data can more accurately predict the lymph node metastasis (LNM) of papillary thyroid carcinoma before surgery to avoid unnecessary lymph node resection. However, the huge-sized WSI provides much more high-dimensional information than low-dimensional tabular clinical data, making the information alignment challenging in the multi-modal WSI analysis tasks. This paper presents a novel transformer-guided multi-modal multi-instance learning framework to predict lymph node metastasis from both WSIs and tabular clinical data. We first propose an effective multi-instance grouping scheme, named siamese attention-based feature grouping (SAG), to group high-dimensional WSIs into representative low-dimensional feature embeddings for fusion. We then design a novel bottleneck shared-specific feature transfer module (BSFT) to explore the shared and specific features between different modalities, where a few learnable bottleneck tokens are utilized for knowledge transfer between modalities. Moreover, a modal adaptation and orthogonal projection scheme were incorporated to further encourage BSFT to learn shared and specific features from multi-modal data. Finally, the shared and specific features are dynamically aggregated via an attention mechanism for slide-level prediction. Experimental results on our collected lymph node metastasis dataset demonstrate the efficiency of our proposed components and our framework achieves the best performance with AUC (area under the curve) of 97.34%, outperforming the state-of-the-art methods by over 1.27%.


Asunto(s)
Procesamiento de Imagen Asistido por Computador , Medicina de Precisión , Humanos , Metástasis Linfática/diagnóstico por imagen , Biopsia
13.
IEEE Trans Pattern Anal Mach Intell ; 45(11): 13408-13421, 2023 11.
Artículo en Inglés | MEDLINE | ID: mdl-37363838

RESUMEN

Defining the loss function is an important part of neural network design and critically determines the success of deep learning modeling. A significant shortcoming of the conventional loss functions is that they weight all regions in the input image volume equally, despite the fact that the system is known to be heterogeneous (i.e., some regions can achieve high prediction performance more easily than others). Here, we introduce a region-specific loss to lift the implicit assumption of homogeneous weighting for better learning. We divide the entire volume into multiple sub-regions, each with an individualized loss constructed for optimal local performance. Effectively, this scheme imposes higher weightings on the sub-regions that are more difficult to segment, and vice versa. Furthermore, the regional false positive and false negative errors are computed for each input image during a training step and the regional penalty is adjusted accordingly to enhance the overall accuracy of the prediction. Using different public and in-house medical image datasets, we demonstrate that the proposed regionally adaptive loss paradigm outperforms conventional methods in the multi-organ segmentations, without any modification to the neural network architecture or additional data preparation.


Asunto(s)
Algoritmos , Redes Neurales de la Computación , Procesamiento de Imagen Asistido por Computador/métodos
14.
Int J Radiat Oncol Biol Phys ; 117(2): 505-514, 2023 10 01.
Artículo en Inglés | MEDLINE | ID: mdl-37141982

RESUMEN

PURPOSE: This study explored deep-learning-based patient-specific auto-segmentation using transfer learning on daily RefleXion kilovoltage computed tomography (kVCT) images to facilitate adaptive radiation therapy, based on data from the first group of patients treated with the innovative RefleXion system. METHODS AND MATERIALS: For head and neck (HaN) and pelvic cancers, a deep convolutional segmentation network was initially trained on a population data set that contained 67 and 56 patient cases, respectively. Then the pretrained population network was adapted to the specific RefleXion patient by fine-tuning the network weights with a transfer learning method. For each of the 6 collected RefleXion HaN cases and 4 pelvic cases, initial planning computed tomography (CT) scans and 5 to 26 sets of daily kVCT images were used for the patient-specific learning and evaluation separately. The performance of the patient-specific network was compared with the population network and the clinical rigid registration method and evaluated by the Dice similarity coefficient (DSC) with manual contours being the reference. The corresponding dosimetric effects resulting from different auto-segmentation and registration methods were also investigated. RESULTS: The proposed patient-specific network achieved mean DSC results of 0.88 for 3 HaN organs at risk (OARs) of interest and 0.90 for 8 pelvic target and OARs, outperforming the population network (0.70 and 0.63) and the registration method (0.72 and 0.72). The DSC of the patient-specific network gradually increased with the increment of longitudinal training cases and approached saturation with more than 6 training cases. Compared with using the registration contour, the target and OAR mean doses and dose-volume histograms obtained using the patient-specific auto-segmentation were closer to the results using the manual contour. CONCLUSIONS: Auto-segmentation of RefleXion kVCT images based on the patient-specific transfer learning could achieve higher accuracy, outperforming a common population network and clinical registration-based method. This approach shows promise in improving dose evaluation accuracy in RefleXion adaptive radiation therapy.


Asunto(s)
Procesamiento de Imagen Asistido por Computador , Planificación de la Radioterapia Asistida por Computador , Humanos , Planificación de la Radioterapia Asistida por Computador/métodos , Procesamiento de Imagen Asistido por Computador/métodos , Órganos en Riesgo/diagnóstico por imagen , Órganos en Riesgo/efectos de la radiación , Radiometría , Tomografía Computarizada por Rayos X
15.
IEEE J Biomed Health Inform ; 27(7): 3258-3269, 2023 Jul.
Artículo en Inglés | MEDLINE | ID: mdl-37099476

RESUMEN

Anatomical resection (AR) based on anatomical sub-regions is a promising method of precise surgical resection, which has been proven to improve long-term survival by reducing local recurrence. The fine-grained segmentation of an organ's surgical anatomy (FGS-OSA), i.e., segmenting an organ into multiple anatomic regions, is critical for localizing tumors in AR surgical planning. However, automatically obtaining FGS-OSA results in computer-aided methods faces the challenges of appearance ambiguities among sub-regions (i.e., inter-sub-region appearance ambiguities) caused by similar HU distributions in different sub-regions of an organ's surgical anatomy, invisible boundaries, and similarities between anatomical landmarks and other anatomical information. In this paper, we propose a novel fine-grained segmentation framework termed the "anatomic relation reasoning graph convolutional network" (ARR-GCN), which incorporates prior anatomic relations into the framework learning. In ARR-GCN, a graph is constructed based on the sub-regions to model the class and their relations. Further, to obtain discriminative initial node representations of graph space, a sub-region center module is designed. Most importantly, to explicitly learn the anatomic relations, the prior anatomic-relations among the sub-regions are encoded in the form of an adjacency matrix and embedded into the intermediate node representations to guide framework learning. The ARR-GCN was validated on two FGS-OSA tasks: i) liver segments segmentation, and ii) lung lobes segmentation. Experimental results on both tasks outperformed other state-of-the-art segmentation methods and yielded promising performances by ARR-GCN for suppressing ambiguities among sub-regions.


Asunto(s)
Hígado , Humanos , Hígado/anatomía & histología , Hígado/diagnóstico por imagen , Hígado/cirugía , Neoplasias
16.
IEEE Trans Med Imaging ; 42(8): 2462-2473, 2023 08.
Artículo en Inglés | MEDLINE | ID: mdl-37028064

RESUMEN

Cancer survival prediction requires exploiting related multimodal information (e.g., pathological, clinical and genomic features, etc.) and it is even more challenging in clinical practices due to the incompleteness of patient's multimodal data. Furthermore, existing methods lack sufficient intra- and inter-modal interactions, and suffer from significant performance degradation caused by missing modalities. This manuscript proposes a novel hybrid graph convolutional network, entitled HGCN, which is equipped with an online masked autoencoder paradigm for robust multimodal cancer survival prediction. Particularly, we pioneer modeling the patient's multimodal data into flexible and interpretable multimodal graphs with modality-specific preprocessing. HGCN integrates the advantages of graph convolutional networks (GCNs) and a hypergraph convolutional network (HCN) through node message passing and a hyperedge mixing mechanism to facilitate intra-modal and inter-modal interactions between multimodal graphs. With HGCN, the potential for multimodal data to create more reliable predictions of patient's survival risk is dramatically increased compared to prior methods. Most importantly, to compensate for missing patient modalities in clinical scenarios, we incorporated an online masked autoencoder paradigm into HGCN, which can effectively capture intrinsic dependence between modalities and seamlessly generate missing hyperedges for model inference. Extensive experiments and analysis on six cancer cohorts from TCGA show that our method significantly outperforms the state-of-the-arts in both complete and missing modal settings. Our codes are made available at https://github.com/lin-lcx/HGCN.


Asunto(s)
Genómica , Neoplasias , Humanos , Neoplasias/diagnóstico por imagen
17.
IEEE Trans Med Imaging ; 42(5): 1337-1348, 2023 05.
Artículo en Inglés | MEDLINE | ID: mdl-37015475

RESUMEN

Multi-instance learning (MIL) is widely adop- ted for automatic whole slide image (WSI) analysis and it usually consists of two stages, i.e., instance feature extraction and feature aggregation. However, due to the "weak supervision" of slide-level labels, the feature aggregation stage would suffer from severe over-fitting in training an effective MIL model. In this case, mining more information from limited slide-level data is pivotal to WSI analysis. Different from previous works on improving instance feature extraction, this paper investigates how to exploit the latent relationship of different instances (patches) to combat overfitting in MIL for more generalizable WSI classification. In particular, we propose a novel Multi-instance Rein- forcement Contrastive Learning framework (MuRCL) to deeply mine the inherent semantic relationships of different patches to advance WSI classification. Specifically, the proposed framework is first trained in a self-supervised manner and then finetuned with WSI slide-level labels. We formulate the first stage as a contrastive learning (CL) process, where positive/negative discriminative feature sets are constructed from the same patch-level feature bags of WSIs. To facilitate the CL training, we design a novel reinforcement learning-based agent to progressively update the selection of discriminative feature sets according to an online reward for slide-level feature aggregation. Then, we further update the model with labeled WSI data to regularize the learned features for the final WSI classification. Experimental results on three public WSI classification datasets (Camelyon16, TCGA-Lung and TCGA-Kidney) demonstrate that the proposed MuRCL outperforms state-of-the-art MIL models. In addition, MuRCL can achieve comparable performance to other state-of-the-art MIL models on TCGA-Esca dataset.


Asunto(s)
Procesamiento de Imagen Asistido por Computador , Aprendizaje Automático Supervisado , Humanos , Conjuntos de Datos como Asunto , Pulmón/diagnóstico por imagen , Riñón/diagnóstico por imagen
18.
IEEE Trans Med Imaging ; 42(5): 1431-1445, 2023 05.
Artículo en Inglés | MEDLINE | ID: mdl-37015694

RESUMEN

Collecting sufficient high-quality training data for deep neural networks is often expensive or even unaffordable in medical image segmentation tasks. We thus propose to train the network by using external data that can be collected in a cheaper way, e.g., crowd-sourcing. We show that by data discernment, the network is able to mine valuable knowledge from external data, even though the data distribution is very different from that of the original (internal) data. We discern the external data by learning an importance weight for each of them, with the goal to enhance the contribution of informative external data to network updating, while suppressing the data that are 'useless' or even 'harmful'. An iterative algorithm that alternatively estimates the importance weight and updates the network is developed by formulating the data discernment as a constrained nonlinear programming problem. It estimates the importance weight according to the distribution discrepancy between the external data and the internal dataset, and imposes a constraint to drive the network to learn more effectively, compared with the network without using the external data. We evaluate the proposed algorithm on two tasks: abdominal CT image and cervical smear image segmentation, using totally 6 publicly available datasets. The effectiveness of the algorithm is demonstrated by extensive experiments. Source codes are available at: https://github.com/YouyiSong/Data-Discernment.


Asunto(s)
Algoritmos , Colaboración de las Masas , Redes Neurales de la Computación , Programas Informáticos , Procesamiento de Imagen Asistido por Computador
19.
IEEE Trans Pattern Anal Mach Intell ; 45(5): 5988-6005, 2023 May.
Artículo en Inglés | MEDLINE | ID: mdl-36094969

RESUMEN

Co-occurrent visual pattern makes context aggregation become an essential paradigm for semantic segmentation. The existing studies focus on modeling the contexts within image while neglecting the valuable semantics of the corresponding category beyond image. To this end, we propose a novel soft mining contextual information beyond image paradigm named MCIBI++ to further boost the pixel-level representations. Specifically, we first set up a dynamically updated memory module to store the dataset-level distribution information of various categories and then leverage the information to yield the dataset-level category representations during network forward. After that, we generate a class probability distribution for each pixel representation and conduct the dataset-level context aggregation with the class probability distribution as weights. Finally, the original pixel representations are augmented with the aggregated dataset-level and the conventional image-level contextual information. Moreover, in the inference phase, we additionally design a coarse-to-fine iterative inference strategy to further boost the segmentation results. MCIBI++ can be effortlessly incorporated into the existing segmentation frameworks and bring consistent performance improvements. Also, MCIBI++ can be extended into the video semantic segmentation framework with considerable improvements over the baseline. Equipped with MCIBI++, we achieved the state-of-the-art performance on seven challenging image or video semantic segmentation benchmarks.

20.
IEEE Trans Med Imaging ; 42(3): 570-581, 2023 03.
Artículo en Inglés | MEDLINE | ID: mdl-36191115

RESUMEN

Contemporary methods have shown promising results on cardiac image segmentation, but merely in static learning, i.e., optimizing the network once for all, ignoring potential needs for model updating. In real-world scenarios, new data continues to be gathered from multiple institutions over time and new demands keep growing to pursue more satisfying performance. The desired model should incrementally learn from each incoming dataset and progressively update with improved functionality as time goes by. As the datasets sequentially delivered from multiple sites are normally heterogenous with domain discrepancy, each updated model should not catastrophically forget previously learned domains while well generalizing to currently arrived domains or even unseen domains. In medical scenarios, this is particularly challenging as accessing or storing past data is commonly not allowed due to data privacy. To this end, we propose a novel domain-incremental learning framework to recover past domain inputs first and then regularly replay them during model optimization. Particularly, we first present a style-oriented replay module to enable structure-realistic and memory-efficient reproduction of past data, and then incorporate the replayed past data to jointly optimize the model with current data to alleviate catastrophic forgetting. During optimization, we additionally perform domain-sensitive feature whitening to suppress model's dependency on features that are sensitive to domain changes (e.g., domain-distinctive style features) to assist domain-invariant feature exploration and gradually improve the generalization performance of the network. We have extensively evaluated our approach with the M&Ms Dataset in single-domain and compound-domain incremental learning settings. Our approach outperforms other comparison methods with less forgetting on past domains and better generalization on current domains and unseen domains.


Asunto(s)
Corazón , Procesamiento de Imagen Asistido por Computador , Corazón/diagnóstico por imagen
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA