Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 5 de 5
Filter
Add more filters











Database
Language
Publication year range
1.
Radiat Oncol ; 19(1): 96, 2024 Jul 29.
Article in English | MEDLINE | ID: mdl-39080735

ABSTRACT

BACKGROUND: In this work, we compare input level, feature level and decision level data fusion techniques for automatic detection of clinically significant prostate lesions (csPCa). METHODS: Multiple deep learning CNN architectures were developed using the Unet as the baseline. The CNNs use both multiparametric MRI images (T2W, ADC, and High b-value) and quantitative clinical data (prostate specific antigen (PSA), PSA density (PSAD), prostate gland volume & gross tumor volume (GTV)), and only mp-MRI images (n = 118), as input. In addition, co-registered ground truth data from whole mount histopathology images (n = 22) were used as a test set for evaluation. RESULTS: The CNNs achieved for early/intermediate / late level fusion a precision of 0.41/0.51/0.61, recall value of 0.18/0.22/0.25, an average precision of 0.13 / 0.19 / 0.27, and F scores of 0.55/0.67/ 0.76. Dice Sorensen Coefficient (DSC) was used to evaluate the influence of combining mpMRI with parametric clinical data for the detection of csPCa. We compared the DSC between the predictions of CNN's trained with mpMRI and parametric clinical and the CNN's trained with only mpMRI images as input with the ground truth. We obtained a DSC of data 0.30/0.34/0.36 and 0.26/0.33/0.34 respectively. Additionally, we evaluated the influence of each mpMRI input channel for the task of csPCa detection and obtained a DSC of 0.14 / 0.25 / 0.28. CONCLUSION: The results show that the decision level fusion network performs better for the task of prostate lesion detection. Combining mpMRI data with quantitative clinical data does not show significant differences between these networks (p = 0.26/0.62/0.85). The results show that CNNs trained with all mpMRI data outperform CNNs with less input channels which is consistent with current clinical protocols where the same input is used for PI-RADS lesion scoring. TRIAL REGISTRATION: The trial was registered retrospectively at the German Register for Clinical Studies (DRKS) under proposal number Nr. 476/14 & 476/19.


Subject(s)
Deep Learning , Multiparametric Magnetic Resonance Imaging , Prostatic Neoplasms , Humans , Male , Prostatic Neoplasms/pathology , Prostatic Neoplasms/diagnostic imaging , Multiparametric Magnetic Resonance Imaging/methods , Aged , Image Interpretation, Computer-Assisted/methods , Middle Aged
2.
Magn Reson Med ; 90(6): 2388-2399, 2023 Dec.
Article in English | MEDLINE | ID: mdl-37427459

ABSTRACT

PURPOSE: MR guidance is used during therapy to detect and compensate for lesion motion. T2 -weighted MRI often has a superior lesion contrast in comparison to T1 -weighted real-time imaging. The purpose of this work was to design a fast T2 -weighted sequence capable of simultaneously acquiring two orthogonal slices, enabling real-time tracking of lesions. METHODS: To generate a T2 contrast in two orthogonal slices simultaneously, a sequence (Ortho-SFFP-Echo) was designed that samples the T2 -weighted spin echo (S- ) signal in a TR-interleaved acquisition of two slices. Slice selection and phase-encoding directions are swapped between the slices, leading to a unique set of spin-echo signal conditions. To minimize motion-related signal dephasing, additional flow-compensation strategies are implemented. In both the abdominal breathing phantom and in vivo experiments, a time series was acquired using Ortho-SSFP-Echo. The centroid of the target was tracked in postprocessing steps. RESULTS: In the phantom, the lesion could be identified and delineated in the dynamic images. In the volunteer experiments, the kidney was visualized with a T2 contrast at a temporal resolution of 0.45 s under free-breathing conditions. A respiratory belt demonstrated a strong correlation with the time course of the kidney centroid in the head-foot direction. A hypointense saturation band at the slice overlap did not inhibit lesion tracking in the semi-automatic postprocessing steps. CONCLUSION: The Ortho-SFFP-Echo sequence delivers real-time images with a T2 -weighted contrast in two orthogonal slices. The sequence allows for simultaneous acquisition, which could be beneficial for real-time motion tracking in radiotherapy or interventional MRI.

3.
Radiat Oncol ; 17(1): 65, 2022 Apr 02.
Article in English | MEDLINE | ID: mdl-35366918

ABSTRACT

Automatic prostate tumor segmentation is often unable to identify the lesion even if multi-parametric MRI data is used as input, and the segmentation output is difficult to verify due to the lack of clinically established ground truth images. In this work we use an explainable deep learning model to interpret the predictions of a convolutional neural network (CNN) for prostate tumor segmentation. The CNN uses a U-Net architecture which was trained on multi-parametric MRI data from 122 patients to automatically segment the prostate gland and prostate tumor lesions. In addition, co-registered ground truth data from whole mount histopathology images were available in 15 patients that were used as a test set during CNN testing. To be able to interpret the segmentation results of the CNN, heat maps were generated using the Gradient Weighted Class Activation Map (Grad-CAM) method. The CNN achieved a mean Dice Sorensen Coefficient 0.62 and 0.31 for the prostate gland and the tumor lesions -with the radiologist drawn ground truth and 0.32 with whole-mount histology ground truth for tumor lesions. Dice Sorensen Coefficient between CNN predictions and manual segmentations from MRI and histology data were not significantly different. In the prostate the Grad-CAM heat maps could differentiate between tumor and healthy prostate tissue, which indicates that the image information in the tumor was essential for the CNN segmentation.


Subject(s)
Multiparametric Magnetic Resonance Imaging , Prostatic Neoplasms , Humans , Magnetic Resonance Imaging/methods , Male , Neural Networks, Computer , Prostatic Neoplasms/diagnostic imaging
4.
In Vivo ; 34(6): 3473-3481, 2020.
Article in English | MEDLINE | ID: mdl-33144456

ABSTRACT

BACKGROUND/AIM: We examined the prognostic value of intraprostatic gross tumour volume (GTV) as measured by multiparametric MRI (mpMRI) in patients with prostate cancer following (primary) external beam radiation therapy (EBRT). PATIENTS AND METHODS: In a retrospective monocentric study, we analysed patients with prostate cancer (PCa) after EBRT. GTV was delineated in pre-treatment mpMRI (GTV-MRI) using T2-weighted images. Cox-regression analyses were performed considering biochemical failure recurrence-free survival (BRFS) as outcome variable. RESULTS: Among 131 patients, after a median follow-up of 57 months, biochemical failure occurred in 27 (21%). GTV-MRI was not correlated with % of positive biopsy cores, Gleason score and initial PSA (all r<0.2) and only moderately correlated with cT stage (r=0.32). In univariate analysis, cT stage, Gleason score and GTV-MRI were higher in subjects with shorter BRFS (p<0.05). GTV-MRI remained a significant predictor for BRFS in multivariate analyses, independent of Gleason score and cT stage. CONCLUSION: GTV, defined using mpMRI, provides incremental prognostic value for BRFS, independent of established risk factors. This supports the implementation of imaging-based GTV for risk-stratification, although further validation is needed.


Subject(s)
Multiparametric Magnetic Resonance Imaging , Prostatic Neoplasms , Humans , Male , Neoplasm Grading , Neoplasm Staging , Prostate-Specific Antigen , Prostatic Neoplasms/diagnostic imaging , Prostatic Neoplasms/pathology , Prostatic Neoplasms/radiotherapy , Retrospective Studies , Tumor Burden
5.
Radiat Oncol ; 15(1): 181, 2020 Jul 29.
Article in English | MEDLINE | ID: mdl-32727525

ABSTRACT

BACKGROUND: Automatic tumor segmentation based on Convolutional Neural Networks (CNNs) has shown to be a valuable tool in treatment planning and clinical decision making. We investigate the influence of 7 MRI input channels of a CNN with respect to the segmentation performance of head&neck cancer. METHODS: Head&neck cancer patients underwent multi-parametric MRI including T2w, pre- and post-contrast T1w, T2*, perfusion (ktrans, ve) and diffusion (ADC) measurements at 3 time points before and during radiochemotherapy. The 7 different MRI contrasts (input channels) and manually defined gross tumor volumes (primary tumor and lymph node metastases) were used to train CNNs for lesion segmentation. A reference CNN with all input channels was compared to individually trained CNNs where one of the input channels was left out to identify which MRI contrast contributes the most to the tumor segmentation task. A statistical analysis was employed to account for random fluctuations in the segmentation performance. RESULTS: The CNN segmentation performance scored up to a Dice similarity coefficient (DSC) of 0.65. The network trained without T2* data generally yielded the worst results, with ΔDSCGTV-T = 5.7% for primary tumor and ΔDSCGTV-Ln = 5.8% for lymph node metastases compared to the network containing all input channels. Overall, the ADC input channel showed the least impact on segmentation performance, with ΔDSCGTV-T = 2.4% for primary tumor and ΔDSCGTV-Ln = 2.2% respectively. CONCLUSIONS: We developed a method to reduce overall scan times in MRI protocols by prioritizing those sequences that add most unique information for the task of automatic tumor segmentation. The optimized CNNs could be used to aid in the definition of the GTVs in radiotherapy planning, and the faster imaging protocols will reduce patient scan times which can increase patient compliance. TRIAL REGISTRATION: The trial was registered retrospectively at the German Register for Clinical Studies (DRKS) under register number DRKS00003830 on August 20th, 2015.


Subject(s)
Head and Neck Neoplasms/diagnostic imaging , Multiparametric Magnetic Resonance Imaging/methods , Neural Networks, Computer , Head and Neck Neoplasms/pathology , Humans , Radiotherapy Planning, Computer-Assisted
SELECTION OF CITATIONS
SEARCH DETAIL