Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 115
Filtrar
1.
J Appl Clin Med Phys ; 25(2): e14155, 2024 Feb.
Artigo em Inglês | MEDLINE | ID: mdl-37712893

RESUMO

Recent advances in MRI-guided radiation therapy (MRgRT) and deep learning techniques encourage fully adaptive radiation therapy (ART), real-time MRI monitoring, and the MRI-only treatment planning workflow. Given the rapid growth and emergence of new state-of-the-art methods in these fields, we systematically review 197 studies written on or before December 31, 2022, and categorize the studies into the areas of image segmentation, image synthesis, radiomics, and real time MRI. Building from the underlying deep learning methods, we discuss their clinical importance and current challenges in facilitating small tumor segmentation, accurate x-ray attenuation information from MRI, tumor characterization and prognosis, and tumor motion tracking. In particular, we highlight the recent trends in deep learning such as the emergence of multi-modal, visual transformer, and diffusion models.


Assuntos
Aprendizado Profundo , Neoplasias , Humanos , Imageamento por Ressonância Magnética/métodos , Neoplasias/diagnóstico por imagem , Neoplasias/radioterapia
2.
J Appl Clin Med Phys ; 24(10): e14064, 2023 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-37345557

RESUMO

In this work, we demonstrate a method for rapid synthesis of high-quality CT images from unpaired, low-quality CBCT images, permitting CBCT-based adaptive radiotherapy. We adapt contrastive unpaired translation (CUT) to be used with medical images and evaluate the results on an institutional pelvic CT dataset. We compare the method against cycleGAN using mean absolute error, structural similarity index, root mean squared error, and Frèchet Inception Distance and show that CUT significantly outperforms cycleGAN while requiring less time and fewer resources. The investigated method improves the feasibility of online adaptive radiotherapy over the present state-of-the-art.


Assuntos
Tomografia Computadorizada de Feixe Cônico Espiral , Humanos , Tomografia Computadorizada de Feixe Cônico/métodos , Processamento de Imagem Assistida por Computador/métodos , Dosagem Radioterapêutica , Planejamento da Radioterapia Assistida por Computador/métodos
3.
Eur Radiol ; 31(6): 3826-3836, 2021 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-33206226

RESUMO

OBJECTIVES: To develop a deep learning-based method for simultaneous myocardium and pericardial fat quantification from coronary computed tomography angiography (CCTA) for the diagnosis and treatment of cardiovascular disease (CVD). METHODS: We retrospectively identified CCTA data obtained between May 2008 and July 2018 in a multicenter (six centers) CVD study. The proposed method was evaluated on 422 patients' data by two studies. The first overall study involves training model on CVD patients and testing on non-CVD patients, as well as training on non-CVD patients and testing on CVD patients. The second study was performed using the leave-center-out approach. The method performance was evaluated using Dice similarity coefficient (DSC), Jaccard index (JAC), 95% Hausdorff distance (HD95), mean surface distance (MSD), residual mean square distance (RMSD), and the center of mass distance (CMD). The robustness of the proposed method was tested using the nonparametric Kruskal-Wallis test and post hoc test to assess the equality of distribution of DSC values among different tests. RESULTS: The automatic segmentation achieved a strong correlation with contour (ICC and R > 0.97, p value < 0.001 throughout all tests). The accuracy of the proposed method remained high through all the tests, with the median DSC higher than 0.88 for pericardial fat and 0.96 for myocardium. The proposed method also resulted in mean MSD, RMSD, HD95, and CMD of less than 1.36 mm for pericardial fat and 1.00 mm for myocardium. CONCLUSIONS: The proposed deep learning-based segmentation method enables accurate simultaneous quantification of myocardium and pericardial fat in a multicenter study. KEY POINTS: • Deep learning-based myocardium and pericardial fat segmentation method tested on 422 patients' coronary computed tomography angiography in a multicenter study. • The proposed method provides segmentations with high volumetric accuracy (ICC and R > 0.97, p value < 0.001) and similar shape as manual annotation by experienced radiologists (median Dice similarity coefficient ≥ 0.88 for pericardial fat and 0.96 for myocardium).


Assuntos
Angiografia por Tomografia Computadorizada , Tomografia Computadorizada por Raios X , Humanos , Processamento de Imagem Assistida por Computador , Miocárdio , Pericárdio/diagnóstico por imagem , Estudos Retrospectivos
4.
J Appl Clin Med Phys ; 22(1): 11-36, 2021 Jan.
Artigo em Inglês | MEDLINE | ID: mdl-33305538

RESUMO

This paper reviewed the deep learning-based studies for medical imaging synthesis and its clinical application. Specifically, we summarized the recent developments of deep learning-based methods in inter- and intra-modality image synthesis by listing and highlighting the proposed methods, study designs, and reported performances with related clinical applications on representative studies. The challenges among the reviewed studies were then summarized with discussion.


Assuntos
Aprendizado Profundo , Diagnóstico por Imagem , Humanos , Processamento de Imagem Assistida por Computador , Radiografia , Projetos de Pesquisa
5.
J Appl Clin Med Phys ; 22(7): 10-26, 2021 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-34164913

RESUMO

Medical imaging is widely used in the diagnosis and treatment of cancer, and artificial intelligence (AI) has achieved tremendous success in medical image analysis. This paper reviews AI-based tumor subregion analysis in medical imaging. We summarize the latest AI-based methods for tumor subregion analysis and their applications. Specifically, we categorize the AI-based methods by training strategy: supervised and unsupervised. A detailed review of each category is presented, highlighting important contributions and achievements. Specific challenges and potential applications of AI in tumor subregion analysis are discussed.


Assuntos
Inteligência Artificial , Neoplasias , Diagnóstico por Imagem , Humanos , Neoplasias/diagnóstico por imagem
6.
J Nucl Cardiol ; 27(3): 976-987, 2020 06.
Artigo em Inglês | MEDLINE | ID: mdl-30693428

RESUMO

BACKGROUND: The performance of left ventricular (LV) functional assessment using gated myocardial perfusion SPECT (MPS) relies on the accuracy of segmentation. Current methods require manual adjustments that are tedious and subjective. We propose a novel machine-learning-based method to automatically segment LV myocardium and measure its volume in gated MPS imaging without human intervention. METHODS: We used an end-to-end fully convolutional neural network to segment LV myocardium by delineating its endocardial and epicardial surface. A novel compound loss function, which encourages similarity and penalizes discrepancy between prediction and training dataset, is utilized in training stage to achieve excellent performance. We retrospectively investigated 32 normal patients and 24 abnormal patients, whose LV myocardial contours automatically segmented by our method were compared with those delineated by physicians as the ground truth. RESULTS: The results of our method demonstrated very good agreement with the ground truth. The average DSC metrics and Hausdorff distance of the contours delineated by our method are larger than 0.900 and less than 1 cm, respectively, among all 32 + 24 patients of all phases. The correlation coefficient of the LV myocardium volume between ground truth and our results is 0.910 ± 0.061 (P < 0.001), and the mean relative error of LV myocardium volume is - 1.09 ± 3.66%. CONCLUSION: These results strongly indicate the feasibility of our method in accurately quantifying LV myocardium volume change over the cardiac cycle. The learning-based segmentation method in gated MPS imaging has great promise for clinical use.


Assuntos
Ventrículos do Coração/diagnóstico por imagem , Miocárdio/patologia , Tomografia Computadorizada de Emissão de Fóton Único/métodos , Idoso , Algoritmos , Estudos de Viabilidade , Feminino , Coração/diagnóstico por imagem , Humanos , Processamento de Imagem Assistida por Computador/métodos , Aprendizado de Máquina , Masculino , Pessoa de Meia-Idade , Imagem de Perfusão do Miocárdio , Redes Neurais de Computação , Reconhecimento Automatizado de Padrão , Perfusão , Probabilidade , Reprodutibilidade dos Testes , Estudos Retrospectivos
7.
J Appl Clin Med Phys ; 20(2): 121-128, 2019 Feb.
Artigo em Inglês | MEDLINE | ID: mdl-30693665

RESUMO

PURPOSE: Dual-energy computed tomography (DECT) using TwinBeam CT (TBCT) is a new option for radiation oncology simulators. TBCT scanning provides virtual monoenergetic images which are attractive in treatment planning since lower energies offer better contrast for soft tissues, and higher energies reduce noise. A protocol is needed to achieve optimal performance of this feature. In this study, we investigated the TBCT scan schema with the head-and-neck radiotherapy workflow at our clinic and selected the optimal energy with best contrast-noise-ratio (CNR) in organs-at-risks (OARs) delineation for head-and-neck treatment planning. METHODS AND MATERIALS: We synthesized monochromatic images from 40 keV to 190 keV at 5 keV increments from data acquired by TBCT. We collected the Hounsfield unit (HU) numbers of OARs (brainstem, mandible, spinal cord, and parotid glands), the HU numbers of marginal regions outside OARs, and the noise levels for each monochromatic image. We then calculated the CNR for the different OARs at each energy level to generate a serial of spectral curves for each OAR. Based on these spectral curves of CNR, the mono-energy corresponding to the max CNR was identified for each OAR of each patient. RESULTS: Computed tomography scans of ten patients by TBCT were used to test the optimal monoenergetic image for the CNR of OAR. Based on the maximized CNR, the optimal energy values were 78.5 ± 5.3 keV for the brainstem, 78.0 ± 4.2 keV for the mandible, 78.5 ± 5.7 keV for the parotid glands, and 78.5 ± 5.3 keV for the spinal cord. Overall, the optimal energy for the maximum CNR of these OARs in head-and-neck cancer patients was 80 keV. CONCLUSION: We have proposed a clinically feasible protocol that selects the optimal energy level of the virtual monoenergetic image in TBCT for OAR delineation based on the CNR in head-and-neck OAR. This protocol can be applied in TBCT simulation.


Assuntos
Carcinoma de Células Escamosas/radioterapia , Neoplasias de Cabeça e Pescoço/radioterapia , Órgãos em Risco/efeitos da radiação , Interpretação de Imagem Radiográfica Assistida por Computador/normas , Imagem Radiográfica a Partir de Emissão de Duplo Fóton/normas , Planejamento da Radioterapia Assistida por Computador/métodos , Tomografia Computadorizada por Raios X/métodos , Adulto , Idoso , Idoso de 80 Anos ou mais , Carcinoma de Células Escamosas/diagnóstico por imagem , Feminino , Neoplasias de Cabeça e Pescoço/diagnóstico por imagem , Humanos , Masculino , Pessoa de Meia-Idade , Órgãos em Risco/diagnóstico por imagem , Prognóstico , Interpretação de Imagem Radiográfica Assistida por Computador/métodos , Imagem Radiográfica a Partir de Emissão de Duplo Fóton/métodos , Dosagem Radioterapêutica , Radioterapia de Intensidade Modulada/métodos , Estudos Retrospectivos , Razão Sinal-Ruído
8.
Cell Rep Med ; 5(4): 101486, 2024 Apr 16.
Artigo em Inglês | MEDLINE | ID: mdl-38631288

RESUMO

PET scans provide additional clinical value but are costly and not universally accessible. Salehjahromi et al.1 developed an AI-based pipeline to synthesize PET images from diagnostic CT scans, demonstrating its potential clinical utility across various clinical tasks for lung cancer.


Assuntos
Neoplasias Pulmonares , Humanos , Fluordesoxiglucose F18 , Tomografia Computadorizada por Raios X/métodos , Prognóstico , Inteligência Artificial
9.
Med Phys ; 51(3): 1974-1984, 2024 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-37708440

RESUMO

BACKGROUND: An automated, accurate, and efficient lung four-dimensional computed tomography (4DCT) image registration method is clinically important to quantify respiratory motion for optimal motion management. PURPOSE: The purpose of this work is to develop a weakly supervised deep learning method for 4DCT lung deformable image registration (DIR). METHODS: The landmark-driven cycle network is proposed as a deep learning platform that performs DIR of individual phase datasets in a simulation 4DCT. This proposed network comprises a generator and a discriminator. The generator accepts moving and target CTs as input and outputs the deformation vector fields (DVFs) to match the two CTs. It is optimized during both forward and backward paths to enhance the bi-directionality of DVF generation. Further, the landmarks are used to weakly supervise the generator network. Landmark-driven loss is used to guide the generator's training. The discriminator then judges the realism of the deformed CT to provide extra DVF regularization. RESULTS: We performed four-fold cross-validation on 10 4DCT datasets from the public DIR-Lab dataset and a hold-out test on our clinic dataset, which included 50 4DCT datasets. The DIR-Lab dataset was used to evaluate the performance of the proposed method against other methods in the literature by calculating the DIR-Lab Target Registration Error (TRE). The proposed method outperformed other deep learning-based methods on the DIR-Lab datasets in terms of TRE. Bi-directional and landmark-driven loss were shown to be effective for obtaining high registration accuracy. The mean and standard deviation of TRE for the DIR-Lab datasets was 1.20 ± 0.72 mm and the mean absolute error (MAE) and structural similarity index (SSIM) for our datasets were 32.1 ± 11.6 HU and 0.979 ± 0.011, respectively. CONCLUSION: The landmark-driven cycle network has been validated and tested for automatic deformable image registration of patients' lung 4DCTs with results comparable to or better than competing methods.


Assuntos
Tomografia Computadorizada Quadridimensional , Processamento de Imagem Assistida por Computador , Humanos , Processamento de Imagem Assistida por Computador/métodos , Pulmão/diagnóstico por imagem , Simulação por Computador , Movimento (Física) , Algoritmos
10.
ArXiv ; 2024 Jun 21.
Artigo em Inglês | MEDLINE | ID: mdl-38947928

RESUMO

BACKGROUND: Cone-beam computed tomography (CBCT) scans, performed fractionally (e.g., daily or weekly), are widely utilized for patient alignment in the image-guided radiotherapy (IGRT) process, thereby making it a potential imaging modality for the implementation of adaptive radiotherapy (ART) protocols. Nonetheless, significant artifacts and incorrect Hounsfield unit (HU) values hinder their application in quantitative tasks such as target and organ segmentations and dose calculation. Therefore, acquiring CT-quality images from the CBCT scans is essential to implement online ART in clinical settings. PURPOSE: This work aims to develop an unsupervised learning method using the patient-specific diffusion model for CBCT-based synthetic CT (sCT) generation to improve the image quality of CBCT. METHODS: The proposed method is in an unsupervised framework that utilizes a patient-specific score-based model as the image prior alongside a customized total variation (TV) regularization to enforce coherence across different transverse slices. The score-based model is unconditionally trained using the same patient's planning CT (pCT) images to characterize the manifold of CT-quality images and capture the unique anatomical information of the specific patient. The efficacy of the proposed method was assessed on images from anatomical sites including head and neck (H&N) cancer, pancreatic cancer, and lung cancer. The performance of the proposed CBCT correction method was evaluated using quantitative metrics including mean absolute error (MAE), peak signal-to-noise ratio (PSNR), and normalized cross-correlation (NCC). Additionally, the proposed algorithm was benchmarked against two other unsupervised diffusion model-based CBCT correction algorithms.

11.
Med Phys ; 2024 Jun 12.
Artigo em Inglês | MEDLINE | ID: mdl-38865687

RESUMO

BACKGROUND: Dual-energy computed tomography (DECT) and material decomposition play vital roles in quantitative medical imaging. However, the decomposition process may suffer from significant noise amplification, leading to severely degraded image signal-to-noise ratios (SNRs). While existing iterative algorithms perform noise suppression using different image priors, these heuristic image priors cannot accurately represent the features of the target image manifold. Although deep learning-based decomposition methods have been reported, these methods are in the supervised-learning framework requiring paired data for training, which is not readily available in clinical settings. PURPOSE: This work aims to develop an unsupervised-learning framework with data-measurement consistency for image-domain material decomposition in DECT. METHODS: The proposed framework combines iterative decomposition and deep learning-based image prior in a generative adversarial network (GAN) architecture. In the generator module, a data-fidelity loss is introduced to enforce the measurement consistency in material decomposition. In the discriminator module, the discriminator is trained to differentiate the low-noise material-specific images from the high-noise images. In this scheme, paired images of DECT and ground-truth material-specific images are not required for the model training. Once trained, the generator can perform image-domain material decomposition with noise suppression in a single step. RESULTS: In the simulation studies of head and lung digital phantoms, the proposed method reduced the standard deviation (SD) in decomposed images by 97% and 91% from the values in direct inversion results. It also generated decomposed images with structural similarity index measures (SSIMs) greater than 0.95 against the ground truth. In the clinical head and lung patient studies, the proposed method suppressed the SD by 95% and 93% compared to the decomposed images of matrix inversion. CONCLUSIONS: Since the invention of DECT, noise amplification during material decomposition has been one of the biggest challenges, impeding its quantitative use in clinical practice. The proposed method performs accurate material decomposition with efficient noise suppression. Furthermore, the proposed method is within an unsupervised-learning framework, which does not require paired data for model training and resolves the issue of lack of ground-truth data in clinical scenarios.

12.
Med Phys ; 2024 Jun 18.
Artigo em Inglês | MEDLINE | ID: mdl-38889368

RESUMO

BACKGROUND: Iodine maps, derived from image-processing of contrast-enhanced dual-energy computed tomography (DECT) scans, highlight the differences in tissue iodine intake. It finds multiple applications in radiology, including vascular imaging, pulmonary evaluation, kidney assessment, and cancer diagnosis. In radiation oncology, it can contribute to designing more accurate and personalized treatment plans. However, DECT scanners are not commonly available in radiation therapy centers. Additionally, the use of iodine contrast agents is not suitable for all patients, especially those allergic to iodine agents, posing further limitations to the accessibility of this technology. PURPOSE: The purpose of this work is to generate synthetic iodine map images from non-contrast single-energy CT (SECT) images using conditional denoising diffusion probabilistic model (DDPM). METHODS: One-hundered twenty-six head-and-neck patients' images were retrospectively investigated in this work. Each patient underwent non-contrast SECT and contrast DECT scans. Ground truth iodine maps were generated from contrast DECT scans using commercial software syngo.via installed in the clinic. A conditional DDPM was implemented in this work to synthesize iodine maps. Three-fold cross-validation was conducted, with each iteration selecting the data from 42 patients as the test dataset and the remainder as the training dataset. Pixel-to-pixel generative adversarial network (GAN) and CycleGAN served as reference methods for evaluating the proposed DDPM method. RESULTS: The accuracy of the proposed DDPM was evaluated using three quantitative metrics: mean absolute error (MAE) (1.039 ± 0.345 mg/mL), structural similarity index measure (SSIM) (0.89 ± 0.10) and peak signal-to-noise ratio (PSNR) (25.4 ± 3.5 db) respectively. Compared to the reference methods, the proposed technique showcased superior performance across the evaluated metrics, further validated by the paired two-tailed t-tests. CONCLUSION: The proposed conditional DDPM framework has demonstrated the feasibility of generating synthetic iodine map images from non-contrast SECT images. This method presents a potential clinical application, which is providing accurate iodine contrast map in instances where only non-contrast SECT is accessible.

13.
Med Phys ; 2024 Apr 08.
Artigo em Inglês | MEDLINE | ID: mdl-38588512

RESUMO

PURPOSE: Positron Emission Tomography (PET) has been a commonly used imaging modality in broad clinical applications. One of the most important tradeoffs in PET imaging is between image quality and radiation dose: high image quality comes with high radiation exposure. Improving image quality is desirable for all clinical applications while minimizing radiation exposure is needed to reduce risk to patients. METHODS: We introduce PET Consistency Model (PET-CM), an efficient diffusion-based method for generating high-quality full-dose PET images from low-dose PET images. It employs a two-step process, adding Gaussian noise to full-dose PET images in the forward diffusion, and then denoising them using a PET Shifted-window Vision Transformer (PET-VIT) network in the reverse diffusion. The PET-VIT network learns a consistency function that enables direct denoising of Gaussian noise into clean full-dose PET images. PET-CM achieves state-of-the-art image quality while requiring significantly less computation time than other methods. Evaluation with normalized mean absolute error (NMAE), peak signal-to-noise ratio (PSNR), multi-scale structure similarity index (SSIM), normalized cross-correlation (NCC), and clinical evaluation including Human Ranking Score (HRS) and Standardized Uptake Value (SUV) Error analysis shows its superiority in synthesizing full-dose PET images from low-dose inputs. RESULTS: In experiments comparing eighth-dose to full-dose images, PET-CM demonstrated impressive performance with NMAE of 1.278 ± 0.122%, PSNR of 33.783 ± 0.824 dB, SSIM of 0.964 ± 0.009, NCC of 0.968 ± 0.011, HRS of 4.543, and SUV Error of 0.255 ± 0.318%, with an average generation time of 62 s per patient. This is a significant improvement compared to the state-of-the-art diffusion-based model with PET-CM reaching this result 12× faster. Similarly, in the quarter-dose to full-dose image experiments, PET-CM delivered competitive outcomes, achieving an NMAE of 0.973 ± 0.066%, PSNR of 36.172 ± 0.801 dB, SSIM of 0.984 ± 0.004, NCC of 0.990 ± 0.005, HRS of 4.428, and SUV Error of 0.151 ± 0.192% using the same generation process, which underlining its high quantitative and clinical precision in both denoising scenario. CONCLUSIONS: We propose PET-CM, the first efficient diffusion-model-based method, for estimating full-dose PET images from low-dose images. PET-CM provides comparable quality to the state-of-the-art diffusion model with higher efficiency. By utilizing this approach, it becomes possible to maintain high-quality PET images suitable for clinical use while mitigating the risks associated with radiation. The code is availble at https://github.com/shaoyanpan/Full-dose-Whole-body-PET-Synthesis-from-Low-dose-PET-Using-Consistency-Model.

14.
ArXiv ; 2024 May 04.
Artigo em Inglês | MEDLINE | ID: mdl-38745706

RESUMO

Background: Stereotactic body radiotherapy (SBRT) is a well-established treatment modality for liver metastases in patients unsuitable for surgery. Both CT and MRI are useful during treatment planning for accurate target delineation and to reduce potential organs-at-risk (OAR) toxicity from radiation. MRI-CT deformable image registration (DIR) is required to propagate the contours defined on high-contrast MRI to CT images. An accurate DIR method could lead to more precisely defined treatment volumes and superior OAR sparing on the treatment plan. Therefore, it is beneficial to develop an accurate MRI-CT DIR for liver SBRT. Purpose: To create a new deep learning model that can estimate the deformation vector field (DVF) for directly registering abdominal MRI-CT images. Methods: The proposed method assumed a diffeomorphic deformation. By using topology-preserved deformation features extracted from the probabilistic diffeomorphic registration model, abdominal motion can be accurately obtained and utilized for DVF estimation. The model integrated Swin transformers, which have demonstrated superior performance in motion tracking, into the convolutional neural network (CNN) for deformation feature extraction. The model was optimized using a cross-modality image similarity loss and a surface matching loss. To compute the image loss, a modality-independent neighborhood descriptor (MIND) was used between the deformed MRI and CT images. The surface matching loss was determined by measuring the distance between the warped coordinates of the surfaces of contoured structures on the MRI and CT images. To evaluate the performance of the model, a retrospective study was carried out on a group of 50 liver cases that underwent rigid registration of MRI and CT scans. The deformed MRI image was assessed against the CT image using the target registration error (TRE), Dice similarity coefficient (DSC), and mean surface distance (MSD) between the deformed contours of the MRI image and manual contours of the CT image. Results: When compared to only rigid registration, DIR with the proposed method resulted in an increase of the mean DSC values of the liver and portal vein from 0.850±0.102 and 0.628±0.129 to 0.903±0.044 and 0.763±0.073, a decrease of the mean MSD of the liver from 7.216±4.513 mm to 3.232±1.483 mm, and a decrease of the TRE from 26.238±2.769 mm to 8.492±1.058 mm. Conclusion: The proposed DIR method based on a diffeomorphic transformer provides an effective and efficient way to generate an accurate DVF from an MRI-CT image pair of the abdomen. It could be utilized in the current treatment planning workflow for liver SBRT.

15.
Med Phys ; 2024 May 31.
Artigo em Inglês | MEDLINE | ID: mdl-38820286

RESUMO

BACKGROUND: Stereotactic body radiotherapy (SBRT) is a well-established treatment modality for liver metastases in patients unsuitable for surgery. Both CT and MRI are useful during treatment planning for accurate target delineation and to reduce potential organs-at-risk (OAR) toxicity from radiation. MRI-CT deformable image registration (DIR) is required to propagate the contours defined on high-contrast MRI to CT images. An accurate DIR method could lead to more precisely defined treatment volumes and superior OAR sparing on the treatment plan. Therefore, it is beneficial to develop an accurate MRI-CT DIR for liver SBRT. PURPOSE: To create a new deep learning model that can estimate the deformation vector field (DVF) for directly registering abdominal MRI-CT images. METHODS: The proposed method assumed a diffeomorphic deformation. By using topology-preserved deformation features extracted from the probabilistic diffeomorphic registration model, abdominal motion can be accurately obtained and utilized for DVF estimation. The model integrated Swin transformers, which have demonstrated superior performance in motion tracking, into the convolutional neural network (CNN) for deformation feature extraction. The model was optimized using a cross-modality image similarity loss and a surface matching loss. To compute the image loss, a modality-independent neighborhood descriptor (MIND) was used between the deformed MRI and CT images. The surface matching loss was determined by measuring the distance between the warped coordinates of the surfaces of contoured structures on the MRI and CT images. To evaluate the performance of the model, a retrospective study was carried out on a group of 50 liver cases that underwent rigid registration of MRI and CT scans. The deformed MRI image was assessed against the CT image using the target registration error (TRE), Dice similarity coefficient (DSC), and mean surface distance (MSD) between the deformed contours of the MRI image and manual contours of the CT image. RESULTS: When compared to only rigid registration, DIR with the proposed method resulted in an increase of the mean DSC values of the liver and portal vein from 0.850 ± 0.102 and 0.628 ± 0.129 to 0.903 ± 0.044 and 0.763 ± 0.073, a decrease of the mean MSD of the liver from 7.216 ± 4.513 mm to 3.232 ± 1.483 mm, and a decrease of the TRE from 26.238 ± 2.769 mm to 8.492 ± 1.058 mm. CONCLUSION: The proposed DIR method based on a diffeomorphic transformer provides an effective and efficient way to generate an accurate DVF from an MRI-CT image pair of the abdomen. It could be utilized in the current treatment planning workflow for liver SBRT.

16.
Med Phys ; 51(6): 4380-4388, 2024 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-38630982

RESUMO

BACKGROUND: 7 Tesla (7T) apparent diffusion coefficient (ADC) maps derived from diffusion-weighted imaging (DWI) demonstrate improved image quality and spatial resolution over 3 Tesla (3T) ADC maps. However, 7T magnetic resonance imaging (MRI) currently suffers from limited clinical unavailability, higher cost, and increased susceptibility to artifacts. PURPOSE: To address these issues, we propose a hybrid CNN-transformer model to synthesize high-resolution 7T ADC maps from multimodal 3T MRI. METHODS: The Vision CNN-Transformer (VCT), composed of both Vision Transformer (ViT) blocks and convolutional layers, is proposed to produce high-resolution synthetic 7T ADC maps from 3T ADC maps and 3T T1-weighted (T1w) MRI. ViT blocks enabled global image context while convolutional layers efficiently captured fine detail. The VCT model was validated on the publicly available Human Connectome Project Young Adult dataset, comprising 3T T1w, 3T DWI, and 7T DWI brain scans. The Diffusion Imaging in Python library was used to compute ADC maps from the DWI scans. A total of 171 patient cases were randomly divided into 130 training cases, 20 validation cases, and 21 test cases. The synthetic ADC maps were evaluated by comparing their similarity to the ground truth volumes with the following metrics: peak signal-to-noise ratio (PSNR), structural similarity index measure (SSIM), and mean squared error (MSE). In addition, RESULTS: The results are as follows: PSNR: 27.0 ± 0.9 dB, SSIM: 0.945 ± 0.010, and MSE: 2.0E-3 ± 0.4E-3. Both qualitative and quantitative results demonstrate that VCT performs favorably against other state-of-the-art methods. We have introduced various efficiency improvements, including the implementation of flash attention and training on 176×208 resolution images. These enhancements have resulted in the reduction of parameters and training time per epoch by 50% in comparison to ResViT. Specifically, the training time per epoch has been shortened from 7.67 min to 3.86 min. CONCLUSION: We propose a novel method to predict high-resolution 7T ADC maps from low-resolution 3T ADC maps and T1w MRI. Our predicted images demonstrate better spatial resolution and contrast compared to 3T MRI and prediction results made by ResViT and pix2pix. These high-quality synthetic 7T MR images could be beneficial for disease diagnosis and intervention, producing higher resolution and conformal contours, and as an intermediate step in generating synthetic CT for radiation therapy, especially when 7T MRI scanners are unavailable.


Assuntos
Processamento de Imagem Assistida por Computador , Redes Neurais de Computação , Processamento de Imagem Assistida por Computador/métodos , Humanos , Imagem de Difusão por Ressonância Magnética , Encéfalo/diagnóstico por imagem , Imageamento por Ressonância Magnética
17.
Med Phys ; 51(3): 1847-1859, 2024 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-37646491

RESUMO

BACKGROUND: Daily or weekly cone-beam computed tomography (CBCT) scans are commonly used for accurate patient positioning during the image-guided radiotherapy (IGRT) process, making it an ideal option for adaptive radiotherapy (ART) replanning. However, the presence of severe artifacts and inaccurate Hounsfield unit (HU) values prevent its use for quantitative applications such as organ segmentation and dose calculation. To enable the clinical practice of online ART, it is crucial to obtain CBCT scans with a quality comparable to that of a CT scan. PURPOSE: This work aims to develop a conditional diffusion model to perform image translation from the CBCT to the CT distribution for the image quality improvement of CBCT. METHODS: The proposed method is a conditional denoising diffusion probabilistic model (DDPM) that utilizes a time-embedded U-net architecture with residual and attention blocks to gradually transform the white Gaussian noise sample to the target CT distribution conditioned on the CBCT. The model was trained on deformed planning CT (dpCT) and CBCT image pairs, and its feasibility was verified in brain patient study and head-and-neck (H&N) patient study. The performance of the proposed algorithm was evaluated using mean absolute error (MAE), peak signal-to-noise ratio (PSNR) and normalized cross-correlation (NCC) metrics on generated synthetic CT (sCT) samples. The proposed method was also compared to four other diffusion model-based sCT generation methods. RESULTS: In the brain patient study, the MAE, PSNR, and NCC of the generated sCT were 25.99 HU, 30.49 dB, and 0.99, respectively, compared to 40.63 HU, 27.87 dB, and 0.98 of the CBCT images. In the H&N patient study, the metrics were 32.56 HU, 27.65 dB, 0.98 and 38.99 HU, 27.00, 0.98 for sCT and CBCT, respectively. Compared to the other four diffusion models and one Cycle generative adversarial network (Cycle GAN), the proposed method showed superior results in both visual quality and quantitative analysis. CONCLUSIONS: The proposed conditional DDPM method can generate sCT from CBCT with accurate HU numbers and reduced artifacts, enabling accurate CBCT-based organ segmentation and dose calculation for online ART.


Assuntos
Bisacodil/análogos & derivados , Processamento de Imagem Assistida por Computador , Tomografia Computadorizada de Feixe Cônico Espiral , Humanos , Processamento de Imagem Assistida por Computador/métodos , Tomografia Computadorizada de Feixe Cônico , Tomografia Computadorizada por Raios X , Modelos Estatísticos , Planejamento da Radioterapia Assistida por Computador/métodos
18.
Med Phys ; 51(4): 2538-2548, 2024 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-38011588

RESUMO

BACKGROUND AND PURPOSE: Magnetic resonance imaging (MRI)-based synthetic computed tomography (sCT) simplifies radiation therapy treatment planning by eliminating the need for CT simulation and error-prone image registration, ultimately reducing patient radiation dose and setup uncertainty. In this work, we propose a MRI-to-CT transformer-based improved denoising diffusion probabilistic model (MC-IDDPM) to translate MRI into high-quality sCT to facilitate radiation treatment planning. METHODS: MC-IDDPM implements diffusion processes with a shifted-window transformer network to generate sCT from MRI. The proposed model consists of two processes: a forward process, which involves adding Gaussian noise to real CT scans to create noisy images, and a reverse process, in which a shifted-window transformer V-net (Swin-Vnet) denoises the noisy CT scans conditioned on the MRI from the same patient to produce noise-free CT scans. With an optimally trained Swin-Vnet, the reverse diffusion process was used to generate noise-free sCT scans matching MRI anatomy. We evaluated the proposed method by generating sCT from MRI on an institutional brain dataset and an institutional prostate dataset. Quantitative evaluations were conducted using several metrics, including Mean Absolute Error (MAE), Peak Signal-to-Noise Ratio (PSNR), Multi-scale Structure Similarity Index (SSIM), and Normalized Cross Correlation (NCC). Dosimetry analyses were also performed, including comparisons of mean dose and target dose coverages for 95% and 99%. RESULTS: MC-IDDPM generated brain sCTs with state-of-the-art quantitative results with MAE 48.825 ± 21.491 HU, PSNR 26.491 ± 2.814 dB, SSIM 0.947 ± 0.032, and NCC 0.976 ± 0.019. For the prostate dataset: MAE 55.124 ± 9.414 HU, PSNR 28.708 ± 2.112 dB, SSIM 0.878 ± 0.040, and NCC 0.940 ± 0.039. MC-IDDPM demonstrates a statistically significant improvement (with p < 0.05) in most metrics when compared to competing networks, for both brain and prostate synthetic CT. Dosimetry analyses indicated that the target dose coverage differences by using CT and sCT were within ± 0.34%. CONCLUSIONS: We have developed and validated a novel approach for generating CT images from routine MRIs using a transformer-based improved DDPM. This model effectively captures the complex relationship between CT and MRI images, allowing for robust and high-quality synthetic CT images to be generated in a matter of minutes. This approach has the potential to greatly simplify the treatment planning process for radiation therapy by eliminating the need for additional CT scans, reducing the amount of time patients spend in treatment planning, and enhancing the accuracy of treatment delivery.


Assuntos
Cabeça , Tomografia Computadorizada por Raios X , Masculino , Humanos , Tomografia Computadorizada por Raios X/métodos , Imageamento por Ressonância Magnética/métodos , Planejamento da Radioterapia Assistida por Computador/métodos , Radiometria , Processamento de Imagem Assistida por Computador/métodos
19.
Med Phys ; 50(5): 3027-3038, 2023 May.
Artigo em Inglês | MEDLINE | ID: mdl-36463516

RESUMO

BACKGROUND: Manual contouring is very labor-intensive, time-consuming, and subject to intra- and inter-observer variability. An automated deep learning approach to fast and accurate contouring and segmentation is desirable during radiotherapy treatment planning. PURPOSE: This work investigates an efficient deep-learning-based segmentation algorithm in abdomen computed tomography (CT) to facilitate radiation treatment planning. METHODS: In this work, we propose a novel deep-learning model utilizing U-shaped multi-layer perceptron mixer (MLP-Mixer) and convolutional neural network (CNN) for multi-organ segmentation in abdomen CT images. The proposed model has a similar structure to V-net, while a proposed MLP-Convolutional block replaces each convolutional block. The MLP-Convolutional block consists of three components: an early convolutional block for local features extraction and feature resampling, a token-based MLP-Mixer layer for capturing global features with high efficiency, and a token projector for pixel-level detail recovery. We evaluate our proposed network using: (1) an institutional dataset with 60 patient cases and (2) a public dataset (BCTV) with 30 patient cases. The network performance was quantitatively evaluated in three domains: (1) volume similarity between the ground truth contours and the network predictions using the Dice score coefficient (DSC), sensitivity, and precision; (2) surface similarity using Hausdorff distance (HD), mean surface distance (MSD) and residual mean square distance (RMS); and (3) the computational complexity reported by the number of network parameters, training time, and inference time. The performance of the proposed network is compared with other state-of-the-art networks. RESULTS: In the institutional dataset, the proposed network achieved the following volume similarity measures when averaged over all organs: DSC = 0.912, sensitivity = 0.917, precision = 0.917, average surface similarities were HD = 11.95 mm, MSD = 1.90 mm, RMS = 3.86 mm. The proposed network achieved DSC = 0.786 and HD = 9.04 mm on the public dataset. The network also shows statistically significant improvement, which is evaluated by a two-tailed Wilcoxon Mann-Whitney U test, on right lung (MSD where the maximum p-value is 0.001), spinal cord (sensitivity, precision, HD, RMSD where p-value ranges from 0.001 to 0.039), and stomach (DSC where the maximum p-value is 0.01) over all other competing networks. On the public dataset, the network report statistically significant improvement, which is shown by the Wilcoxon Mann-Whitney test, on pancreas (HD where the maximum p-value is 0.006), left (HD where the maximum p-value is 0.022) and right adrenal glands (DSC where the maximum p-value is 0.026). In both datasets, the proposed method can generate contours in less than 5 s. Overall, the proposed MLP-Vnet demonstrates comparable or better performance than competing methods with much lower memory complexity and higher speed. CONCLUSIONS: The proposed MLP-Vnet demonstrates superior segmentation performance, in terms of accuracy and efficiency, relative to state-of-the-art methods. This reliable and efficient method demonstrates potential to streamline clinical workflows in abdominal radiotherapy, which may be especially important for online adaptive treatments.


Assuntos
Redes Neurais de Computação , Tomografia Computadorizada por Raios X , Humanos , Abdome/diagnóstico por imagem , Algoritmos , Pulmão , Processamento de Imagem Assistida por Computador/métodos
20.
ArXiv ; 2023 Jun 14.
Artigo em Inglês | MEDLINE | ID: mdl-37396614

RESUMO

Background: The hippocampus plays a crucial role in memory and cognition. Because of the associated toxicity from whole brain radiotherapy, more advanced treatment planning techniques prioritize hippocampal avoidance, which depends on an accurate segmentation of the small and complexly shaped hippocampus. Purpose: To achieve accurate segmentation of the anterior and posterior regions of the hippocampus from T1 weighted (T1w) MRI images, we developed a novel model, Hippo-Net, which uses a mutually enhanced strategy. Methods: The proposed model consists of two major parts: 1) a localization model is used to detect the volume-of-interest (VOI) of hippocampus. 2) An end-to-end morphological vision transformer network is used to perform substructures segmentation within the hippocampus VOI. The substructures include the anterior and posterior regions of the hippocampus, which are defined as the hippocampus proper and parts of the subiculum. The vision transformer incorporates the dominant features extracted from MRI images, which are further improved by learning-based morphological operators. The integration of these morphological operators into the vision transformer increases the accuracy and ability to separate hippocampus structure into its two distinct substructures.A total of 260 T1w MRI datasets from Medical Segmentation Decathlon dataset were used in this study. We conducted a five-fold cross-validation on the first 200 T1w MR images and then performed a hold-out test on the remaining 60 T1w MR images with the model trained on the first 200 images. The segmentations were evaluated with two indicators, 1) multiple metrics including the Dice similarity coefficient (DSC), 95th percentile Hausdorff distance (HD95), mean surface distance (MSD), volume difference (VD) and center-of-mass distance (COMD); 2) Volumetric Pearson correlation analysis. Results: In five-fold cross-validation, the DSCs were 0.900±0.029 and 0.886±0.031 for the hippocampus proper and parts of the subiculum, respectively. The MSD were 0.426±0.115mm and 0.401±0.100 mm for the hippocampus proper and parts of the subiculum, respectively. Conclusions: The proposed method showed great promise in automatically delineating hippocampus substructures on T1w MRI images. It may facilitate the current clinical workflow and reduce the physicians' effort.

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA