Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 122
Filtrar
1.
Comput Med Imaging Graph ; 115: 102389, 2024 Jul.
Artículo en Inglés | MEDLINE | ID: mdl-38692199

RESUMEN

Accurate reconstruction of a high-resolution 3D volume of the heart is critical for comprehensive cardiac assessments. However, cardiac magnetic resonance (CMR) data is usually acquired as a stack of 2D short-axis (SAX) slices, which suffers from the inter-slice misalignment due to cardiac motion and data sparsity from large gaps between SAX slices. Therefore, we aim to propose an end-to-end deep learning (DL) model to address these two challenges simultaneously, employing specific model components for each challenge. The objective is to reconstruct a high-resolution 3D volume of the heart (VHR) from acquired CMR SAX slices (VLR). We define the transformation from VLR to VHR as a sequential process of motion correction and super-resolution. Accordingly, our DL model incorporates two distinct components. The first component conducts motion correction by predicting displacement vectors to re-position each SAX slice accurately. The second component takes the motion-corrected SAX slices from the first component and performs the super-resolution to fill the data gaps. These two components operate in a sequential way, and the entire model is trained end-to-end. Our model significantly reduced inter-slice misalignment from originally 3.33±0.74 mm to 1.36±0.63 mm and generated accurate high resolution 3D volumes with Dice of 0.974±0.010 for left ventricle (LV) and 0.938±0.017 for myocardium in a simulation dataset. When compared to the LAX contours in a real-world dataset, our model achieved Dice of 0.945±0.023 for LV and 0.786±0.060 for myocardium. In both datasets, our model with specific components for motion correction and super-resolution significantly enhance the performance compared to the model without such design considerations. The codes for our model are available at https://github.com/zhennongchen/CMR_MC_SR_End2End.


Asunto(s)
Aprendizaje Profundo , Corazón , Imagenología Tridimensional , Imagen por Resonancia Magnética , Humanos , Imagenología Tridimensional/métodos , Corazón/diagnóstico por imagen , Imagen por Resonancia Magnética/métodos , Movimiento (Física) , Procesamiento de Imagen Asistido por Computador/métodos
2.
Phys Med Biol ; 69(11)2024 May 21.
Artículo en Inglés | MEDLINE | ID: mdl-38688292

RESUMEN

Objective.The mean squared error (MSE), also known asL2loss, has been widely used as a loss function to optimize image denoising models due to its strong performance as a mean estimator of the Gaussian noise model. Recently, various low-dose computed tomography (LDCT) image denoising methods using deep learning combined with the MSE loss have been developed; however, this approach has been observed to suffer from the regression-to-the-mean problem, leading to over-smoothed edges and degradation of texture in the image.Approach.To overcome this issue, we propose a stochastic function in the loss function to improve the texture of the denoised CT images, rather than relying on complicated networks or feature space losses. The proposed loss function includes the MSE loss to learn the mean distribution and the Pearson divergence loss to learn feature textures. Specifically, the Pearson divergence loss is computed in an image space to measure the distance between two intensity measures of denoised low-dose and normal-dose CT images. The evaluation of the proposed model employs a novel approach of multi-metric quantitative analysis utilizing relative texture feature distance.Results.Our experimental results show that the proposed Pearson divergence loss leads to a significant improvement in texture compared to the conventional MSE loss and generative adversarial network (GAN), both qualitatively and quantitatively.Significance.Achieving consistent texture preservation in LDCT is a challenge in conventional GAN-type methods due to adversarial aspects aimed at minimizing noise while preserving texture. By incorporating the Pearson regularizer in the loss function, we can easily achieve a balance between two conflicting properties. Consistent high-quality CT images can significantly help clinicians in diagnoses and supporting researchers in the development of AI-diagnostic models.


Asunto(s)
Procesamiento de Imagen Asistido por Computador , Dosis de Radiación , Relación Señal-Ruido , Tomografía Computarizada por Rayos X , Tomografía Computarizada por Rayos X/métodos , Procesamiento de Imagen Asistido por Computador/métodos , Humanos , Aprendizaje Profundo
3.
Med Phys ; 51(5): 3309-3321, 2024 May.
Artículo en Inglés | MEDLINE | ID: mdl-38569143

RESUMEN

BACKGROUND: Patient head motion is a common source of image artifacts in computed tomography (CT) of the head, leading to degraded image quality and potentially incorrect diagnoses. The partial angle reconstruction (PAR) means dividing the CT projection into several consecutive angular segments and reconstructing each segment individually. Although motion estimation and compensation using PAR has been developed and investigated in cardiac CT scans, its potential for reducing motion artifacts in head CT scans remains unexplored. PURPOSE: To develop a deep learning (DL) model capable of directly estimating head motion from PAR images of head CT scans and to integrate the estimated motion into an iterative reconstruction process to compensate for the motion. METHODS: Head motion is considered as a rigid transformation described by six time-variant variables, including the three variables for translation and three variables for rotation. Each motion variable is modeled using a B-spline defined by five control points (CP) along time. We split the full projections from 360° into 25 consecutive PARs and subsequently input them into a convolutional neural network (CNN) that outputs the estimated CPs for each motion variable. The estimated CPs are used to calculate the object motion in each projection, which are incorporated into the forward and backprojection of an iterative reconstruction algorithm to reconstruct the motion-compensated image. The performance of our DL model is evaluated through both simulation and phantom studies. RESULTS: The DL model achieved high accuracy in estimating head motion, as demonstrated in both the simulation study (mean absolute error (MAE) ranging from 0.28 to 0.45 mm or degree across different motion variables) and the phantom study (MAE ranging from 0.40 to 0.48 mm or degree). The resulting motion-corrected image, I D L , P A R ${I}_{DL,\ PAR}$ , exhibited a significant reduction in motion artifacts when compared to the traditional filtered back-projection reconstructions, which is evidenced both in the simulation study (image MAE drops from 178 ± $ \pm $ 33HU to 37 ± $ \pm $ 9HU, structural similarity index (SSIM) increases from 0.60 ± $ \pm $ 0.06 to 0.98 ± $ \pm $ 0.01) and the phantom study (image MAE drops from 117 ± $ \pm $ 17HU to 42 ± $ \pm $ 19HU, SSIM increases from 0.83 ± $ \pm $ 0.04 to 0.98 ± $ \pm $ 0.02). CONCLUSIONS: We demonstrate that using PAR and our proposed deep learning model enables accurate estimation of patient head motion and effectively reduces motion artifacts in the resulting head CT images.


Asunto(s)
Artefactos , Aprendizaje Profundo , Procesamiento de Imagen Asistido por Computador , Tomografía Computarizada por Rayos X , Humanos , Tomografía Computarizada por Rayos X/métodos , Procesamiento de Imagen Asistido por Computador/métodos , Cabeza/diagnóstico por imagen , Movimientos de la Cabeza , Fantasmas de Imagen
4.
Med Phys ; 51(3): 2187-2199, 2024 Mar.
Artículo en Italiano | MEDLINE | ID: mdl-38319676

RESUMEN

BACKGROUND: Efficient and accurate delineation of organs at risk (OARs) is a critical procedure for treatment planning and dose evaluation. Deep learning-based auto-segmentation of OARs has shown promising results and is increasingly being used in radiation therapy. However, existing deep learning-based auto-segmentation approaches face two challenges in clinical practice: generalizability and human-AI interaction. A generalizable and promptable auto-segmentation model, which segments OARs of multiple disease sites simultaneously and supports on-the-fly human-AI interaction, can significantly enhance the efficiency of radiation therapy treatment planning. PURPOSE: Meta's segment anything model (SAM) was proposed as a generalizable and promptable model for next-generation natural image segmentation. We further evaluated the performance of SAM in radiotherapy segmentation. METHODS: Computed tomography (CT) images of clinical cases from four disease sites at our institute were collected: prostate, lung, gastrointestinal, and head & neck. For each case, we selected the OARs important in radiotherapy treatment planning. We then compared both the Dice coefficients and Jaccard indices derived from three distinct methods: manual delineation (ground truth), automatic segmentation using SAM's 'segment anything' mode, and automatic segmentation using SAM's 'box prompt' mode that implements manual interaction via live prompts during segmentation. RESULTS: Our results indicate that SAM's segment anything mode can achieve clinically acceptable segmentation results in most OARs with Dice scores higher than 0.7. SAM's box prompt mode further improves Dice scores by 0.1∼0.5. Similar results were observed for Jaccard indices. The results show that SAM performs better for prostate and lung, but worse for gastrointestinal and head & neck. When considering the size of organs and the distinctiveness of their boundaries, SAM shows better performance for large organs with distinct boundaries, such as lung and liver, and worse for smaller organs with less distinct boundaries, like parotid and cochlea. CONCLUSIONS: Our results demonstrate SAM's robust generalizability with consistent accuracy in automatic segmentation for radiotherapy. Furthermore, the advanced box-prompt method enables the users to augment auto-segmentation interactively and dynamically, leading to patient-specific auto-segmentation in radiation therapy. SAM's generalizability across different disease sites and different modalities makes it feasible to develop a generic auto-segmentation model in radiotherapy.


Asunto(s)
Aprendizaje Profundo , Oncología por Radiación , Masculino , Humanos , Inteligencia Artificial , Redes Neurales de la Computación , Tomografía Computarizada por Rayos X/métodos , Órganos en Riesgo , Planificación de la Radioterapia Asistida por Computador/métodos , Procesamiento de Imagen Asistido por Computador/métodos
5.
IEEE Trans Med Imaging ; 43(6): 2098-2112, 2024 Jun.
Artículo en Inglés | MEDLINE | ID: mdl-38241121

RESUMEN

To address the lack of high-quality training labels in positron emission tomography (PET) imaging, weakly-supervised reconstruction methods that generate network-based mappings between prior images and noisy targets have been developed. However, the learned model has an intrinsic variance proportional to the average variance of the target image. To suppress noise and improve the accuracy and generalizability of the learned model, we propose a conditional weakly-supervised multi-task learning (MTL) strategy, in which an auxiliary task is introduced serving as an anatomical regularizer for the PET reconstruction main task. In the proposed MTL approach, we devise a novel multi-channel self-attention (MCSA) module that helps learn an optimal combination of shared and task-specific features by capturing both local and global channel-spatial dependencies. The proposed reconstruction method was evaluated on NEMA phantom PET datasets acquired at different positions in a PET/CT scanner and 26 clinical whole-body PET datasets. The phantom results demonstrate that our method outperforms state-of-the-art learning-free and weakly-supervised approaches obtaining the best noise/contrast tradeoff with a significant noise reduction of approximately 50.0% relative to the maximum likelihood (ML) reconstruction. The patient study results demonstrate that our method achieves the largest noise reductions of 67.3% and 35.5% in the liver and lung, respectively, as well as consistently small biases in 8 tumors with various volumes and intensities. In addition, network visualization reveals that adding the auxiliary task introduces more anatomical information into PET reconstruction than adding only the anatomical loss, and the developed MCSA can abstract features and retain PET image details.


Asunto(s)
Algoritmos , Procesamiento de Imagen Asistido por Computador , Fantasmas de Imagen , Tomografía de Emisión de Positrones , Aprendizaje Automático Supervisado , Humanos , Procesamiento de Imagen Asistido por Computador/métodos , Tomografía de Emisión de Positrones/métodos , Imagen de Cuerpo Entero/métodos , Tomografía Computarizada por Tomografía de Emisión de Positrones/métodos
6.
Phys Med Biol ; 69(4)2024 Feb 12.
Artículo en Inglés | MEDLINE | ID: mdl-38271727

RESUMEN

Objective. This paper presents a novel approach for addressing the intricate task of diagnosing aortic valve regurgitation (AR), a valvular disease characterized by blood leakage due to incompetence of the valve closure. Conventional diagnostic techniques require detailed evaluations of multi-modal clinical data, frequently resulting in labor-intensive and time-consuming procedures that are vulnerable to varying subjective assessment of regurgitation severity.Approach. In our research, we introduce the multi-view video contrastive network, designed to leverage multiple color Doppler imaging inputs for multi-view video processing. We leverage supervised contrastive learning as a strategic approach to tackle class imbalance and enhance the effectiveness of our feature representation learning. Specifically, we introduce a contrastive learning framework to enhance representation learning within the embedding space through inter-patient and intra-patient contrastive loss terms.Main results. We conducted extensive experiments using an in-house dataset comprising 250 echocardiography video series. Our results exhibit a substantial improvement in diagnostic accuracy for AR compared to state-of-the-art methods in terms of accuracy by 9.60%, precision by 8.67%, recall by 9.01%, andF1-score by 8.92%. These results emphasize the capacity of our approach to provide a more precise and efficient method for evaluating the severity of AR.Significance. The proposed model could quickly and accurately make decisions about the severity of AR, potentially serving as a useful prescreening tool.


Asunto(s)
Catéteres , Enfermedades de las Válvulas Cardíacas , Humanos , Ecocardiografía
7.
Med Phys ; 51(3): 2096-2107, 2024 Mar.
Artículo en Inglés | MEDLINE | ID: mdl-37776263

RESUMEN

BACKGROUND: Radiotherapy (RT) combined with cetuximab is the standard treatment for patients with inoperable head and neck cancers. Segmentation of head and neck (H&N) tumors is a prerequisite for radiotherapy planning but a time-consuming process. In recent years, deep convolutional neural networks (DCNN) have become the de facto standard for automated image segmentation. However, due to the expensive computational cost associated with enlarging the field of view in DCNNs, their ability to model long-range dependency is still limited, and this can result in sub-optimal segmentation performance for objects with background context spanning over long distances. On the other hand, Transformer models have demonstrated excellent capabilities in capturing such long-range information in several semantic segmentation tasks performed on medical images. PURPOSE: Despite the impressive representation capacity of vision transformer models, current vision transformer-based segmentation models still suffer from inconsistent and incorrect dense predictions when fed with multi-modal input data. We suspect that the power of their self-attention mechanism may be limited in extracting the complementary information that exists in multi-modal data. To this end, we propose a novel segmentation model, debuted, Cross-modal Swin Transformer (SwinCross), with cross-modal attention (CMA) module to incorporate cross-modal feature extraction at multiple resolutions. METHODS: We propose a novel architecture for cross-modal 3D semantic segmentation with two main components: (1) a cross-modal 3D Swin Transformer for integrating information from multiple modalities (PET and CT), and (2) a cross-modal shifted window attention block for learning complementary information from the modalities. To evaluate the efficacy of our approach, we conducted experiments and ablation studies on the HECKTOR 2021 challenge dataset. We compared our method against nnU-Net (the backbone of the top-5 methods in HECKTOR 2021) and other state-of-the-art transformer-based models, including UNETR and Swin UNETR. The experiments employed a five-fold cross-validation setup using PET and CT images. RESULTS: Empirical evidence demonstrates that our proposed method consistently outperforms the comparative techniques. This success can be attributed to the CMA module's capacity to enhance inter-modality feature representations between PET and CT during head-and-neck tumor segmentation. Notably, SwinCross consistently surpasses Swin UNETR across all five folds, showcasing its proficiency in learning multi-modal feature representations at varying resolutions through the cross-modal attention modules. CONCLUSIONS: We introduced a cross-modal Swin Transformer for automating the delineation of head and neck tumors in PET and CT images. Our model incorporates a cross-modality attention module, enabling the exchange of features between modalities at multiple resolutions. The experimental results establish the superiority of our method in capturing improved inter-modality correlations between PET and CT for head-and-neck tumor segmentation. Furthermore, the proposed methodology holds applicability to other semantic segmentation tasks involving different imaging modalities like SPECT/CT or PET/MRI. Code:https://github.com/yli192/SwinCross_CrossModalSwinTransformer_for_Medical_Image_Segmentation.


Asunto(s)
Neoplasias de Cabeza y Cuello , Tomografía Computarizada por Tomografía de Emisión de Positrones , Humanos , Tomografía de Emisión de Positrones , Neoplasias de Cabeza y Cuello/diagnóstico por imagen , Aprendizaje , Redes Neurales de la Computación , Procesamiento de Imagen Asistido por Computador
8.
Eur J Nucl Med Mol Imaging ; 51(2): 358-368, 2024 Jan.
Artículo en Inglés | MEDLINE | ID: mdl-37787849

RESUMEN

PURPOSE: Due to various physical degradation factors and limited counts received, PET image quality needs further improvements. The denoising diffusion probabilistic model (DDPM) was a distribution learning-based model, which tried to transform a normal distribution into a specific data distribution based on iterative refinements. In this work, we proposed and evaluated different DDPM-based methods for PET image denoising. METHODS: Under the DDPM framework, one way to perform PET image denoising was to provide the PET image and/or the prior image as the input. Another way was to supply the prior image as the network input with the PET image included in the refinement steps, which could fit for scenarios of different noise levels. 150 brain [[Formula: see text]F]FDG datasets and 140 brain [[Formula: see text]F]MK-6240 (imaging neurofibrillary tangles deposition) datasets were utilized to evaluate the proposed DDPM-based methods. RESULTS: Quantification showed that the DDPM-based frameworks with PET information included generated better results than the nonlocal mean, Unet and generative adversarial network (GAN)-based denoising methods. Adding additional MR prior in the model helped achieved better performance and further reduced the uncertainty during image denoising. Solely relying on MR prior while ignoring the PET information resulted in large bias. Regional and surface quantification showed that employing MR prior as the network input while embedding PET image as a data-consistency constraint during inference achieved the best performance. CONCLUSION: DDPM-based PET image denoising is a flexible framework, which can efficiently utilize prior information and achieve better performance than the nonlocal mean, Unet and GAN-based denoising methods.


Asunto(s)
Procesamiento de Imagen Asistido por Computador , Tomografía de Emisión de Positrones , Humanos , Procesamiento de Imagen Asistido por Computador/métodos , Tomografía de Emisión de Positrones/métodos , Relación Señal-Ruido , Modelos Estadísticos , Algoritmos
9.
JMIR Med Educ ; 9: e48904, 2023 Dec 28.
Artículo en Inglés | MEDLINE | ID: mdl-38153785

RESUMEN

BACKGROUND: Large language models, such as ChatGPT, are capable of generating grammatically perfect and human-like text content, and a large number of ChatGPT-generated texts have appeared on the internet. However, medical texts, such as clinical notes and diagnoses, require rigorous validation, and erroneous medical content generated by ChatGPT could potentially lead to disinformation that poses significant harm to health care and the general public. OBJECTIVE: This study is among the first on responsible artificial intelligence-generated content in medicine. We focus on analyzing the differences between medical texts written by human experts and those generated by ChatGPT and designing machine learning workflows to effectively detect and differentiate medical texts generated by ChatGPT. METHODS: We first constructed a suite of data sets containing medical texts written by human experts and generated by ChatGPT. We analyzed the linguistic features of these 2 types of content and uncovered differences in vocabulary, parts-of-speech, dependency, sentiment, perplexity, and other aspects. Finally, we designed and implemented machine learning methods to detect medical text generated by ChatGPT. The data and code used in this paper are published on GitHub. RESULTS: Medical texts written by humans were more concrete, more diverse, and typically contained more useful information, while medical texts generated by ChatGPT paid more attention to fluency and logic and usually expressed general terminologies rather than effective information specific to the context of the problem. A bidirectional encoder representations from transformers-based model effectively detected medical texts generated by ChatGPT, and the F1 score exceeded 95%. CONCLUSIONS: Although text generated by ChatGPT is grammatically perfect and human-like, the linguistic characteristics of generated medical texts were different from those written by human experts. Medical text generated by ChatGPT could be effectively detected by the proposed machine learning algorithms. This study provides a pathway toward trustworthy and accountable use of large language models in medicine.


Asunto(s)
Algoritmos , Inteligencia Artificial , Humanos , Desinformación , Suministros de Energía Eléctrica , Instituciones de Salud
10.
IEEE Trans Med Imaging ; PP2023 Nov 23.
Artículo en Inglés | MEDLINE | ID: mdl-37995174

RESUMEN

Position emission tomography (PET) is widely used in clinics and research due to its quantitative merits and high sensitivity, but suffers from low signal-to-noise ratio (SNR). Recently convolutional neural networks (CNNs) have been widely used to improve PET image quality. Though successful and efficient in local feature extraction, CNN cannot capture long-range dependencies well due to its limited receptive field. Global multi-head self-attention (MSA) is a popular approach to capture long-range information. However, the calculation of global MSA for 3D images has high computational costs. In this work, we proposed an efficient spatial and channel-wise encoder-decoder transformer, Spach Transformer, that can leverage spatial and channel information based on local and global MSAs. Experiments based on datasets of different PET tracers, i.e., 18F-FDG, 18F-ACBC, 18F-DCFPyL, and 68Ga-DOTATATE, were conducted to evaluate the proposed framework. Quantitative results show that the proposed Spach Transformer framework outperforms state-of-the-art deep learning architectures.

11.
Phys Med Biol ; 68(20)2023 Oct 06.
Artículo en Inglés | MEDLINE | ID: mdl-37726013

RESUMEN

Objective. Ultrasound is extensively utilized as a convenient and cost-effective method in emergency situations. Unfortunately, the limited availability of skilled clinicians in emergency hinders the wider adoption of point-of-care ultrasound. To overcome this challenge, this paper aims to aid less experienced healthcare providers in emergency lung ultrasound scans.Approach. To assist healthcare providers, it is important to have a comprehensive model that can automatically guide the entire process of lung ultrasound based on the clinician's workflow. In this paper, we propose a framework for diagnosing pneumothorax using artificial intelligence (AI) assistance. Specifically, the proposed framework for lung ultrasound scan follows the steps taken by skilled physicians. It begins with finding the appropriate transducer position on the chest to locate the pleural line accurately in B-mode. The next step involves acquiring temporal M-mode data to determine the presence of lung sliding, a crucial indicator for pneumothorax. To mimic the sequential process of clinicians, two DL models were developed. The first model focuses on quality assurance (QA) and regression of the pleural line region-of-interest, while the second model classifies lung sliding. To achieve the inference on a mobile device, a size of EfficientNet-Lite0 model was further reduced to have fewer than 3 million parameters.Main results. The results showed that both the QA and lung sliding classification models achieved over 95% in area under the receiver operating characteristic (AUC), while the ROI performance reached 89% in the dice similarity coefficient. The entire stepwise pipeline was simulated using retrospective data, yielding an AUC of 89%.Significance. The step-wise AI framework for the pneumothorax diagnosis with QA offers an intelligible guide for each clinical workflow, which achieved significantly high precision and real-time inferences.


Asunto(s)
Neumotórax , Humanos , Neumotórax/diagnóstico por imagen , Estudios Retrospectivos , Sistemas de Atención de Punto , Inteligencia Artificial , Ultrasonografía/métodos
12.
J Nucl Med ; 64(2): 188-196, 2023 02.
Artículo en Inglés | MEDLINE | ID: mdl-36522184

RESUMEN

Trustworthiness is a core tenet of medicine. The patient-physician relationship is evolving from a dyad to a broader ecosystem of health care. With the emergence of artificial intelligence (AI) in medicine, the elements of trust must be revisited. We envision a road map for the establishment of trustworthy AI ecosystems in nuclear medicine. In this report, AI is contextualized in the history of technologic revolutions. Opportunities for AI applications in nuclear medicine related to diagnosis, therapy, and workflow efficiency, as well as emerging challenges and critical responsibilities, are discussed. Establishing and maintaining leadership in AI require a concerted effort to promote the rational and safe deployment of this innovative technology by engaging patients, nuclear medicine physicians, scientists, technologists, and referring providers, among other stakeholders, while protecting our patients and society. This strategic plan was prepared by the AI task force of the Society of Nuclear Medicine and Molecular Imaging.


Asunto(s)
Inteligencia Artificial , Medicina Nuclear , Humanos , Ecosistema , Cintigrafía , Imagen Molecular
13.
Meta Radiol ; 1(3)2023 Nov.
Artículo en Inglés | MEDLINE | ID: mdl-38344271

RESUMEN

The emergence of artificial general intelligence (AGI) is transforming radiation oncology. As prominent vanguards of AGI, large language models (LLMs) such as GPT-4 and PaLM 2 can process extensive texts and large vision models (LVMs) such as the Segment Anything Model (SAM) can process extensive imaging data to enhance the efficiency and precision of radiation therapy. This paper explores full-spectrum applications of AGI across radiation oncology including initial consultation, simulation, treatment planning, treatment delivery, treatment verification, and patient follow-up. The fusion of vision data with LLMs also creates powerful multimodal models that elucidate nuanced clinical patterns. Together, AGI promises to catalyze a shift towards data-driven, personalized radiation therapy. However, these models should complement human expertise and care. This paper provides an overview of how AGI can transform radiation oncology to elevate the standard of patient care in radiation oncology, with the key insight being AGI's ability to exploit multimodal clinical data at scale.

14.
Front Radiol ; 3: 1224682, 2023.
Artículo en Inglés | MEDLINE | ID: mdl-38464946

RESUMEN

At the dawn of of Artificial General Intelligence (AGI), the emergence of large language models such as ChatGPT show promise in revolutionizing healthcare by improving patient care, expanding medical access, and optimizing clinical processes. However, their integration into healthcare systems requires careful consideration of potential risks, such as inaccurate medical advice, patient privacy violations, the creation of falsified documents or images, overreliance on AGI in medical education, and the perpetuation of biases. It is crucial to implement proper oversight and regulation to address these risks, ensuring the safe and effective incorporation of AGI technologies into healthcare systems. By acknowledging and mitigating these challenges, AGI can be harnessed to enhance patient care, medical knowledge, and healthcare processes, ultimately benefiting society as a whole.

15.
Med Image Anal ; 80: 102519, 2022 08.
Artículo en Inglés | MEDLINE | ID: mdl-35767910

RESUMEN

Recently, deep learning-based denoising methods have been gradually used for PET images denoising and have shown great achievements. Among these methods, one interesting framework is conditional deep image prior (CDIP) which is an unsupervised method that does not need prior training or a large number of training pairs. In this work, we combined CDIP with Logan parametric image estimation to generate high-quality parametric images. In our method, the kinetic model is the Logan reference tissue model that can avoid arterial sampling. The neural network was utilized to represent the images of Logan slope and intercept. The patient's computed tomography (CT) image or magnetic resonance (MR) image was used as the network input to provide anatomical information. The optimization function was constructed and solved by the alternating direction method of multipliers (ADMM) algorithm. Both simulation and clinical patient datasets demonstrated that the proposed method could generate parametric images with more detailed structures. Quantification results showed that the proposed method results had higher contrast-to-noise (CNR) improvement ratios (PET/CT datasets: 62.25%±29.93%; striatum of brain PET datasets : 129.51%±32.13%, thalamus of brain PET datasets: 128.24%±31.18%) than Gaussian filtered results (PET/CT datasets: 23.33%±18.63%; striatum of brain PET datasets: 74.71%±8.71%, thalamus of brain PET datasets: 73.02%±9.34%) and nonlocal mean (NLM) denoised results (PET/CT datasets: 37.55%±26.56%; striatum of brain PET datasets: 100.89%±16.13%, thalamus of brain PET datasets: 103.59%±16.37%).


Asunto(s)
Procesamiento de Imagen Asistido por Computador , Tomografía de Emisión de Positrones , Algoritmos , Simulación por Computador , Humanos , Procesamiento de Imagen Asistido por Computador/métodos , Tomografía Computarizada por Tomografía de Emisión de Positrones , Tomografía de Emisión de Positrones/métodos
16.
Phys Med Biol ; 67(11)2022 05 16.
Artículo en Inglés | MEDLINE | ID: mdl-35390782

RESUMEN

Objective.There are several x-ray computed tomography (CT) scanning strategies used to reduce radiation dose, such as (1) sparse-view CT, (2) low-dose CT and (3) region-of-interest (ROI) CT (called interior tomography). To further reduce the dose, sparse-view and/or low-dose CT settings can be applied together with interior tomography. Interior tomography has various advantages in terms of reducing the number of detectors and decreasing the x-ray radiation dose. However, a large patient or a small field-of-view (FOV) detector can cause truncated projections, and then the reconstructed images suffer from severe cupping artifacts. In addition, although low-dose CT can reduce the radiation exposure dose, analytic reconstruction algorithms produce image noise. Recently, many researchers have utilized image-domain deep learning (DL) approaches to remove each artifact and demonstrated impressive performances, and the theory of deep convolutional framelets supports the reason for the performance improvement.Approach.In this paper, we found that it is difficult to solve coupled artifacts using an image-domain convolutional neural network (CNN) based on deep convolutional framelets.Significance.To address the coupled problem, we decouple it into two sub-problems: (i) image-domain noise reduction inside the truncated projection to solve low-dose CT problem and (ii) extrapolation of the projection outside the truncated projection to solve the ROI CT problem. The decoupled sub-problems are solved directly with a novel proposed end-to-end learning method using dual-domain CNNs.Main results.We demonstrate that the proposed method outperforms the conventional image-domain DL methods, and a projection-domain CNN shows better performance than the image-domain CNNs commonly used by many researchers.


Asunto(s)
Aprendizaje Profundo , Algoritmos , Artefactos , Humanos , Procesamiento de Imagen Asistido por Computador/métodos , Tomografía Computarizada por Rayos X/métodos , Rayos X
17.
BMJ Open ; 12(2): e053635, 2022 02 21.
Artículo en Inglés | MEDLINE | ID: mdl-35190428

RESUMEN

OBJECTIVE: To develop simple but clinically informative risk stratification tools using a few top demographic factors and biomarkers at COVID-19 diagnosis to predict acute kidney injury (AKI) and death. DESIGN: Retrospective cohort analysis, follow-up from 1 February through 28 May 2020. SETTING: 3 teaching hospitals, 2 urban and 1 community-based in the Boston area. PARTICIPANTS: Eligible patients were at least 18 years old, tested COVID-19 positive from 1 February through 28 May 2020, and had at least two serum creatinine measurements within 30 days of a new COVID-19 diagnosis. Exclusion criteria were having chronic kidney disease or having a previous AKI within 3 months of a new COVID-19 diagnosis. MAIN OUTCOMES AND MEASURES: Time from new COVID-19 diagnosis until AKI event, time until death event. RESULTS: Among 3716 patients, there were 1855 (49.9%) males and the average age was 58.6 years (SD 19.2 years). Age, sex, white blood cell, haemoglobin, platelet, C reactive protein (CRP) and D-dimer levels were most strongly associated with AKI and/or death. We created risk scores using these variables predicting AKI within 3 days and death within 30 days of a new COVID-19 diagnosis. Area under the curve (AUC) for predicting AKI within 3 days was 0.785 (95% CI 0.758 to 0.813) and AUC for death within 30 days was 0.861 (95% CI 0.843 to 0.878). Haemoglobin was the most predictive component for AKI, and age the most predictive for death. Predictive accuracies using all study variables were similar to using the simplified scores. CONCLUSION: Simple risk scores using age, sex, a complete blood cell count, CRP and D-dimer were highly predictive of AKI and death and can help simplify and better inform clinical decision making.


Asunto(s)
Lesión Renal Aguda , COVID-19 , Insuficiencia Renal Crónica , Lesión Renal Aguda/complicaciones , Lesión Renal Aguda/diagnóstico , Adolescente , Prueba de COVID-19 , Estudios de Cohortes , Hospitales , Humanos , Masculino , Persona de Mediana Edad , Insuficiencia Renal Crónica/complicaciones , Insuficiencia Renal Crónica/diagnóstico , Estudios Retrospectivos , Medición de Riesgo , Factores de Riesgo , SARS-CoV-2
18.
Med Phys ; 49(4): 2373-2385, 2022 Apr.
Artículo en Inglés | MEDLINE | ID: mdl-35048390

RESUMEN

PURPOSE: Arterial spin labeling (ASL) magnetic resonance imaging (MRI) is an advanced noninvasive imaging technology that can measure cerebral blood flow (CBF) quantitatively without a contrast agent injection or radiation exposure. However, because of the weak labeling, conventional ASL images usually suffer from low signal-to-noise ratio (SNR), poor spatial resolution, and long acquisition time. Therefore, a method that can simultaneously improve the spatial resolution and SNR is needed. METHODS: In this work, we proposed an unsupervised superresolution (SR) method to improve ASL image resolution based on a pyramid of generative adversarial networks (GAN). Through layer-by-layer training, the generators can learn features from the coarsest to the finest. The last layer's generator that contains fine details and textures was used to generate the final SR ASL images. In our proposed framework, the corresponding T1-weighted MR image was supplied as a second-channel input of the generators to provide high-resolution prior information. In addition, a low-pass-filter loss term was included to suppress the noise of the original ASL images. To evaluate the performance of the proposed framework, a simulation study and two real-patient experiments based on the in vivo datasets obtained from three healthy subjects on a 3T MR scanner were conducted, regarding the low-resolution (LR) to normal-resolution (NR) and the NR-to-SR tasks. The proposed method was compared to the nearest neighbor interpolation, trilinear interpolation, third-order B-splines interpolation methods, and deep image prior (DIP) with the peak signal-to-noise ratio (PSNR) and structural similarity index (SSIM) as the quantification metrics. The averaged ASL images acquired with 44 min acquisition time were used as the ground truth for real-patient LR-to-NR study. The ablation studies of low-pass-filter loss term and T1-weighted MR image were performed based on simulation data. RESULTS: For the simulation study, results show that the proposed method achieved significantly higher PSNR ( p $p$ -value < $<$ 0.05) and SSIM ( p $p$ -value < $<$ 0.05) than the nearest neighbor interpolation, trilinear interpolation, third-order B-splines interpolation, and DIP methods. For the real-patient LR-to-NR experiment, results show that the proposed method can generate high-quality SR ASL images with clearer structure boundaries and low noise levels and has the highest mean PSNR and SSIM. For real-patient NR-to-SR tasks, the structure of the results using the proposed method is sharper and clearer, which are the most similar to the structure of the reference 44 min acquisition image than other methods. The proposed method also shows the ability to remove artifacts in the NR image while superresolution. The ablation study verified that the low-pass-filter loss term and T1-weighted MR image are necessary for the proposed method. CONCLUSIONS: The proposed unsupervised multiscale GAN framework can simultaneously improve spatial resolution and reduce image noise. Experiment results from simulation data and three healthy subjects show that the proposed method achieves better performance than the nearest neighbor interpolation, the trilinear interpolation, the third-order B-splines interpolation, and DIP methods.


Asunto(s)
Procesamiento de Imagen Asistido por Computador , Imagen por Resonancia Magnética , Artefactos , Humanos , Procesamiento de Imagen Asistido por Computador/métodos , Relación Señal-Ruido , Marcadores de Spin
19.
Diagnostics (Basel) ; 12(1)2022 Jan 03.
Artículo en Inglés | MEDLINE | ID: mdl-35054267

RESUMEN

Imaging plays an important role in assessing the severity of COVID-19 pneumonia. Recent COVID-19 research indicates that the disease progress propagates from the bottom of the lungs to the top. However, chest radiography (CXR) cannot directly provide a quantitative metric of radiographic opacities, and existing AI-assisted CXR analysis methods do not quantify the regional severity. In this paper, to assist the regional analysis, we developed a fully automated framework using deep learning-based four-region segmentation and detection models to assist the quantification of COVID-19 pneumonia. Specifically, a segmentation model is first applied to separate left and right lungs, and then a detection network of the carina and left hilum is used to separate upper and lower lungs. To improve the segmentation performance, an ensemble strategy with five models is exploited. We evaluated the clinical relevance of the proposed method compared with the radiographic assessment of the quality of lung edema (RALE) annotated by physicians. Mean intensities of segmented four regions indicate a positive correlation to the regional extent and density scores of pulmonary opacities based on the RALE. Therefore, the proposed method can accurately assist the quantification of regional pulmonary opacities of COVID-19 pneumonia patients.

20.
Eur Radiol ; 32(4): 2235-2245, 2022 Apr.
Artículo en Inglés | MEDLINE | ID: mdl-34988656

RESUMEN

BACKGROUND: Main challenges for COVID-19 include the lack of a rapid diagnostic test, a suitable tool to monitor and predict a patient's clinical course and an efficient way for data sharing among multicenters. We thus developed a novel artificial intelligence system based on deep learning (DL) and federated learning (FL) for the diagnosis, monitoring, and prediction of a patient's clinical course. METHODS: CT imaging derived from 6 different multicenter cohorts were used for stepwise diagnostic algorithm to diagnose COVID-19, with or without clinical data. Patients with more than 3 consecutive CT images were trained for the monitoring algorithm. FL has been applied for decentralized refinement of independently built DL models. RESULTS: A total of 1,552,988 CT slices from 4804 patients were used. The model can diagnose COVID-19 based on CT alone with the AUC being 0.98 (95% CI 0.97-0.99), and outperforms the radiologist's assessment. We have also successfully tested the incorporation of the DL diagnostic model with the FL framework. Its auto-segmentation analyses co-related well with those by radiologists and achieved a high Dice's coefficient of 0.77. It can produce a predictive curve of a patient's clinical course if serial CT assessments are available. INTERPRETATION: The system has high consistency in diagnosing COVID-19 based on CT, with or without clinical data. Alternatively, it can be implemented on a FL platform, which would potentially encourage the data sharing in the future. It also can produce an objective predictive curve of a patient's clinical course for visualization. KEY POINTS: • CoviDet could diagnose COVID-19 based on chest CT with high consistency; this outperformed the radiologist's assessment. Its auto-segmentation analyses co-related well with those by radiologists and could potentially monitor and predict a patient's clinical course if serial CT assessments are available. It can be integrated into the federated learning framework. • CoviDet can be used as an adjunct to aid clinicians with the CT diagnosis of COVID-19 and can potentially be used for disease monitoring; federated learning can potentially open opportunities for global collaboration.


Asunto(s)
Inteligencia Artificial , COVID-19 , Algoritmos , Humanos , Radiólogos , Tomografía Computarizada por Rayos X/métodos
SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA
...