Your browser doesn't support javascript.
loading
Montrer: 20 | 50 | 100
Résultats 1 - 20 de 126
Filtrer
1.
Article de Anglais | MEDLINE | ID: mdl-38968015

RÉSUMÉ

Despite promising advancements in deep learning in medical domains, challenges still remain owing to data scarcity, compounded by privacy concerns and data ownership disputes. Recent explorations of distributed-learning paradigms, particularly federated learning, have aimed to mitigate these challenges. However, these approaches are often encumbered by substantial communication and computational overhead, and potential vulnerabilities in privacy safeguards. Therefore, we propose a self-supervised masked sampling distillation technique called MS-DINO, tailored to the vision transformer architecture. This approach removes the need for incessant communication and strengthens privacy using a modified encryption mechanism inherent to the vision transformer while minimizing the computational burden on client-side devices. Rigorous evaluations across various tasks confirmed that our method outperforms existing self-supervised distributed learning strategies and fine-tuned baselines.

2.
Neural Netw ; 178: 106435, 2024 Jun 22.
Article de Anglais | MEDLINE | ID: mdl-38970945

RÉSUMÉ

Understanding the training dynamics of deep ReLU networks is a significant area of interest in deep learning. However, there remains a lack of complete elucidation regarding the weight vector dynamics, even for single ReLU neurons. To bridge this gap, our study delves into the training dynamics of the gradient flow w(t) for single ReLU neurons under the square loss, dissecting it into its magnitude ‖w(t)‖ and angle φ(t) components. Through this decomposition, we establish upper and lower bounds on these components to elucidate the convergence dynamics. Furthermore, we demonstrate the empirical extension of our findings to general two-layer multi-neuron networks. All theoretical results are generalized to the gradient descent method and rigorously verified through experiments.

3.
Endocrinol Metab (Seoul) ; 39(3): 500-510, 2024 Jun.
Article de Anglais | MEDLINE | ID: mdl-38721637

RÉSUMÉ

BACKGRUOUND: Osteoporosis is the most common metabolic bone disease and can cause fragility fractures. Despite this, screening utilization rates for osteoporosis remain low among populations at risk. Automated bone mineral density (BMD) estimation using computed tomography (CT) can help bridge this gap and serve as an alternative screening method to dual-energy X-ray absorptiometry (DXA). METHODS: The feasibility of an opportunistic and population agnostic screening method for osteoporosis using abdominal CT scans without bone densitometry phantom-based calibration was investigated in this retrospective study. A total of 268 abdominal CT-DXA pairs and 99 abdominal CT studies without DXA scores were obtained from an oncology specialty clinic in the Republic of Korea. The center axial CT slices from the L1, L2, L3, and L4 lumbar vertebrae were annotated with the CT slice level and spine segmentation labels for each subject. Deep learning models were trained to localize the center axial slice from the CT scan of the torso, segment the vertebral bone, and estimate BMD for the top four lumbar vertebrae. RESULTS: Automated vertebra-level DXA measurements showed a mean absolute error (MAE) of 0.079, Pearson's r of 0.852 (P<0.001), and R2 of 0.714. Subject-level predictions on the held-out test set had a MAE of 0.066, Pearson's r of 0.907 (P<0.001), and R2 of 0.781. CONCLUSION: CT scans collected during routine examinations without bone densitometry calibration can be used to generate DXA BMD predictions.


Sujet(s)
Absorptiométrie photonique , Densité osseuse , Vertèbres lombales , Ostéoporose , Tomodensitométrie , Humains , Ostéoporose/imagerie diagnostique , Ostéoporose/diagnostic , Tomodensitométrie/méthodes , Femelle , Absorptiométrie photonique/méthodes , Études rétrospectives , Adulte d'âge moyen , Mâle , Sujet âgé , Vertèbres lombales/imagerie diagnostique , Dépistage de masse/méthodes , République de Corée , Apprentissage profond
4.
J Cardiovasc Comput Tomogr ; 18(4): 401-407, 2024.
Article de Anglais | MEDLINE | ID: mdl-38677958

RÉSUMÉ

BACKGROUND: Positive remodeling is an integral part of the vascular adaptation process during the development of atherosclerosis, which can be detected by coronary computed tomography angiography (CTA). METHODS: A total of 426 patients who underwent both coronary CTA and optical coherence tomography (OCT) were included. Four machine learning (ML) models, gradient boosting machine (GBM), random forest (RF), deep learning (DL), and support vector machine (SVM), were employed to detect specific plaque features. A total of 15 plaque features assessed by OCT were analyzed. The variable importance ranking was used to identify the features most closely associated with positive remodeling. RESULTS: In the variable importance ranking, lipid index and maximal calcification arc were consistently ranked high across all four ML models. Lipid index and maximal calcification arc were correlated with positive remodeling, showing pronounced influence at the lower range and diminishing influence at the higher range. Patients with more plaques with positive remodeling throughout their entire coronary trees had higher low-density lipoprotein cholesterol levels and were associated with a higher incidence of cardiovascular events during 5-year follow-up (Hazard ratio 2.10 [1.26-3.48], P â€‹= â€‹0.004). CONCLUSION: Greater lipid accumulation and less calcium burden were important features associated with positive remodeling in the coronary arteries. The number of coronary plaques with positive remodeling was associated with a higher incidence of cardiovascular events.


Sujet(s)
Angiographie par tomodensitométrie , Coronarographie , Maladie des artères coronaires , Vaisseaux coronaires , Phénotype , Plaque d'athérosclérose , Valeur prédictive des tests , Tomographie par cohérence optique , Calcification vasculaire , Remodelage vasculaire , Humains , Mâle , Femelle , Maladie des artères coronaires/imagerie diagnostique , Adulte d'âge moyen , Vaisseaux coronaires/imagerie diagnostique , Sujet âgé , Calcification vasculaire/imagerie diagnostique , Pronostic , Études rétrospectives , Marqueurs biologiques/sang , Facteurs temps , Lipides/sang , Facteurs de risque , Apprentissage profond
5.
Nat Commun ; 15(1): 2323, 2024 Mar 14.
Article de Anglais | MEDLINE | ID: mdl-38485914

RÉSUMÉ

Recent successes of foundation models in artificial intelligence have prompted the emergence of large-scale chemical pre-trained models. Despite the growing interest in large molecular pre-trained models that provide informative representations for downstream tasks, attempts for multimodal pre-training approaches on the molecule domain were limited. To address this, here we present a multimodal molecular pre-trained model that incorporates the modalities of structure and biochemical properties, drawing inspiration from recent advances in multimodal learning techniques. Our proposed model pipeline of data handling and training objectives aligns the structure/property features in a common embedding space, which enables the model to regard bidirectional information between the molecules' structure and properties. These contributions emerge synergistic knowledge, allowing us to tackle both multimodal and unimodal downstream tasks through a single model. Through extensive experiments, we demonstrate that our model has the capabilities to solve various meaningful chemical challenges, including conditional molecule generation, property prediction, molecule classification, and reaction prediction.

6.
Neuroimage ; 291: 120571, 2024 May 01.
Article de Anglais | MEDLINE | ID: mdl-38518829

RÉSUMÉ

DCE-MRI provides information about vascular permeability and tissue perfusion through the acquisition of pharmacokinetic parameters. However, traditional methods for estimating these pharmacokinetic parameters involve fitting tracer kinetic models, which often suffer from computational complexity and low accuracy due to noisy arterial input function (AIF) measurements. Although some deep learning approaches have been proposed to tackle these challenges, most existing methods rely on supervised learning that requires paired input DCE-MRI and labeled pharmacokinetic parameter maps. This dependency on labeled data introduces significant time and resource constraints and potential noise in the labels, making supervised learning methods often impractical. To address these limitations, we present a novel unpaired deep learning method for estimating pharmacokinetic parameters and the AIF using a physics-driven CycleGAN approach. Our proposed CycleGAN framework is designed based on the underlying physics model, resulting in a simpler architecture with a single generator and discriminator pair. Crucially, our experimental results indicate that our method does not necessitate separate AIF measurements and produces more reliable pharmacokinetic parameters than other techniques.


Sujet(s)
Produits de contraste , Apprentissage profond , Humains , Produits de contraste/pharmacocinétique , Simulation numérique , Amélioration d'image/méthodes , Imagerie par résonance magnétique/méthodes , Algorithmes , Reproductibilité des résultats
7.
IEEE J Biomed Health Inform ; 28(3): 1692-1703, 2024 Mar.
Article de Anglais | MEDLINE | ID: mdl-38133977

RÉSUMÉ

Automatic Speech Recognition (ASR) is a technology that converts spoken words into text, facilitating interaction between humans and machines. One of the most common applications of ASR is Speech-To-Text (STT) technology, which simplifies user workflows by transcribing spoken words into text. In the medical field, STT has the potential to significantly reduce the workload of clinicians who rely on typists to transcribe their voice recordings. However, developing an STT model for the medical domain is challenging due to the lack of sufficient speech and text datasets. To address this issue, we propose a medical-domain text correction method that modifies the output text of a general STT system using the Vision Language Pre-training (VLP) method. VLP combines textual and visual information to correct text based on image knowledge. Our extensive experiments demonstrate that the proposed method offers quantitatively and clinically significant improvements in STT performance in the medical field. We further show that multi-modal understanding of image and text information outperforms single-modal understanding using only text information.


Sujet(s)
Parole , Voix , Humains , Langage
8.
Med Image Anal ; 91: 103021, 2024 Jan.
Article de Anglais | MEDLINE | ID: mdl-37952385

RÉSUMÉ

The escalating demand for artificial intelligence (AI) systems that can monitor and supervise human errors and abnormalities in healthcare presents unique challenges. Recent advances in vision-language models reveal the challenges of monitoring AI by understanding both visual and textual concepts and their semantic correspondences. However, there has been limited success in the application of vision-language models in the medical domain. Current vision-language models and learning strategies for photographic images and captions call for a web-scale data corpus of image and text pairs which is not often feasible in the medical domain. To address this, we present a model named medical cross-attention vision-language model (Medical X-VL), which leverages key components to be tailored for the medical domain. The model is based on the following components: self-supervised unimodal models in medical domain and a fusion encoder to bridge them, momentum distillation, sentencewise contrastive learning for medical reports, and sentence similarity-adjusted hard negative mining. We experimentally demonstrated that our model enables various zero-shot tasks for monitoring AI, ranging from the zero-shot classification to zero-shot error correction. Our model outperformed current state-of-the-art models in two medical image datasets, suggesting a novel clinical application of our monitoring AI model to alleviate human errors. Our method demonstrates a more specialized capacity for fine-grained understanding, which presents a distinct advantage particularly applicable to the medical domain.


Sujet(s)
Intelligence artificielle , Radiologie , Humains , Radiographie , Apprentissage , Langage
9.
Med Image Anal ; 91: 103022, 2024 Jan.
Article de Anglais | MEDLINE | ID: mdl-37976870

RÉSUMÉ

Blood vessel segmentation in medical imaging is one of the essential steps for vascular disease diagnosis and interventional planning in a broad spectrum of clinical scenarios in image-based medicine and interventional medicine. Unfortunately, manual annotation of the vessel masks is challenging and resource-intensive due to subtle branches and complex structures. To overcome this issue, this paper presents a self-supervised vessel segmentation method, dubbed the contrastive diffusion adversarial representation learning (C-DARL) model. Our model is composed of a diffusion module and a generation module that learns the distribution of multi-domain blood vessel data by generating synthetic vessel images from diffusion latent. Moreover, we employ contrastive learning through a mask-based contrastive loss so that the model can learn more realistic vessel representations. To validate the efficacy, C-DARL is trained using various vessel datasets, including coronary angiograms, abdominal digital subtraction angiograms, and retinal imaging. Experimental results confirm that our model achieves performance improvement over baseline methods with noise robustness, suggesting the effectiveness of C-DARL for vessel segmentation.Our source code is available at https://github.com/boahK/MEDIA_CDARL.2.


Sujet(s)
Apprentissage , Rétine , Humains , Coronarographie , Diffusion , Logiciel , Traitement d'image par ordinateur
10.
Sci Rep ; 13(1): 22992, 2023 12 27.
Article de Anglais | MEDLINE | ID: mdl-38151502

RÉSUMÉ

Patients with acute coronary syndromes caused by plaque erosion might be managed conservatively without stenting. Currently, the diagnosis of plaque erosion requires an invasive imaging procedure. We sought to develop a deep learning (DL) model that enables an accurate diagnosis of plaque erosion using coronary computed tomography angiography (CTA). A total of 532 CTA scans from 395 patients were used to develop a DL model: 426 CTA scans from 316 patients for training and internal validation, and 106 separate scans from 79 patients for validation. Momentum Distillation-enhanced Composite Transformer Attention (MD-CTA), a novel DL model that can effectively process the entire set of CTA scans to diagnose plaque erosion, was developed. The novel DL model, compared to the convolution neural network, showed significantly improved AUC (0.899 [0.841-0.957] vs. 0.724 [0.622-0.826]), sensitivity (87.1 [70.2-96.4] vs. 71.0 [52.0-85.8]), and specificity (85.3 [75.3-92.4] vs. 68.0 [56.2-78.3]), respectively, for the patient-level prediction. Similar results were obtained at the slice-level prediction AUC (0.897 [0.890-0.904] vs. 0.757 [0.744-0.770]), sensitivity (82.2 [79.8-84.3] vs. 68.9 [66.2-71.6]), and specificity (80.1 [79.1-81.0] vs. 67.3 [66.3-68.4]), respectively. This newly developed DL model enables an accurate CT diagnosis of plaque erosion, which might enable cardiologists to provide tailored therapy without invasive procedures.Clinical Trial Registration: http://www.clinicaltrials.gov , NCT04523194.


Sujet(s)
Maladie des artères coronaires , Apprentissage profond , Plaque d'athérosclérose , Humains , Maladie des artères coronaires/imagerie diagnostique , Coronarographie/méthodes , Plaque d'athérosclérose/imagerie diagnostique , Tomodensitométrie , Vaisseaux coronaires/imagerie diagnostique
11.
ArXiv ; 2023 Jul 31.
Article de Anglais | MEDLINE | ID: mdl-37791106

RÉSUMÉ

Blood vessel segmentation in medical imaging is one of the essential steps for vascular disease diagnosis and interventional planning in a broad spectrum of clinical scenarios in image-based medicine and interventional medicine. Unfortunately, manual annotation of the vessel masks is challenging and resource-intensive due to subtle branches and complex structures. To overcome this issue, this paper presents a self-supervised vessel segmentation method, dubbed the contrastive diffusion adversarial representation learning (C-DARL) model. Our model is composed of a diffusion module and a generation module that learns the distribution of multi-domain blood vessel data by generating synthetic vessel images from diffusion latent. Moreover, we employ contrastive learning through a mask-based contrastive loss so that the model can learn more realistic vessel representations. To validate the efficacy, C-DARL is trained using various vessel datasets, including coronary angiograms, abdominal digital subtraction angiograms, and retinal imaging. Experimental results confirm that our model achieves performance improvement over baseline methods with noise robustness, suggesting the effectiveness of C-DARL for vessel segmentation.

12.
Korean J Radiol ; 24(8): 807-820, 2023 08.
Article de Anglais | MEDLINE | ID: mdl-37500581

RÉSUMÉ

OBJECTIVE: To assess whether computed tomography (CT) conversion across different scan parameters and manufacturers using a routable generative adversarial network (RouteGAN) can improve the accuracy and variability in quantifying interstitial lung disease (ILD) using a deep learning-based automated software. MATERIALS AND METHODS: This study included patients with ILD who underwent thin-section CT. Unmatched CT images obtained using scanners from four manufacturers (vendors A-D), standard- or low-radiation doses, and sharp or medium kernels were classified into groups 1-7 according to acquisition conditions. CT images in groups 2-7 were converted into the target CT style (Group 1: vendor A, standard dose, and sharp kernel) using a RouteGAN. ILD was quantified on original and converted CT images using a deep learning-based software (Aview, Coreline Soft). The accuracy of quantification was analyzed using the dice similarity coefficient (DSC) and pixel-wise overlap accuracy metrics against manual quantification by a radiologist. Five radiologists evaluated quantification accuracy using a 10-point visual scoring system. RESULTS: Three hundred and fifty CT slices from 150 patients (mean age: 67.6 ± 10.7 years; 56 females) were included. The overlap accuracies for quantifying total abnormalities in groups 2-7 improved after CT conversion (original vs. converted: 0.63 vs. 0.68 for DSC, 0.66 vs. 0.70 for pixel-wise recall, and 0.68 vs. 0.73 for pixel-wise precision; P < 0.002 for all). The DSCs of fibrosis score, honeycombing, and reticulation significantly increased after CT conversion (0.32 vs. 0.64, 0.19 vs. 0.47, and 0.23 vs. 0.54, P < 0.002 for all), whereas those of ground-glass opacity, consolidation, and emphysema did not change significantly or decreased slightly. The radiologists' scores were significantly higher (P < 0.001) and less variable on converted CT. CONCLUSION: CT conversion using a RouteGAN can improve the accuracy and variability of CT images obtained using different scan parameters and manufacturers in deep learning-based quantification of ILD.


Sujet(s)
Emphysème , Pneumopathies interstitielles , Emphysème pulmonaire , Femelle , Humains , Adulte d'âge moyen , Sujet âgé , Pneumopathies interstitielles/imagerie diagnostique , Tomodensitométrie/méthodes , Poumon/imagerie diagnostique
13.
Korean J Radiol ; 24(6): 541-552, 2023 06.
Article de Anglais | MEDLINE | ID: mdl-37271208

RÉSUMÉ

OBJECTIVE: Detection of pneumoperitoneum using abdominal radiography, particularly in the supine position, is often challenging. This study aimed to develop and externally validate a deep learning model for the detection of pneumoperitoneum using supine and erect abdominal radiography. MATERIALS AND METHODS: A model that can utilize "pneumoperitoneum" and "non-pneumoperitoneum" classes was developed through knowledge distillation. To train the proposed model with limited training data and weak labels, it was trained using a recently proposed semi-supervised learning method called distillation for self-supervised and self-train learning (DISTL), which leverages the Vision Transformer. The proposed model was first pre-trained with chest radiographs to utilize common knowledge between modalities, fine-tuned, and self-trained on labeled and unlabeled abdominal radiographs. The proposed model was trained using data from supine and erect abdominal radiographs. In total, 191212 chest radiographs (CheXpert data) were used for pre-training, and 5518 labeled and 16671 unlabeled abdominal radiographs were used for fine-tuning and self-supervised learning, respectively. The proposed model was internally validated on 389 abdominal radiographs and externally validated on 475 and 798 abdominal radiographs from the two institutions. We evaluated the performance in diagnosing pneumoperitoneum using the area under the receiver operating characteristic curve (AUC) and compared it with that of radiologists. RESULTS: In the internal validation, the proposed model had an AUC, sensitivity, and specificity of 0.881, 85.4%, and 73.3% and 0.968, 91.1, and 95.0 for supine and erect positions, respectively. In the external validation at the two institutions, the AUCs were 0.835 and 0.852 for the supine position and 0.909 and 0.944 for the erect position. In the reader study, the readers' performances improved with the assistance of the proposed model. CONCLUSION: The proposed model trained with the DISTL method can accurately detect pneumoperitoneum on abdominal radiography in both the supine and erect positions.


Sujet(s)
Apprentissage profond , Humains , Études rétrospectives , Radiographie abdominale , Radiographie , Apprentissage machine supervisé , Radiographie thoracique/méthodes
14.
IEEE Trans Pattern Anal Mach Intell ; 45(10): 12179-12191, 2023 Oct.
Article de Anglais | MEDLINE | ID: mdl-37352089

RÉSUMÉ

There are many recent research efforts to fine-tune a pre-trained generator with a few target images to generate images of a novel domain. Unfortunately, these methods often suffer from overfitting or under-fitting when fine-tuned with a single target image. To address this, here we present a novel single-shot GAN adaptation method through unified CLIP space manipulations. Specifically, our model employs a two-step training strategy: reference image search in the source generator using a CLIP-guided latent optimization, followed by generator fine-tuning with a novel loss function that imposes CLIP space consistency between the source and adapted generators. To further improve the adapted model to produce spatially consistent samples with respect to the source generator, we also propose contrastive regularization for patchwise relationships in the CLIP space. Experimental results show that our model generates diverse outputs with the target texture and outperforms the baseline models both qualitatively and quantitatively. Furthermore, we show that our CLIP space manipulation strategy allows more effective attribute editing.

15.
IEEE J Biomed Health Inform ; 27(8): 4143-4153, 2023 08.
Article de Anglais | MEDLINE | ID: mdl-37192031

RÉSUMÉ

Gastric endoscopic screening is an effective way to decide appropriate gastric cancer treatment at an early stage, reducing gastric cancer-associated mortality rate. Although artificial intelligence has brought a great promise to assist pathologist to screen digitalized endoscopic biopsies, existing artificial intelligence systems are limited to be utilized in planning gastric cancer treatment. We propose a practical artificial intelligence-based decision support system that enables five subclassifications of gastric cancer pathology, which can be directly matched to general gastric cancer treatment guidance. The proposed framework is designed to efficiently differentiate multi-classes of gastric cancer through multiscale self-attention mechanism using 2-stage hybrid vision transformer networks, by mimicking the way how human pathologists understand histology. The proposed system demonstrates its reliable diagnostic performance by achieving class-average sensitivity of above 0.85 for multicentric cohort tests. Moreover, the proposed system demonstrates its great generalization capability on gastrointestinal track organ cancer by achieving the best class-average sensitivity among contemporary networks. Furthermore, in the observational study, artificial intelligence-assisted pathologists show significantly improved diagnostic sensitivity within saved screening time compared to human pathologists. Our results demonstrate that the proposed artificial intelligence system has a great potential for providing presumptive pathologic opinion and supporting decision of appropriate gastric cancer treatment in practical clinical settings.


Sujet(s)
Tumeurs de l'estomac , Humains , Intelligence artificielle , Biopsie , Alimentations électriques , Endoscopie , Tumeurs de l'estomac/thérapie
16.
Sci Rep ; 13(1): 2432, 2023 02 10.
Article de Anglais | MEDLINE | ID: mdl-36765086

RÉSUMÉ

Healed coronary plaques, morphologically characterized by a layered phenotype, are signs of previous plaque destabilization and healing. Recent optical coherence tomography (OCT) studies demonstrated that layered plaque is associated with higher levels of local and systemic inflammation and rapid plaque progression. However, the diagnosis of layered plaque needs expertise in OCT image analysis and is susceptible to inter-observer variability. We developed a deep learning (DL) model for an accurate diagnosis of layered plaque. A Visual Transformer (ViT)-based DL model that integrates information from adjacent frames emulating the cardiologists who review consecutive OCT frames to make a diagnosis was developed and compared with the standard convolutional neural network (CNN) model. A total of 237,021 cross-sectional OCT images from 581 patients collected from 8 sites were used for training and internal validation, and 65,394 images from 292 patients collected from another site were used for external validation. In the five-fold cross-validation, the ViT-based model provided better performance (area under the curve [AUC]: 0.860; 95% confidence interval [CI]: 0.855-0.866) than the standard CNN-based model (AUC: 0.799; 95% CI: 0.792-0.805). The ViT-based model (AUC: 0.845; 95% CI: 0.837-0.853) also surpassed the standard CNN-based model (AUC: 0.791; 95% CI: 0.782-0.800) in the external validation. The ViT-based DL model can accurately diagnose a layered plaque, which could help risk stratification for cardiac events.


Sujet(s)
Apprentissage profond , Infarctus du myocarde , Plaque d'athérosclérose , Humains , Études transversales , Plaque d'athérosclérose/imagerie diagnostique , , Tomographie par cohérence optique/méthodes
17.
Med Phys ; 50(4): 2263-2278, 2023 Apr.
Article de Anglais | MEDLINE | ID: mdl-36341576

RÉSUMÉ

BACKGROUND: Chest X-ray (CXR) images are commonly used to show the internal structure of the human body without invasive intervention. The quality of CXR is an important factor as it affects the accuracy of a clinical diagnosis. Unfortunately, it is difficult to always get good quality CXR scans due to noises and scatters. PURPOSE: Recently, wavelet directional CycleGAN (WavCycleGAN) has shown promising results in image restoration tasks by removing noise and artifacts without sacrificing high-frequency components of the input image. Unfortunately, WavCycleGAN directly reconstructs wavelet directional images that require a wavelet transform in both the training and test phases, resulting in additional processing steps and unnatural artifacts originating from the wavelet domain image. In addition, WavCycleGAN can only process artifact-related subbands, so it is difficult to apply WavCycleGAN when different levels of artifacts are present in all subbands. To address this, here we present a novel unsupervised CXR image restoration scheme with similar or even better artifact removal performance than WavCycleGAN in spite of wavelet transform being only applied in the training phase. METHODS: We introduce a novel wavelet subband discriminator which can be combined with CycleGAN or switchable CycleGAN, where wavelet transform is applied only in the training phase for discriminators to match the distribution of wavelet subband components. In our framework, the image restoration network can be still applied in the image domain to prevent unnatural artifacts of the wavelet domain image with the help of the image-domain cycle-consistency loss. In addition, using wavelet subband discriminator makes it possible to remove artifacts in all subbands by utilizing frequency-specific wavelet subband discriminators. RESULTS: Through extensive experiments for noise and scatter removal in CXRs, we confirm that our method provides competitive performance compared to existing approaches without additional processing steps in the test phase. Furthermore, we show that our wavelet subband discriminator combined with the switchable CycleGAN can provide the flexibility by generating different levels of artifact removal. CONCLUSIONS: The proposed wavelet subband discriminator can be combined with the existing CycleGAN or switchable CycleGAN structures to construct an efficient unsupervised CXR image reconstruction. The advantage of our wavelet subband discriminator-based CXR image restoration is that, unlike traditional WavCycleGAN, it does not require any additional processing steps in the testing phase and does not generate unnatural artifacts originating from the wavelet domain image. We believe that our wavelet subband discriminator can be applied to various CXR image applications.


Sujet(s)
Algorithmes , Traitement d'image par ordinateur , Humains , Rayons X , Traitement d'image par ordinateur/méthodes , Analyse en ondelettes , Tomodensitométrie/méthodes , Artéfacts
18.
IEEE Trans Med Imaging ; 42(4): 922-934, 2023 04.
Article de Anglais | MEDLINE | ID: mdl-36342993

RÉSUMÉ

Patient scans from MRI often suffer from noise, which hampers the diagnostic capability of such images. As a method to mitigate such artifacts, denoising is largely studied both within the medical imaging community and beyond the community as a general subject. However, recent deep neural network-based approaches mostly rely on the minimum mean squared error (MMSE) estimates, which tend to produce a blurred output. Moreover, such models suffer when deployed in real-world situations: out-of-distribution data, and complex noise distributions that deviate from the usual parametric noise models. In this work, we propose a new denoising method based on score-based reverse diffusion sampling, which overcomes all the aforementioned drawbacks. Our network, trained only with coronal knee scans, excels even on out-of-distribution in vivo liver MRI data, contaminated with a complex mixture of noise. Even more, we propose a method to enhance the resolution of the denoised image with the same network. With extensive experiments, we show that our method establishes state-of-the-art performance while having desirable properties which prior MMSE denoisers did not have: flexibly choosing the extent of denoising, and quantifying uncertainty.


Sujet(s)
Imagerie par résonance magnétique , , Humains , Artéfacts
19.
Med Image Anal ; 83: 102651, 2023 01.
Article de Anglais | MEDLINE | ID: mdl-36327653

RÉSUMÉ

In contrast to 2-D ultrasound (US) for uniaxial plane imaging, a 3-D US imaging system can visualize a volume along three axial planes. This allows for a full view of the anatomy, which is useful for gynecological (GYN) and obstetrical (OB) applications. Unfortunately, the 3-D US has an inherent limitation in resolution compared to the 2-D US. In the case of 3-D US with a 3-D mechanical probe, for example, the image quality is comparable along the beam direction, but significant deterioration in image quality is often observed in the other two axial image planes. To address this, here we propose a novel unsupervised deep learning approach to improve 3-D US image quality. In particular, using unmatched high-quality 2-D US images as a reference, we trained a recently proposed switchable CycleGAN architecture so that every mapping plane in 3-D US can learn the image quality of 2-D US images. Thanks to the switchable architecture, our network can also provide real-time control of image enhancement level based on user preference, which is ideal for a user-centric scanner setup. Extensive experiments with clinical evaluation confirm that our method offers significantly improved image quality as well user-friendly flexibility.


Sujet(s)
Contrôle de qualité , Humains
20.
IEEE Trans Med Imaging ; 42(7): 2091-2105, 2023 Jul.
Article de Anglais | MEDLINE | ID: mdl-36318558

RÉSUMÉ

The widespread application of artificial intelligence in health research is currently hampered by limitations in data availability. Distributed learning methods such as federated learning (FL) and split learning (SL) are introduced to solve this problem as well as data management and ownership issues with their different strengths and weaknesses. The recent proposal of federated split task-agnostic (F eSTA) learning tries to reconcile the distinct merits of FL and SL by enabling the multi-task collaboration between participants through Vision Transformer (ViT) architecture, but they suffer from higher communication overhead. To address this, here we present a multi-task distributed learning using ViT with random patch permutation, dubbed p -F eSTA. Instead of using a CNN-based head as in F eSTA, p -F eSTA adopts a simple patch embedder with random permutation, improving the multi-task learning performance without sacrificing privacy. Experimental results confirm that the proposed method significantly enhances the benefit of multi-task collaboration, communication efficiency, and privacy preservation, shedding light on practical multi-task distributed learning in the field of medical imaging.


Sujet(s)
Intelligence artificielle , Humains
SÉLECTION CITATIONS
DÉTAIL DE RECHERCHE
...