Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 129
Filtrar
1.
J Magn Reson Imaging ; 59(3): 1021-1031, 2024 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-37921361

RESUMO

BACKGROUND: Amyloid-beta and brain atrophy are hallmarks for Alzheimer's Disease that can be targeted with positron emission tomography (PET) and MRI, respectively. MRI is cheaper, less-invasive, and more available than PET. There is a known relationship between amyloid-beta and brain atrophy, meaning PET images could be inferred from MRI. PURPOSE: To build an image translation model using a Conditional Generative Adversarial Network able to synthesize Amyloid-beta PET images from structural MRI. STUDY TYPE: Retrospective. POPULATION: Eight hundred eighty-two adults (348 males/534 females) with different stages of cognitive decline (control, mild cognitive impairment, moderate cognitive impairment, and severe cognitive impairment). Five hundred fifty-two subjects for model training and 331 for testing (80%:20%). FIELD STRENGTH/SEQUENCE: 3 T, T1-weighted structural (T1w). ASSESSMENT: The testing cohort was used to evaluate the performance of the model using the Structural Similarity Index Measure (SSIM) and Peak Signal-to-Noise Ratio (PSNR), comparing the likeness of the overall synthetic PET images created from structural MRI with the overall true PET images. SSIM was computed in the overall image to include the luminance, contrast, and structural similarity components. Experienced observers reviewed the images for quality, performance and tried to determine if they could tell the difference between real and synthetic images. STATISTICAL TESTS: Pixel wise Pearson correlation was significant, and had an R2 greater than 0.96 in example images. From blinded readings, a Pearson Chi-squared test showed that there was no significant difference between the real and synthetic images by the observers (P = 0.68). RESULTS: A high degree of likeness across the evaluation set, which had a mean SSIM = 0.905 and PSNR = 2.685. The two observers were not able to determine the difference between the real and synthetic images, with accuracies of 54% and 46%, respectively. CONCLUSION: Amyloid-beta PET images can be synthesized from structural MRI with a high degree of similarity to the real PET images. EVIDENCE LEVEL: 3 TECHNICAL EFFICACY: Stage 1.


Assuntos
Peptídeos beta-Amiloides , Tomografia por Emissão de Pósitrons , Masculino , Adulto , Feminino , Humanos , Estudos Retrospectivos , Tomografia por Emissão de Pósitrons/métodos , Imageamento por Ressonância Magnética/métodos , Atrofia , Processamento de Imagem Assistida por Computador/métodos
2.
J Microsc ; 295(3): 236-242, 2024 Sep.
Artigo em Inglês | MEDLINE | ID: mdl-38563195

RESUMO

Fibre bundle (FB)-based endoscopes are indispensable in biology and medical science due to their minimally invasive nature. However, resolution and contrast for fluorescence imaging are limited due to characteristic features of the FBs, such as low numerical aperture (NA) and individual fibre core sizes. In this study, we improved the resolution and contrast of sample fluorescence images acquired using in-house fabricated high-NA FBs by utilising generative adversarial networks (GANs). In order to train our deep learning model, we built an FB-based multifocal structured illumination microscope (MSIM) based on a digital micromirror device (DMD) which improves the resolution and the contrast substantially compared to basic FB-based fluorescence microscopes. After network training, the GAN model, employing image-to-image translation techniques, effectively transformed wide-field images into high-resolution MSIM images without the need for any additional optical hardware. The results demonstrated that GAN-generated outputs significantly enhanced both contrast and resolution compared to the original wide-field images. These findings highlight the potential of GAN-based models trained using MSIM data to enhance resolution and contrast in wide-field imaging for fibre bundle-based fluorescence microscopy. Lay Description: Fibre bundle (FB) endoscopes are essential in biology and medicine but suffer from limited resolution and contrast for fluorescence imaging. Here we improved these limitations using high-NA FBs and generative adversarial networks (GANs). We trained a GAN model with data from an FB-based multifocal structured illumination microscope (MSIM) to enhance resolution and contrast without additional optical hardware. Results showed significant enhancement in contrast and resolution, showcasing the potential of GAN-based models for fibre bundle-based fluorescence microscopy.

3.
BMC Med Imaging ; 24(1): 67, 2024 Mar 20.
Artigo em Inglês | MEDLINE | ID: mdl-38504179

RESUMO

BACKGROUND: Clinical data warehouses provide access to massive amounts of medical images, but these images are often heterogeneous. They can for instance include images acquired both with or without the injection of a gadolinium-based contrast agent. Harmonizing such data sets is thus fundamental to guarantee unbiased results, for example when performing differential diagnosis. Furthermore, classical neuroimaging software tools for feature extraction are typically applied only to images without gadolinium. The objective of this work is to evaluate how image translation can be useful to exploit a highly heterogeneous data set containing both contrast-enhanced and non-contrast-enhanced images from a clinical data warehouse. METHODS: We propose and compare different 3D U-Net and conditional GAN models to convert contrast-enhanced T1-weighted (T1ce) into non-contrast-enhanced (T1nce) brain MRI. These models were trained using 230 image pairs and tested on 77 image pairs from the clinical data warehouse of the Greater Paris area. RESULTS: Validation using standard image similarity measures demonstrated that the similarity between real and synthetic T1nce images was higher than between real T1nce and T1ce images for all the models compared. The best performing models were further validated on a segmentation task. We showed that tissue volumes extracted from synthetic T1nce images were closer to those of real T1nce images than volumes extracted from T1ce images. CONCLUSION: We showed that deep learning models initially developed with research quality data could synthesize T1nce from T1ce images of clinical quality and that reliable features could be extracted from the synthetic images, thus demonstrating the ability of such methods to help exploit a data set coming from a clinical data warehouse.


Assuntos
Data Warehousing , Gadolínio , Humanos , Encéfalo/diagnóstico por imagem , Imageamento por Ressonância Magnética/métodos , Neuroimagem/métodos , Processamento de Imagem Assistida por Computador/métodos
4.
Sensors (Basel) ; 24(7)2024 Apr 08.
Artigo em Inglês | MEDLINE | ID: mdl-38610569

RESUMO

The performance of three-dimensional (3D) point cloud reconstruction is affected by dynamic features such as vegetation. Vegetation can be detected by near-infrared (NIR)-based indices; however, the sensors providing multispectral data are resource intensive. To address this issue, this study proposes a two-stage framework to firstly improve the performance of the 3D point cloud generation of buildings with a two-view SfM algorithm, and secondly, reduce noise caused by vegetation. The proposed framework can also overcome the lack of near-infrared data when identifying vegetation areas for reducing interferences in the SfM process. The first stage includes cross-sensor training, model selection and the evaluation of image-to-image RGB to color infrared (CIR) translation with Generative Adversarial Networks (GANs). The second stage includes feature detection with multiple feature detector operators, feature removal with respect to the NDVI-based vegetation classification, masking, matching, pose estimation and triangulation to generate sparse 3D point clouds. The materials utilized in both stages are a publicly available RGB-NIR dataset, and satellite and UAV imagery. The experimental results indicate that the cross-sensor and category-wise validation achieves an accuracy of 0.9466 and 0.9024, with a kappa coefficient of 0.8932 and 0.9110, respectively. The histogram-based evaluation demonstrates that the predicted NIR band is consistent with the original NIR data of the satellite test dataset. Finally, the test on the UAV RGB and artificially generated NIR with a segmentation-driven two-view SfM proves that the proposed framework can effectively translate RGB to CIR for NDVI calculation. Further, the artificially generated NDVI is able to segment and classify vegetation. As a result, the generated point cloud is less noisy, and the 3D model is enhanced.

5.
Sensors (Basel) ; 24(4)2024 Feb 09.
Artigo em Inglês | MEDLINE | ID: mdl-38400307

RESUMO

This work explores the generation of James Webb Space Telescope (JWSP) imagery via image-to-image translation from the available Hubble Space Telescope (HST) data. Comparative analysis encompasses the Pix2Pix, CycleGAN, TURBO, and DDPM-based Palette methodologies, assessing the criticality of image registration in astronomy. While the focus of this study is not on the scientific evaluation of model fairness, we note that the techniques employed may bear some limitations and the translated images could include elements that are not present in actual astronomical phenomena. To mitigate this, uncertainty estimation is integrated into our methodology, enhancing the translation's integrity and assisting astronomers in distinguishing between reliable predictions and those of questionable certainty. The evaluation was performed using metrics including MSE, SSIM, PSNR, LPIPS, and FID. The paper introduces a novel approach to quantifying uncertainty within image translation, leveraging the stochastic nature of DDPMs. This innovation not only bolsters our confidence in the translated images but also provides a valuable tool for future astronomical experiment planning. By offering predictive insights when JWST data are unavailable, our approach allows for informed preparatory strategies for making observations with the upcoming JWST, potentially optimizing its precious observational resources. To the best of our knowledge, this work is the first attempt to apply image-to-image translation for astronomical sensor-to-sensor translation.

6.
Sensors (Basel) ; 24(4)2024 Feb 19.
Artigo em Inglês | MEDLINE | ID: mdl-38400497

RESUMO

Autonomous driving systems heavily depend on perception tasks for optimal performance. However, the prevailing datasets are primarily focused on scenarios with clear visibility (i.e., sunny and daytime). This concentration poses challenges in training deep-learning-based perception models for environments with adverse conditions (e.g., rainy and nighttime). In this paper, we propose an unsupervised network designed for the translation of images from day-to-night to solve the ill-posed problem of learning the mapping between domains with unpaired data. The proposed method involves extracting both semantic and geometric information from input images in the form of attention maps. We assume that the multi-task network can extract semantic and geometric information during the estimation of semantic segmentation and depth maps, respectively. The image-to-image translation network integrates the two distinct types of extracted information, employing them as spatial attention maps. We compare our method with related works both qualitatively and quantitatively. The proposed method shows both qualitative and qualitative improvements in visual presentation over related work.

7.
J Struct Biol ; 215(2): 107965, 2023 06.
Artigo em Inglês | MEDLINE | ID: mdl-37100102

RESUMO

In cryo-transmission electron microscopy (cryo-TEM), sample thickness is one of the most important parameters that governs image quality. When combining cryo-TEM with other imaging methods, such as light microscopy, measuring and controlling the sample thickness to ensure suitability of samples becomes even more critical due to the low throughput of such correlated imaging experiments. Here, we present a method to assess the sample thickness using reflected light microscopy and machine learning that can be used prior to TEM imaging of a sample. The method makes use of the thin-film interference effect that is observed when imaging narrow-band LED light sources reflected by thin samples. By training a neural network to translate such reflection images into maps of the underlying sample thickness, we are able to accurately predict the thickness of cryo-TEM samples using a light microscope. We exemplify our approach using mammalian cells grown on TEM grids, and demonstrate that the thickness predictions are highly similar to the measured sample thickness. The open-source software described herein, including the neural network and algorithms to generate training datasets, is freely available at github.com/bionanopatterning/thicknessprediction. With the recent development of in situ cellular structural biology using cryo-TEM, there is a need for fast and accurate assessment of sample thickness prior to high-resolution imaging. We anticipate that our method will improve the throughput of this assessment by providing an alternative method to screening using cryo-TEM. Furthermore, we demonstrate that our method can be incorporated into correlative imaging workflows to locate intracellular proteins at sites ideal for high-resolution cryo-TEM imaging.


Assuntos
Aprendizado de Máquina , Proteínas , Animais , Microscopia Eletrônica de Transmissão , Microscopia Crioeletrônica/métodos , Software , Mamíferos
8.
Network ; 34(4): 282-305, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-37668425

RESUMO

Neural Style Transfer (NST) has been a widely researched topic as of late enabling new forms of image manipulation. Here we perform an extensive study on NST algorithms and extend the existing methods with custom modifications for application to Indian art styles. In this paper, we aim to provide a comprehensive analysis of various methods ranging from the seminal work of Gatys et al which demonstrated the power of Convolutional Neural Networks (CNNs) in creating artistic imagery by separating and recombining image content and style, to the state of the art image-to-image translation models which use Generative Adversarial Networks (GANs) to learn the mapping between two domain of images. We observe and infer based on the results produced by the models on which one could be a more suitable approach for Indian art styles, especially Tanjore paintings which are unique compared to the Western art styles. We then propose the method which is more suitable for the domain of Indian Art style along with custom architecture which includes an enhancement and evaluation module. We then present evaluation methods, both qualitative and quantitative which includes our proposed metric, to evaluate the results produced by the model.


Assuntos
Algoritmos , Povo Asiático , Cultura , Humanos , Aprendizagem , Índia , Arte
9.
J Appl Clin Med Phys ; 24(10): e14064, 2023 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-37345557

RESUMO

In this work, we demonstrate a method for rapid synthesis of high-quality CT images from unpaired, low-quality CBCT images, permitting CBCT-based adaptive radiotherapy. We adapt contrastive unpaired translation (CUT) to be used with medical images and evaluate the results on an institutional pelvic CT dataset. We compare the method against cycleGAN using mean absolute error, structural similarity index, root mean squared error, and Frèchet Inception Distance and show that CUT significantly outperforms cycleGAN while requiring less time and fewer resources. The investigated method improves the feasibility of online adaptive radiotherapy over the present state-of-the-art.


Assuntos
Tomografia Computadorizada de Feixe Cônico Espiral , Humanos , Tomografia Computadorizada de Feixe Cônico/métodos , Processamento de Imagem Assistida por Computador/métodos , Dosagem Radioterapêutica , Planejamento da Radioterapia Assistida por Computador/métodos
10.
Sensors (Basel) ; 23(3)2023 Jan 17.
Artigo em Inglês | MEDLINE | ID: mdl-36772129

RESUMO

Research exploring CycleGAN-based synthetic image generation has recently accelerated in the medical community due to its ability to leverage unpaired images effectively. However, a commonly established drawback of the CycleGAN, the introduction of artifacts in generated images, makes it unreliable for medical imaging use cases. In an attempt to address this, we explore the effect of structure losses on the CycleGAN and propose a generalized frequency-based loss that aims at preserving the content in the frequency domain. We apply this loss to the use-case of cone-beam computed tomography (CBCT) translation to computed tomography (CT)-like quality. Synthetic CT (sCT) images generated from our methods are compared against baseline CycleGAN along with other existing structure losses proposed in the literature. Our methods (MAE: 85.5, MSE: 20433, NMSE: 0.026, PSNR: 30.02, SSIM: 0.935) quantitatively and qualitatively improve over the baseline CycleGAN (MAE: 88.8, MSE: 24244, NMSE: 0.03, PSNR: 29.37, SSIM: 0.935) across all investigated metrics and are more robust than existing methods. Furthermore, no observable artifacts or loss in image quality were observed. Finally, we demonstrated that sCTs generated using our methods have superior performance compared to the original CBCT images on selected downstream tasks.


Assuntos
Tomografia Computadorizada de Feixe Cônico , Processamento de Imagem Assistida por Computador , Processamento de Imagem Assistida por Computador/métodos , Tomografia Computadorizada por Raios X , Artefatos , Benchmarking
11.
Sensors (Basel) ; 23(15)2023 Aug 01.
Artigo em Inglês | MEDLINE | ID: mdl-37571641

RESUMO

Unsupervised image-to-image translation has received considerable attention due to the recent remarkable advancements in generative adversarial networks (GANs). In image-to-image translation, state-of-the-art methods use unpaired image data to learn mappings between the source and target domains. However, despite their promising results, existing approaches often fail in challenging conditions, particularly when images have various target instances and a translation task involves significant transitions in shape and visual artifacts when translating low-level information rather than high-level semantics. To tackle the problem, we propose a novel framework called Progressive Unsupervised Generative Attentional Networks with Adaptive Layer-Instance Normalization (PRO-U-GAT-IT) for the unsupervised image-to-image translation task. In contrast to existing attention-based models that fail to handle geometric transitions between the source and target domains, our model can translate images requiring extensive and holistic changes in shape. Experimental results show the superiority of the proposed approach compared to the existing state-of-the-art models on different datasets.

12.
Sensors (Basel) ; 23(7)2023 Apr 06.
Artigo em Inglês | MEDLINE | ID: mdl-37050825

RESUMO

The range-gated laser imaging instrument can capture face images in a dark environment, which provides a new idea for long-distance face recognition at night. However, the laser image has low contrast, low SNR and no color information, which affects observation and recognition. Therefore, it becomes important to convert laser images into visible images and then identify them. For image translation, we propose a laser-visible face image translation model combined with spectral normalization (SN-CycleGAN). We add spectral normalization layers to the discriminator to solve the problem of low image translation quality caused by the difficulty of training the generative adversarial network. The content reconstruction loss function based on the Y channel is added to reduce the error mapping. The face generated by the improved model on the self-built laser-visible face image dataset has better visual quality, which reduces the error mapping and basically retains the structural features of the target compared with other models. The FID value of evaluation index is 36.845, which is 16.902, 13.781, 10.056, 57.722, 62.598 and 0.761 lower than the CycleGAN, Pix2Pix, UNIT, UGATIT, StarGAN and DCLGAN models, respectively. For the face recognition of translated images, we propose a laser-visible face recognition model based on feature retention. The shallow feature maps with identity information are directly connected to the decoder to solve the problem of identity information loss in network transmission. The domain loss function based on triplet loss is added to constrain the style between domains. We use pre-trained FaceNet to recognize generated visible face images and obtain the recognition accuracy of Rank-1. The recognition accuracy of the images generated by the improved model reaches 76.9%, which is greatly improved compared with the above models and 19.2% higher than that of laser face recognition.

13.
Sensors (Basel) ; 23(20)2023 Oct 12.
Artigo em Inglês | MEDLINE | ID: mdl-37896492

RESUMO

In the field of intelligent vehicle technology, there is a high dependence on images captured under challenging conditions to develop robust perception algorithms. However, acquiring these images can be both time-consuming and dangerous. To address this issue, unpaired image-to-image translation models offer a solution by synthesizing samples of the desired domain, thus eliminating the reliance on ground truth supervision. However, the current methods predominantly focus on single projections rather than multiple solutions, not to mention controlling the direction of generation, which creates a scope for enhancement. In this study, we propose a generative adversarial network (GAN)-based model, which incorporates both a style encoder and a content encoder, specifically designed to extract relevant information from an image. Further, we employ a decoder to reconstruct an image using these encoded features, while ensuring that the generated output remains within a permissible range by applying a self-regression module to constrain the style latent space. By modifying the hyperparameters, we can generate controllable outputs with specific style codes. We evaluate the performance of our model by generating snow scenes on the Cityscapes and the EuroCity Persons datasets. The results reveal the effectiveness of our proposed methodology, thereby reinforcing the benefits of our approach in the ongoing evolution of intelligent vehicle technology.

14.
J Cell Sci ; 133(7)2020 04 14.
Artigo em Inglês | MEDLINE | ID: mdl-32094267

RESUMO

Measuring the physical size of a cell is valuable in understanding cell growth control. Current single-cell volume measurement methods for mammalian cells are labor intensive, inflexible and can cause cell damage. We introduce CTRL: Cell Topography Reconstruction Learner, a label-free technique incorporating the deep learning algorithm and the fluorescence exclusion method for reconstructing cell topography and estimating mammalian cell volume from differential interference contrast (DIC) microscopy images alone. The method achieves quantitative accuracy, requires minimal sample preparation, and applies to a wide range of biological and experimental conditions. The method can be used to track single-cell volume dynamics over arbitrarily long time periods. For HT1080 fibrosarcoma cells, we observe that the cell size at division is positively correlated with the cell size at birth (sizer), and there is a noticeable reduction in cell size fluctuations at 25% completion of the cell cycle in HT1080 fibrosarcoma cells.


Assuntos
Algoritmos , Inteligência Artificial , Animais , Divisão Celular , Tamanho Celular
15.
BMC Med Imaging ; 22(1): 124, 2022 07 14.
Artigo em Inglês | MEDLINE | ID: mdl-35836126

RESUMO

BACKGROUND: Current medical image translation is implemented in the image domain. Considering the medical image acquisition is essentially a temporally continuous process, we attempt to develop a novel image translation framework via deep learning trained in video domain for generating synthesized computed tomography (CT) images from cone-beam computed tomography (CBCT) images. METHODS: For a proof-of-concept demonstration, CBCT and CT images from 100 patients were collected to demonstrate the feasibility and reliability of the proposed framework. The CBCT and CT images were further registered as paired samples and used as the input data for the supervised model training. A vid2vid framework based on the conditional GAN network, with carefully-designed generators, discriminators and a new spatio-temporal learning objective, was applied to realize the CBCT-CT image translation in the video domain. Four evaluation metrics, including mean absolute error (MAE), peak signal-to-noise ratio (PSNR), normalized cross-correlation (NCC), and structural similarity (SSIM), were calculated on all the real and synthetic CT images from 10 new testing patients to illustrate the model performance. RESULTS: The average values for four evaluation metrics, including MAE, PSNR, NCC, and SSIM, are 23.27 ± 5.53, 32.67 ± 1.98, 0.99 ± 0.0059, and 0.97 ± 0.028, respectively. Most of the pixel-wise hounsfield units value differences between real and synthetic CT images are within 50. The synthetic CT images have great agreement with the real CT images and the image quality is improved with lower noise and artifacts compared with CBCT images. CONCLUSIONS: We developed a deep-learning-based approach to perform the medical image translation problem in the video domain. Although the feasibility and reliability of the proposed framework were demonstrated by CBCT-CT image translation, it can be easily extended to other types of medical images. The current results illustrate that it is a very promising method that may pave a new path for medical image translation research.


Assuntos
Aprendizado Profundo , Tomografia Computadorizada de Feixe Cônico/métodos , Humanos , Processamento de Imagem Assistida por Computador/métodos , Reprodutibilidade dos Testes , Razão Sinal-Ruído
16.
Sensors (Basel) ; 22(21)2022 Oct 29.
Artigo em Inglês | MEDLINE | ID: mdl-36366007

RESUMO

In this paper, we revisit the paired image-to-image translation using the conditional generative adversarial network, the so-called "Pix2Pix", and propose efficient optimization techniques for the architecture and the training method to maximize the architecture's performance to boost the realism of the generated images. We propose a generative adversarial network-based technique to create new artificial indoor scenes using a user-defined semantic segmentation map as an input to define the location, shape, and category of each object in the scene, exactly similar to Pix2Pix. We train different residual connections-based architectures of the generator and discriminator on the NYU depth-v2 dataset and a selected indoor subset from the ADE20K dataset, showing that the proposed models have fewer parameters, less computational complexity, and can generate better quality images than the state of the art methods following the same technique to generate realistic indoor images. We also prove that using extra specific labels and more training samples increases the quality of the generated images; however, the proposed residual connections-based models can learn better from small datasets (i.e., NYU depth-v2) and can improve the realism of the generated images in training on bigger datasets (i.e., ADE20K indoor subset) in comparison to Pix2Pix. The proposed method achieves an LPIPS value of 0.505 and an FID value of 81.067, generating better quality images than that produced by Pix2Pix and other recent paired Image-to-image translation methods and outperforming them in terms of LPIPS and FID.

17.
Sensors (Basel) ; 22(21)2022 Nov 06.
Artigo em Inglês | MEDLINE | ID: mdl-36366238

RESUMO

Supervised image-to-image translation has been proven to generate realistic images with sharp details and to have good quantitative performance. Such methods are trained on a paired dataset, where an image from the source domain already has a corresponding translated image in the target domain. However, this paired dataset requirement imposes a huge practical constraint, requires domain knowledge or is even impossible to obtain in certain cases. Due to these problems, unsupervised image-to-image translation has been proposed, which does not require domain expertise and can take advantage of a large unlabeled dataset. Although such models perform well, they are hard to train due to the major constraints induced in their loss functions, which make training unstable. Since CycleGAN has been released, numerous methods have been proposed which try to address various problems from different perspectives. In this review, we firstly describe the general image-to-image translation framework and discuss the datasets and metrics involved in the topic. Furthermore, we revise the current state-of-the-art with a classification of existing works. This part is followed by a small quantitative evaluation, for which results were taken from papers.


Assuntos
Processamento de Imagem Assistida por Computador , Processamento de Imagem Assistida por Computador/métodos
18.
Sensors (Basel) ; 23(1)2022 Dec 23.
Artigo em Inglês | MEDLINE | ID: mdl-36616754

RESUMO

Generative adversarial network (GAN)-based data augmentation is used to enhance the performance of object detection models. It comprises two stages: training the GAN generator to learn the distribution of a small target dataset, and sampling data from the trained generator to enhance model performance. In this paper, we propose a pipelined model, called robust data augmentation GAN (RDAGAN), that aims to augment small datasets used for object detection. First, clean images and a small datasets containing images from various domains are input into the RDAGAN, which then generates images that are similar to those in the input dataset. Thereafter, it divides the image generation task into two networks: an object generation network and image translation network. The object generation network generates images of the objects located within the bounding boxes of the input dataset and the image translation network merges these images with clean images. A quantitative experiment confirmed that the generated images improve the YOLOv5 model's fire detection performance. A comparative evaluation showed that RDAGAN can maintain the background information of input images and localize the object generation location. Moreover, ablation studies demonstrated that all components and objects included in the RDAGAN play pivotal roles.


Assuntos
Incêndios , Aprendizagem , Processamento de Imagem Assistida por Computador
19.
Neuroimage ; 243: 118569, 2021 11.
Artigo em Inglês | MEDLINE | ID: mdl-34506916

RESUMO

In magnetic resonance (MR) imaging, a lack of standardization in acquisition often causes pulse sequence-based contrast variations in MR images from site to site, which impedes consistent measurements in automatic analyses. In this paper, we propose an unsupervised MR image harmonization approach, CALAMITI (Contrast Anatomy Learning and Analysis for MR Intensity Translation and Integration), which aims to alleviate contrast variations in multi-site MR imaging. Designed using information bottleneck theory, CALAMITI learns a globally disentangled latent space containing both anatomical and contrast information, which permits harmonization. In contrast to supervised harmonization methods, our approach does not need a sample population to be imaged across sites. Unlike traditional unsupervised harmonization approaches which often suffer from geometry shifts, CALAMITI better preserves anatomy by design. The proposed method is also able to adapt to a new testing site with a straightforward fine-tuning process. Experiments on MR images acquired from ten sites show that CALAMITI achieves superior performance compared with other harmonization approaches.


Assuntos
Imageamento por Ressonância Magnética/métodos , Humanos , Processamento de Imagem Assistida por Computador/métodos , Teoria da Informação
20.
J Appl Clin Med Phys ; 22(3): 55-62, 2021 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-33527712

RESUMO

PURPOSE AND BACKGROUND: The magnetic resonance (MR)-only radiotherapy workflow is urged by the increasing use of MR image for the identification and delineation of tumors, while a fast generation of synthetic computer tomography (sCT) image from MR image for dose calculation remains one of the key challenges to the workflow. This study aimed to develop a neural network to generate the sCT in brain site and evaluate the dosimetry accuracy. MATERIALS AND METHODS: A generative adversarial network (GAN) was developed to translate T1-weighted MRI to sCT. First, the "U-net" shaped encoder-decoder network with some image translation-specific modifications was trained to generate sCT, then the discriminator network was adversarially trained to distinguish between synthetic and real CT images. We enrolled 37 brain cancer patients acquiring both CT and MRI for treatment position simulation. Twenty-seven pairs of 2D T1-weighted MR images and rigidly registered CT image were used to train the GAN model, and the remaining 10 pairs were used to evaluate the model performance through the metric of mean absolute error. Furthermore, the clinical Volume Modulated Arc Therapy plan was calculated on both sCT and real CT, followed by gamma analysis and comparison of dose-volume histogram. RESULTS: On average, only 15 s were needed to generate one sCT from one T1-weighted MRI. The mean absolute error between synthetic and real CT was 60.52 ± 13.32 Housefield Unit over 5-fold cross validation. For dose distribution on sCT and CT, the average pass rates of gamma analysis using the 3%/3 mm and 2%/2 mm criteria were 99.76% and 97.25% over testing patients, respectively. For parameters of dose-volume histogram for both target and organs at risk, no significant differences were found between both plans. CONCLUSION: The GAN model can generate synthetic CT from one single MRI sequence within seconds, and a state-of-art accuracy of CT number and dosimetry was achieved.


Assuntos
Imageamento por Ressonância Magnética , Tomografia Computadorizada por Raios X , Encéfalo/diagnóstico por imagem , Humanos , Espectroscopia de Ressonância Magnética , Redes Neurais de Computação , Dosagem Radioterapêutica , Planejamento da Radioterapia Assistida por Computador
SELEÇÃO DE REFERÊNCIAS
Detalhe da pesquisa