RESUMO
BACKGROUND: In clinical medicine, low-dose radiographic image noise reduces the quality of the detected image features and may have a negative impact on disease diagnosis. OBJECTIVE: In this study, Adaptive Projection Network (APNet) is proposed to reduce noise from low-dose medical images. METHODS: APNet is developed based on an architecture of the U-shaped network to capture multi-scale data and achieve end-to-end image denoising. To adaptively calibrate important features during information transmission, a residual block of the dual attention method throughout the encoding and decoding phases is integrated. A non-local attention module to separate the noise and texture of the image details by using image adaptive projection during the feature fusion. RESULTS: To verify the effectiveness of APNet, experiments on lung CT images with synthetic noise are performed, and the results demonstrate that the proposed approach outperforms recent methods in both quantitative index and visual quality. In addition, the denoising experiment on the dental CT image is also carried out and it verifies that the network has a certain generalization. CONCLUSIONS: The proposed APNet is an effective method that can reduce image noise and preserve the required image details in low-dose radiographic images.
Assuntos
Algoritmos , Tomografia Computadorizada por Raios X , Razão Sinal-Ruído , Doses de Radiação , Tomografia Computadorizada por Raios X/métodos , Processamento de Imagem Assistida por Computador/métodosRESUMO
Recently, convolutional neural networks (CNNs) have been widely used in image denoising, and their performance has been enhanced through residual learning. However, previous research mostly focused on optimizing the network architecture of CNNs, ignoring the limitations of the commonly used residual learning. This paper identifies two of its limitations, which are the neglect of image information and the lack of effective consideration of image self-similarity. To solve these limitations, this paper proposes a compositional denoising network (CDN), which contains two sub-paths, the image information path (IIP) and the noise estimation path (NEP), respectively. IIP is trained via an image-to-image method to extract image information. For NEP, it utilizes image self-similarity from the perspective of training. This similarity-based training method constrains NEP to output similar estimated noise distributions for different image patches with a specific kind of noise. Finally, image information and noise distribution information are comprehensively considered for image denoising. Experimental results indicate that CDN outperforms other CNN-based methods in both synthetic and real-world image denoising, achieving state-of-the-art performance.
Assuntos
Processamento de Imagem Assistida por Computador , Placa Aterosclerótica , Humanos , Processamento de Imagem Assistida por Computador/métodos , Razão Sinal-Ruído , Redes Neurais de ComputaçãoRESUMO
BACKGROUND: Low dose computed tomography (LDCT) uses lower radiation dose, but the reconstructed images contain higher noise that can have negative impact in disease diagnosis. Although deep learning with the edge extraction operators reserves edge information well, only applying the edge extraction operators to input LDCT images does not yield overall satisfactory results. OBJECTIVE: To improve LDCT images quality, this study proposes and tests a dual edge extraction multi-scale attention mechanism convolution neural network (DEMACNN) based on a compound loss. METHODS: The network uses edge extraction operators to extract edge information from both the input images and the feature maps in the network, improving the utilization of the edge operators and retaining the images edge information. The feature enhancement block is constructed by fusing the attention mechanism and multi-scale module, enhancing effective information, while suppressing useless information. The residual learning method is used to learn the network, improving the performance of the network, and solving the problem of gradient disappearance. Except for the network structure, a compound loss function, which consists of the MSE loss, the proposed joint total variation loss, and the edge loss, is proposed to enhance the denoising ability of the network and reserve the edge of images. RESULTS: Compared with other advanced methods (REDCNN, CT-former and EDCNN), the proposed new network achieves the best PSNR and SSIM values in LDCT images of the abdomen, which are 33.3486 and 0.9104, respectively. In addition, the new network also performs well on head and chest image data. CONCLUSION: The experimental results demonstrate that the proposed new network structure and denoising algorithm not only effectively removes the noise in LDCT images, but also protects the edges and details of the images well.
Assuntos
Algoritmos , Redes Neurais de Computação , Tórax , Tomografia Computadorizada por Raios XRESUMO
Rupture of intracranial aneurysm is the first cause of subarachnoid hemorrhage, second only to cerebral thrombosis and hypertensive cerebral hemorrhage, and the mortality rate is very high. MRI technology plays an irreplaceable role in the early detection and diagnosis of intracranial aneurysms and supports evaluating the size and structure of aneurysms. The increase in many aneurysm images, may be a massive workload for the doctors, which is likely to produce a wrong diagnosis. Therefore, we proposed a simple and effective comprehensive residual attention network (CRANet) to improve the accuracy of aneurysm detection, using a residual network to extract the features of an aneurysm. Many experiments have shown that the proposed CRANet model could detect aneurysms effectively. In addition, on the test set, the accuracy and recall rates reached 97.81% and 94%, which significantly improved the detection rate of aneurysms.
Assuntos
Aneurisma Intracraniano , Hemorragia Subaracnóidea , Humanos , Aneurisma Intracraniano/complicações , Aneurisma Intracraniano/diagnóstico por imagem , Hemorragia Subaracnóidea/etiologiaRESUMO
Diffusion tensor magnetic resonance imaging (DTI) is a widely adopted neuroimaging method for the in vivo mapping of brain tissue microstructure and white matter tracts. Nonetheless, the noise in the diffusion-weighted images (DWIs) decreases the accuracy and precision of DTI derived microstructural parameters and leads to prolonged acquisition time for achieving improved signal-to-noise ratio (SNR). Deep learning-based image denoising using convolutional neural networks (CNNs) has superior performance but often requires additional high-SNR data for supervising the training of CNNs, which reduces the feasibility of supervised learning-based denoising in practice. In this work, we develop a self-supervised deep learning-based method entitled "SDnDTI" for denoising DTI data, which does not require additional high-SNR data for training. Specifically, SDnDTI divides multi-directional DTI data into many subsets of six DWI volumes and transforms DWIs from each subset to along the same diffusion-encoding directions through the diffusion tensor model, generating multiple repetitions of DWIs with identical image contrasts but different noise observations. SDnDTI removes noise by first denoising each repetition of DWIs using a deep 3-dimensional CNN with the average of all repetitions with higher SNR as the training target, following the same approach as normal supervised learning based denoising methods, and then averaging CNN-denoised images for achieving higher SNR. The denoising efficacy of SDnDTI is demonstrated in terms of the similarity of output images and resultant DTI metrics compared to the ground truth generated using substantially more DWI volumes on two datasets with different spatial resolutions, b-values and numbers of input DWI volumes provided by the Human Connectome Project (HCP) and the Lifespan HCP in Aging. The SDnDTI results preserve image sharpness and textural details and substantially improve upon those from the raw data. The results of SDnDTI are comparable to those from supervised learning-based denoising and outperform those from state-of-the-art conventional denoising algorithms including BM4D, AONLM and MPPCA. By leveraging domain knowledge of diffusion MRI physics, SDnDTI makes it easier to use CNN-based denoising methods in practice and has the potential to benefit a wider range of research and clinical applications that require accelerated DTI acquisition and high-quality DTI data for mapping of tissue microstructure, fiber tracts and structural connectivity in the living human brain.
Assuntos
Aprendizado Profundo , Imagem de Tensor de Difusão , Imagem de Tensor de Difusão/métodos , Humanos , Processamento de Imagem Assistida por Computador/métodos , Redes Neurais de Computação , Razão Sinal-RuídoRESUMO
Protein fold recognition is a critical step in protein structure and function prediction, and aims to ascertain the most likely fold type of the query protein. As a typical pattern recognition problem, designing a powerful feature extractor and metric function to extract relevant and representative fold-specific features from protein sequences is the key to improving protein fold recognition. In this study, we propose an effective sequence-based approach, called RattnetFold, to identify protein fold types. The basic concept of RattnetFold is to employ a stack convolutional neural network with the attention mechanism that acts as a feature extractor to extract fold-specific features from protein residue-residue contact maps. Moreover, based on the fold-specific features, we leverage metric learning to project fold-specific features into a subspace where similar proteins are closer together and name this approach RattnetFoldPro. Benchmarking experiments illustrate that RattnetFold and RattnetFoldPro enable the convolutional neural networks to efficiently learn the underlying subtle patterns in residue-residue contact maps, thereby improving the performance of protein fold recognition. An online web server of RattnetFold and the benchmark datasets are freely available at http://csbio.njust.edu.cn/bioinf/rattnetfold/.
Assuntos
Biologia Computacional , Redes Neurais de Computação , Algoritmos , Sequência de Aminoácidos , Proteínas/químicaRESUMO
Despite the extensive efforts, accurate traffic time series forecasting remains challenging. By taking into account the non-linear nature of traffic in-depth, we propose a novel ST-CRMF model consisting of the Compensated Residual Matrix Factorization with Spatial-Temporal regularization for graph-based traffic time series forecasting. Our model inherits the benefits of MF and regularizer optimization and further carries out the compensatory modeling of the spatial-temporal correlations through a well-designed bi-directional residual structure. Of particular concern is that MF modeling and later residual learning share and synchronize iterative updates as equal training parameters, which considerably alleviates the error propagation problem that associates with rolling forecasting. Besides, most of the existing prediction models have neglected the difficult-to-avoid issue of missing traffic data; the ST-CRMF model can repair the possible missing value while fulfilling the forecasting tasks. After testing the effects of key parameters on model performance, the numerous experimental results confirm that our ST-CRMF model can efficiently capture the comprehensive spatial-temporal dependencies and significantly outperform those state-of-the-art models in the short-to-long terms (5-/15-/30-/60-min) traffic forecasting tasks on the open Seattle-Loop and METR-LA traffic datasets.
Assuntos
Fatores de Tempo , PrevisõesRESUMO
Skin cancer is the most common type of cancer that affects humans and is usually diagnosed by initial clinical screening, which is followed by dermoscopic analysis. Automated classification of skin lesions is still a challenging task because of the high visual similarity between melanoma and benign lesions. This paper proposes a new residual deep convolutional neural network (RDCNN) for skin lesions diagnosis. The proposed neural network is trained and tested using six well-known skin cancer datasets, PH2, DermIS and Quest, MED-NODE, ISIC2016, ISIC2017, and ISIC2018. Three different experiments are carried out to measure the performance of the proposed RDCNN. In the first experiment, the proposed RDCNN is trained and tested using the original dataset images without any pre-processing or segmentation. In the second experiment, the proposed RDCNN is tested using segmented images. Finally, the utilized trained model in the second experiment is saved and reused in the third experiment as a pre-trained model. Then, it is trained again using a different dataset. The proposed RDCNN shows significant high performance and outperforms the existing deep convolutional networks.
Assuntos
Melanoma , Dermatopatias , Neoplasias Cutâneas , Dermoscopia , Progressão da Doença , Humanos , Melanoma/diagnóstico por imagem , Redes Neurais de Computação , Dermatopatias/diagnóstico por imagem , Neoplasias Cutâneas/diagnóstico por imagemRESUMO
BACKGROUND: Knee Osteoarthritis (KOA) is the most common type of Osteoarthritis (OA) and it is diagnosed by physicians using a standard 0 -4 Kellgren Lawrence (KL) grading system which sets the KOA on a spectrum of 5 grades; starting from normal (0) to Severe OA (4). OBJECTIVES: In this paper, we propose a transfer learning approach of a very deep wide residual learning-based network (WRN-50-2) which is fine-tuned using X-ray plain radiographs from the Osteoarthritis Initiative (OAI) dataset to learn the KL severity grading of KOA. METHODS: We propose a data augmentation approach of OAI data to avoid data imbalance and reduce overfitting by applying it only to certain KL grades depending on their number of plain radiographs. Then we conduct experiments to test the model based on an independent testing data of original plain radiographs acquired from the OAI dataset. RESULTS: Experimental results showed good generalization power in predicting the KL grade of knee X-rays with an accuracy of 72% and Precision 74%. Moreover, using Grad-Cam, we also observed that network selected some distinctive features that describe the prediction of a KL grade of a knee radiograph. CONCLUSION: This study demonstrates that our proposed new model outperforms several other related works, and it can be further improved to be used to help radiologists make more accurate and precise diagnosis of KOA in future clinical practice.
Assuntos
Osteoartrite do Joelho , Progressão da Doença , Humanos , Osteoartrite do Joelho/diagnóstico por imagem , Radiografia , Raios XRESUMO
Channel estimation is a challenging task in a millimeter-wave (mm Wave) massive multiple-input multiple-output (MIMO) system. The existing deep learning scheme, which learns the mapping from the input to the target channel, has great difficulty in estimating the exact channel state information (CSI). In this paper, we consider the quantized received measurements as a low-resolution image, and we adopt the deep learning-based image super-resolution technique to reconstruct the mm Wave channel. Specifically, we exploit a state-of-the-art channel estimation framework based on residual learning and multi-path feature fusion (RL-MFF-Net). Firstly, residual learning makes the channel estimator focus on learning high-frequency residual information between the quantized received measurements and the mm Wave channel, while abundant low-frequency information is bypassed through skip connections. Moreover, to address the estimator's gradient dispersion problem, a dense connection is added to the residual blocks to ensure the maximum information flow between the layers. Furthermore, the underlying mm Wave channel local features extracted from different residual blocks are preserved by multi-path feature fusion. The simulation results demonstrate that the proposed scheme outperforms traditional methods as well as existing deep learning methods, especially in the low signal-to-noise-ration (SNR) region.
RESUMO
BACKGROUND: Accurate segmentation and recognition algorithm of lung nodules has great important value of reference for early diagnosis of lung cancer. An algorithm is proposed for 3D CT sequence images in this paper based on 3D Res U-Net segmentation network and 3D ResNet50 classification network. The common convolutional layers in encoding and decoding paths of U-Net are replaced by residual units while the loss function is changed to Dice loss after using cross entropy loss to accelerate network convergence. Since the lung nodules are small and rich in 3D information, the ResNet50 is improved by replacing the 2D convolutional layers with 3D convolutional layers and reducing the sizes of some convolution kernels, 3D ResNet50 network is obtained for the diagnosis of benign and malignant lung nodules. RESULTS: 3D Res U-Net was trained and tested on 1044 CT subcases in the LIDC-IDRI database. The segmentation result shows that the Dice coefficient of 3D Res U-Net is above 0.8 for the segmentation of lung nodules larger than 10 mm in diameter. 3D ResNet50 was trained and tested on 2960 lung nodules in the LIDC-IDRI database. The classification result shows that the diagnostic accuracy of 3D ResNet50 is 87.3% and AUC is 0.907. CONCLUSION: The 3D Res U-Net module improves segmentation performance significantly with the comparison of 3D U-Net model based on residual learning mechanism. 3D Res U-Net can identify small nodules more effectively and improve its segmentation accuracy for large nodules. Compared with the original network, the classification performance of 3D ResNet50 is significantly improved, especially for small benign nodules.
Assuntos
Aprendizado Profundo , Neoplasias Pulmonares , Algoritmos , Humanos , Processamento de Imagem Assistida por Computador , Pulmão , Neoplasias Pulmonares/diagnóstico por imagem , Redes Neurais de Computação , Tomografia Computadorizada por Raios XRESUMO
BACKGROUND: A novel multi-level pyramidal pooling residual U-Net with adversarial mechanism was proposed for organ segmentation from medical imaging, and was conducted on the challenging NIH Pancreas-CT dataset. METHODS: The 82 pancreatic contrast-enhanced abdominal CT volumes were split via four-fold cross validation to test the model performance. In order to achieve accurate segmentation, we firstly involved residual learning into an adversarial U-Net to achieve a better gradient information flow for improving segmentation performance. Then, we introduced a multi-level pyramidal pooling module (MLPP), where a novel pyramidal pooling was involved to gather contextual information for segmentation, then four groups of structures consisted of a different number of pyramidal pooling blocks were proposed to search for the structure with the optimal performance, and two types of pooling blocks were applied in the experimental section to further assess the robustness of MLPP for pancreas segmentation. For evaluation, Dice similarity coefficient (DSC) and recall were used as the metrics in this work. RESULTS: The proposed method preceded the baseline network 5.30% and 6.16% on metrics DSC and recall, and achieved competitive results compared with the-state-of-art methods. CONCLUSIONS: Our algorithm showed great segmentation performance even on the particularly challenging pancreas dataset, this indicates that the proposed model is a satisfactory and promising segmentor.
Assuntos
Redes Neurais de Computação , Pâncreas/anatomia & histologia , Pâncreas/diagnóstico por imagem , Tomografia Computadorizada por Raios X , Meios de Contraste , Conjuntos de Dados como Assunto , HumanosRESUMO
Convolutional neural networks (CNNs) are a state-of-the-art technique for speech emotion recognition. However, CNNs have mostly been applied to noise-free emotional speech data, and limited evidence is available for their applicability in emotional speech denoising. In this study, a cascaded denoising CNN (DnCNN)-CNN architecture is proposed to classify emotions from Korean and German speech in noisy conditions. The proposed architecture consists of two stages. In the first stage, the DnCNN exploits the concept of residual learning to perform denoising; in the second stage, the CNN performs the classification. The classification results for real datasets show that the DnCNN-CNN outperforms the baseline CNN in overall accuracy for both languages. For Korean speech, the DnCNN-CNN achieves an accuracy of 95.8%, whereas the accuracy of the CNN is marginally lower (93.6%). For German speech, the DnCNN-CNN has an overall accuracy of 59.3-76.6%, whereas the CNN has an overall accuracy of 39.4-58.1%. These results demonstrate the feasibility of applying the DnCNN with residual learning to speech denoising and the effectiveness of the CNN-based approach in speech emotion recognition. Our findings provide new insights into speech emotion recognition in adverse conditions and have implications for language-universal speech emotion recognition.
Assuntos
Redes Neurais de Computação , Fala , Emoções , Idioma , PercepçãoRESUMO
Automatically recognizing the modulation of radar signals is a necessary survival technique in electronic intelligence systems. In order to avoid the complex process of the feature extracting and realize the intelligent modulation recognition of various radar signals under low signal-to-noise ratios (SNRs), this paper proposes a method based on intrapulse signatures of radar signals using adaptive singular value reconstruction (ASVR) and deep residual learning. Firstly, the time-frequency spectrums of radar signals under low SNRs are improved after ASVR denoising processing. Secondly, a series of image processing techniques, including binarizing and morphologic filtering, are applied to suppress the background noise in the time-frequency distribution images (TFDIs). Thirdly, the training process of the residual network is achieved using TFDIs, and classification under various conditions is realized using the new-trained network. Simulation results show that, for eight kinds of modulation signals, the proposed approach still achieves an overall probability of successful recognition of 94.1% when the SNR is only -8 dB. Outstanding performance proves the superiority and robustness of the proposed method.
RESUMO
Image denoising is a challenging task that is essential in numerous computer vision and image processing problems. This study proposes and applies a generative adversarial network-based image denoising training architecture to multiple-level Gaussian image denoising tasks. Convolutional neural network-based denoising approaches come across a blurriness issue that produces denoised images blurry on texture details. To resolve the blurriness issue, we first performed a theoretical study of the cause of the problem. Subsequently, we proposed an adversarial Gaussian denoiser network, which uses the generative adversarial network-based adversarial learning process for image denoising tasks. This framework resolves the blurriness problem by encouraging the denoiser network to find the distribution of sharp noise-free images instead of blurry images. Experimental results demonstrate that the proposed framework can effectively resolve the blurriness problem and achieve significant denoising efficiency than the state-of-the-art denoising methods.
RESUMO
Surface flatness assessment is necessary for quality control of metal sheets manufactured from steel coils by roll leveling and cutting. Mechanical-contact-based flatness sensors are being replaced by modern laser-based optical sensors that deliver accurate and dense reconstruction of metal sheet surfaces for flatness index computation. However, the surface range images captured by these optical sensors are corrupted by very specific kinds of noise due to vibrations caused by mechanical processes like degreasing, cleaning, polishing, shearing, and transporting roll systems. Therefore, high-quality flatness optical measurement systems strongly depend on the quality of image denoising methods applied to extract the true surface height image. This paper presents a deep learning architecture for removing these specific kinds of noise from the range images obtained by a laser based range sensor installed in a rolling and shearing line, in order to allow accurate flatness measurements from the clean range images. The proposed convolutional blind residual denoising network (CBRDNet) is composed of a noise estimation module and a noise removal module implemented by specific adaptation of semantic convolutional neural networks. The CBRDNet is validated on both synthetic and real noisy range image data that exhibit the most critical kinds of noise that arise throughout the metal sheet production process. Real data were obtained from a single laser line triangulation flatness sensor installed in a roll leveling and cut to length line. Computational experiments over both synthetic and real datasets clearly demonstrate that CBRDNet achieves superior performance in comparison to traditional 1D and 2D filtering methods, and state-of-the-art CNN-based denoising techniques. The experimental validation results show a reduction in error than can be up to 15% relative to solutions based on traditional 1D and 2D filtering methods and between 10% and 3% relative to the other deep learning denoising architectures recently reported in the literature.
RESUMO
Due to the fast transmission speed and severe health damage, COVID-19 has attracted global attention. Early diagnosis and isolation are effective and imperative strategies for epidemic prevention and control. Most diagnostic methods for the COVID-19 is based on nucleic acid testing (NAT), which is expensive and time-consuming. To build an efficient and valid alternative of NAT, this article investigates the feasibility of employing computed tomography images of lungs as the diagnostic signals. Unlike normal lungs, parts of the lungs infected with the COVID-19 developed lesions, ground-glass opacity, and bronchiectasis became apparent. Through a public dataset, in this article, we propose an advanced residual learning diagnosis detection (RLDD) scheme for the COVID-19 technique, which is designed to distinguish positive COVID-19 cases from heterogeneous lung images. Besides the advantage of high diagnosis effectiveness, the designed residual-based COVID-19 detection network can efficiently extract the lung features through small COVID-19 samples, which removes the pretraining requirement on other medical datasets. In the test set, we achieve an accuracy of 91.33%, a precision of 91.30%, and a recall of 90%. For the batch of 150 samples, the assessment time is only 4.7 s. Therefore, RLDD can be integrated into the application programming interface and embedded into the medical instrument to improve the detection efficiency of COVID-19.
RESUMO
In this study, we developed a multi-scale Convolutional neural network based Automated hippocampal subfield Segmentation Toolbox (CAST) for automated segmentation of hippocampal subfields. Although training CAST required approximately three days on a single workstation with a high-quality GPU card, CAST can segment a new subject in less than 1 âmin even with GPU acceleration disabled, thus this method is more time efficient than current automated methods and manual segmentation. This toolbox is highly flexible with either a single modality or multiple modalities and can be easily set up to be trained with a researcher's unique data. A 3D multi-scale deep convolutional neural network is the key algorithm used in the toolbox. The main merit of multi-scale images is the capability to capture more global structural information from down-sampled images without dramatically increasing memory and computational burden. The original images capture more local information to refine the boundary between subfields. Residual learning is applied to alleviate the vanishing gradient problem and improve the performance with a deeper network. We applied CAST with the same settings on two datasets, one 7T dataset (the UMC dataset) with only the T2 image and one 3T dataset (the MNI dataset) with both T1 and T2 images available. The segmentation accuracy of both CAST and the state-of-the-art automated method ASHS, in terms of the dice similarity coefficient (DSC), were comparable. CAST significantly improved the reliability of segmenting small subfields, such as CA2, CA3, and the entorhinal cortex (ERC), in terms of the intraclass correlation coefficient (ICC). Both ASHS and manual segmentation process some subfields (e.g. CA2 and ERC) with high DSC values but low ICC values, consequently increasing the difficulty of judging segmentation quality. CAST produces very consistent DSC and ICC values, with a maximal discrepancy of 0.01 (DSC-ICC) across all subfields. The pre-trained model, source code, and settings for the CAST toolbox are publicly available.
Assuntos
Hipocampo/diagnóstico por imagem , Redes Neurais de Computação , Adulto , Algoritmos , Automação , Bases de Dados Factuais , Aprendizado Profundo , Feminino , Humanos , Processamento de Imagem Assistida por Computador/métodos , Aprendizado de Máquina , Imageamento por Ressonância Magnética , Masculino , Pessoa de Meia-Idade , Reprodutibilidade dos Testes , Adulto JovemRESUMO
Diffusion tensor magnetic resonance imaging (DTI) is unsurpassed in its ability to map tissue microstructure and structural connectivity in the living human brain. Nonetheless, the angular sampling requirement for DTI leads to long scan times and poses a critical barrier to performing high-quality DTI in routine clinical practice and large-scale research studies. In this work we present a new processing framework for DTI entitled DeepDTI that minimizes the data requirement of DTI to six diffusion-weighted images (DWIs) required by conventional voxel-wise fitting methods for deriving the six unique unknowns in a diffusion tensor using data-driven supervised deep learning. DeepDTI maps the input non-diffusion-weighted (b â= â0) image and six DWI volumes sampled along optimized diffusion-encoding directions, along with T1-weighted and T2-weighted image volumes, to the residuals between the input and high-quality output b = 0 image and DWI volumes using a 10-layer three-dimensional convolutional neural network (CNN). The inputs and outputs of DeepDTI are uniquely formulated, which not only enables residual learning to boost CNN performance but also enables tensor fitting of resultant high-quality DWIs to generate orientational DTI metrics for tractography. The very deep CNN used by DeepDTI leverages the redundancy in local and non-local spatial information and across diffusion-encoding directions and image contrasts in the data. The performance of DeepDTI was systematically quantified in terms of the quality of the output images, DTI metrics, DTI-based tractography and tract-specific analysis results. We demonstrate rotationally-invariant and robust estimation of DTI metrics from DeepDTI that are comparable to those obtained with two b â= â0 images and 21 DWIs for the primary eigenvector derived from DTI and two b â= â0 images and 26-30 DWIs for various scalar metrics derived from DTI, achieving 3.3-4.6 × âacceleration, and twice as good as those of a state-of-the-art denoising algorithm at the group level. The twenty major white-matter tracts can be accurately identified from the tractography of DeepDTI results. The mean distance between the core of the major white-matter tracts identified from DeepDTI results and those from the ground-truth results using 18 âb â= â0 images and 90 DWIs measures around 1-1.5 âmm. DeepDTI leverages domain knowledge of diffusion MRI physics and power of deep learning to render DTI, DTI-based tractography, major white-matter tracts identification and tract-specific analysis more feasible for a wider range of neuroscientific and clinical studies.
Assuntos
Encéfalo/diagnóstico por imagem , Conectoma , Aprendizado Profundo , Imagem de Tensor de Difusão/métodos , Rede Nervosa/diagnóstico por imagem , Redes Neurais de Computação , Humanos , Processamento de Imagem Assistida por Computador/métodosRESUMO
Skin lesion classification is an effective approach aided by computer vision for the diagnosis of skin cancer. Though deep learning models presented advantages over traditional methods and brought tremendous breakthroughs, a precise diagnosis is still challenging because of the intra-class variation and inter-class similarity caused by the diversity of imaging methods and clinicopathology. In this paper, we propose a densely connected convolutional network with an attention and residual learning (ARDT-DenseNet) method for skin lesion classification. Each ARDT block consists of dense blocks, transition blocks and attention and residual modules. Compared to a residual network with the same number of convolutional layers, the size of the parameters of the densely connected network proposed in this paper has been reduced by half, while the accuracy of skin lesion classification is preserved. Our improved densely connected network adds an attention mechanism and residual learning after each dense block and transition block without introducing additional parameters. We evaluate the ARDT-DenseNet model with the ISIC 2016 and ISIC 2017 datasets. Our method achieves an ACC of 85.7% and an AUC of 83.7% in skin lesion classification with ISIC 2016 and an average AUC of 91.8% in skin lesion classification with ISIC 2017. The experimental results show that the method proposed in this paper has achieved a significant improvement in skin lesion classification, which is superior to that of the state-of-the-art method.