Your browser doesn't support javascript.
loading
: 20 | 50 | 100
1 - 20 de 112
1.
Med Image Anal ; 95: 103184, 2024 Jul.
Article En | MEDLINE | ID: mdl-38723320

Synthesizing 7T Susceptibility Weighted Imaging (SWI) from 3T SWI could offer significant clinical benefits by combining the high sensitivity of 7T SWI for neurological disorders with the widespread availability of 3T SWI in diagnostic routines. Although methods exist for synthesizing 7T Magnetic Resonance Imaging (MRI), they primarily focus on traditional MRI modalities like T1-weighted imaging, rather than SWI. SWI poses unique challenges, including limited data availability and the invisibility of certain tissues in individual 3T SWI slices. To address these challenges, we propose a Self-supervised Anatomical Continuity Enhancement (SACE) network to synthesize 7T SWI from 3T SWI using plentiful 3T SWI data and limited 3T-7T paired data. The SACE employs two specifically designed pretext tasks to utilize low-level representations from abundant 3T SWI data for assisting 7T SWI synthesis in a downstream task with limited paired data. One pretext task emphasizes input-specific morphology by balancing the elimination of redundant patterns with the preservation of essential morphology, preventing the blurring of synthetic 7T SWI images. The other task improves the synthesis of tissues that are invisible in a single 3T SWI slice by aligning adjacent slices with the current slice and predicting their difference fields. The downstream task innovatively combines clinical knowledge with brain substructure diagrams to selectively enhance clinically relevant features. When evaluated on a dataset comprising 97 cases (5495 slices), the proposed method achieved a Peak Signal-to-Noise Ratio (PSNR) of 23.05 dB and a Structural Similarity Index (SSIM) of 0.688. Due to the absence of specific methods for 7T SWI, our method was compared with existing enhancement techniques for general 7T MRI synthesis, outperforming these techniques in the context of 7T SWI synthesis. Clinical evaluations have shown that our synthetic 7T SWI is clinically effective, demonstrating its potential as a clinical tool.


Magnetic Resonance Imaging , Humans , Magnetic Resonance Imaging/methods , Image Enhancement/methods , Brain/diagnostic imaging , Algorithms , Image Interpretation, Computer-Assisted/methods
2.
IEEE Trans Image Process ; 33: 2197-2212, 2024.
Article En | MEDLINE | ID: mdl-38470587

Anatomical and functional image fusion is an important technique in a variety of medical and biological applications. Recently, deep learning (DL)-based methods have become a mainstream direction in the field of multi-modal image fusion. However, existing DL-based fusion approaches have difficulty in effectively capturing local features and global contextual information simultaneously. In addition, the scale diversity of features, which is a crucial issue in image fusion, often lacks adequate attention in most existing works. In this paper, to address the above problems, we propose a MixFormer-based multi-scale network, termed as MM-Net, for anatomical and functional image fusion. In our method, an improved MixFormer-based backbone is introduced to sufficiently extract both local features and global contextual information at multiple scales from the source images. The features from different source images are fused at multiple scales based on a multi-source spatial attention-based cross-modality feature fusion (CMFF) module. The scale diversity of the fused features is further enriched by a series of multi-scale feature interaction (MSFI) modules and feature aggregation upsample (FAU) modules. Moreover, a loss function consisting of both spatial domain and frequency domain components is devised to train the proposed fusion model. Experimental results demonstrate that our method outperforms several state-of-the-art fusion methods on both qualitative and quantitative comparisons, and the proposed fusion model exhibits good generalization capability. The source code of our fusion method will be available at https://github.com/yuliu316316.

3.
J Magn Reson Imaging ; 59(5): 1620-1629, 2024 May.
Article En | MEDLINE | ID: mdl-37559435

BACKGROUND: Ultra-high field 7T MRI can provide excellent tissue contrast and anatomical details, but is often cost prohibitive, and is not widely accessible in clinical practice. PURPOSE: To generate synthetic 7T images from widely acquired 3T images with deep learning and to evaluate the feasibility of this approach for brain imaging. STUDY TYPE: Prospective. POPULATION: 33 healthy volunteers and 89 patients with brain diseases, divided into training, and evaluation datasets in the ratio 4:1. SEQUENCE AND FIELD STRENGTH: T1-weighted nonenhanced or contrast-enhanced magnetization-prepared rapid acquisition gradient-echo sequence at both 3T and 7T. ASSESSMENT: A generative adversarial network (SynGAN) was developed to produce synthetic 7T images from 3T images as input. SynGAN training and evaluation were performed separately for nonenhanced and contrast-enhanced paired acquisitions. Qualitative image quality of acquired 3T and 7T images and of synthesized 7T images was evaluated by three radiologists in terms of overall image quality, artifacts, sharpness, contrast, and visualization of vessel using 5-point Likert scales. STATISTICAL TESTS: Wilcoxon signed rank tests to compare synthetic 7T images with acquired 7T and 3T images and intraclass correlation coefficients to evaluate interobserver variability. P < 0.05 was considered significant. RESULTS: Of the 122 paired 3T and 7T MRI scans, 66 were acquired without contrast agent and 56 with contrast agent. The average time to generate synthetic images was ~11.4 msec per slice (2.95 sec per participant). The synthetic 7T images achieved significantly improved tissue contrast and sharpness in comparison to 3T images in both nonenhanced and contrast-enhanced subgroups. Meanwhile, there was no significant difference between acquired 7T and synthetic 7T images in terms of all the evaluation criteria for both nonenhanced and contrast-enhanced subgroups (P ≥ 0.180). DATA CONCLUSION: The deep learning model has potential to generate synthetic 7T images with similar image quality to acquired 7T images. LEVEL OF EVIDENCE: 2 TECHNICAL EFFICACY: Stage 1.


Contrast Media , Magnetic Resonance Imaging , Humans , Feasibility Studies , Prospective Studies , Magnetic Resonance Imaging/methods , Brain/diagnostic imaging
4.
Am J Vet Res ; 85(1)2024 Jan 01.
Article En | MEDLINE | ID: mdl-37852296

OBJECTIVE: The objective of this study was to develop a robust machine-learning approach for efficient detection and grading of sesamoiditis in horses using radiographs, specifically in data-limited conditions. SAMPLE: A dataset of 255 dorsolateral-palmaromedial oblique (DLPMO) and dorsomedial-palmarolateral oblique (DMPLO) equine radiographs were retrospectively acquired from Hagyard Equine Medical Institute. These images were anonymized and classified into 3 categories of sesamoiditis severity (normal, mild, and moderate). METHODS: This study was conducted from February 1, 2023, to August 31, 2023. Two RetinaNet models were used in a cascaded manner, with a self-attention module incorporated into the second RetinaNet's classification subnetwork. The first RetinaNet localized the sesamoid bone in the radiographs, while the second RetinaNet graded the severity of sesamoiditis based on the localized region. Model performance was evaluated using the confusion matrix and average precision (AP). RESULTS: The proposed model demonstrated a promising classification performance with 92.7% accuracy, surpassing the base RetinaNet model. It achieved a mean average precision (mAP) of 81.8%, indicating superior object detection ability. Notably, performance metrics for each severity category showed significant improvement. CLINICAL RELEVANCE: The proposed deep learning-based method can accurately localize the position of sesamoid bones and grade the severity of sesamoiditis on equine radiographs, providing corresponding confidence scores. This approach has the potential to be deployed in a clinical environment, improving the diagnostic interpretation of metacarpophalangeal (fetlock) joint radiographs in horses. Furthermore, by expanding the training dataset, the model may learn to assist in the diagnosis of pathologies in other skeletal regions of the horse.


Deep Learning , Horse Diseases , Sesamoid Bones , Animals , Horses , Retrospective Studies , Horse Diseases/diagnostic imaging , Horse Diseases/pathology , Radiography , Sesamoid Bones/diagnostic imaging
5.
Sensors (Basel) ; 23(22)2023 Nov 13.
Article En | MEDLINE | ID: mdl-38005535

The utilization of Artificial Intelligence (AI) for assessing motor performance in Parkinson's Disease (PD) offers substantial potential, particularly if the results can be integrated into clinical decision-making processes. However, the precise quantification of PD symptoms remains a persistent challenge. The current standard Unified Parkinson's Disease Rating Scale (UPDRS) and its variations serve as the primary clinical tools for evaluating motor symptoms in PD, but are time-intensive and prone to inter-rater variability. Recent work has applied data-driven machine learning techniques to analyze videos of PD patients performing motor tasks, such as finger tapping, a UPDRS task to assess bradykinesia. However, these methods often use abstract features that are not closely related to clinical experience. In this paper, we introduce a customized machine learning approach for the automated scoring of UPDRS bradykinesia using single-view RGB videos of finger tapping, based on the extraction of detailed features that rigorously conform to the established UPDRS guidelines. We applied the method to 75 videos from 50 PD patients collected in both a laboratory and a realistic clinic environment. The classification performance agreed well with expert assessors, and the features selected by the Decision Tree aligned with clinical knowledge. Our proposed framework was designed to remain relevant amid ongoing patient recruitment and technological progress. The proposed approach incorporates features that closely resonate with clinical reasoning and shows promise for clinical implementation in the foreseeable future.


Parkinson Disease , Humans , Parkinson Disease/diagnosis , Hypokinesia/diagnosis , Artificial Intelligence , Machine Learning
6.
Med Image Anal ; 89: 102871, 2023 10.
Article En | MEDLINE | ID: mdl-37480795

Motor dysfunction in Parkinson's Disease (PD) patients is typically assessed by clinicians employing the Movement Disorder Society's Unified Parkinson's Disease Rating Scale (MDS-UPDRS). Such comprehensive clinical assessments are time-consuming, expensive, semi-subjective, and may potentially result in conflicting labels across different raters. To address this problem, we propose an automatic, objective, and weakly-supervised method for labeling PD patients' gait videos. The proposed method accepts videos of patients and classifies their gait scores as normal (Gait score in MDS-UPDRS = 0) or PD (MDS-UPDRS ≥ 1). Unlike previous work, the proposed method does not require a priori MDS-UPDRS ratings for training, utilizing only domain-specific knowledge obtained from neurologists. We propose several labeling functions that classify patients' gait and use a generative model to learn the accuracy of each labeling function in a self-supervised manner. Since results depended upon the estimated values of the patients' 3D poses, and existing pre-trained 3D pose estimators did not yield accurate results, we propose a weakly-supervised 3D human pose estimation method for fine-tuning pre-trained models in a clinical setting. Using leave-one-out evaluations, the proposed method obtains an accuracy of 89% on a dataset of 29 PD subjects - a significant improvement compared to previous work by 7%-10% depending upon the dataset. The method obtained state-of-the-art results on the Human3.6M dataset. Our results suggest that the use of labeling functions may provide a robust means to interpret and classify patient-oriented videos involving motor tasks.


Parkinson Disease , Humans , Gait , Learning
7.
IEEE Trans Pattern Anal Mach Intell ; 45(7): 8143-8158, 2023 Jul.
Article En | MEDLINE | ID: mdl-37015376

This article focuses on conditional generative modeling (CGM) for image data with continuous, scalar conditions (termed regression labels). We propose the first model for this task which is called continuous conditional generative adversarial network (CcGAN). Existing conditional GANs (cGANs) are mainly designed for categorical conditions (e.g., class labels). Conditioning on regression labels is mathematically distinct and raises two fundamental problems: (P1) since there may be very few (even zero) real images for some regression labels, minimizing existing empirical versions of cGAN losses (a.k.a. empirical cGAN losses) often fails in practice; and (P2) since regression labels are scalar and infinitely many, conventional label input mechanisms (e.g., combining a hidden map of the generator/discriminator with a one-hot encoded label) are not applicable. We solve these problems by: (S1) reformulating existing empirical cGAN losses to be appropriate for the continuous scenario; and (S2) proposing a naive label input (NLI) mechanism and an improved label input (ILI) mechanism to incorporate regression labels into the generator and the discriminator. The reformulation in (S1) leads to two novel empirical discriminator losses, termed the hard vicinal discriminator loss (HVDL) and the soft vicinal discriminator loss (SVDL) respectively, and a novel empirical generator loss. Hence, we propose four versions of CcGAN employing different proposed losses and label input mechanisms. The error bounds of the discriminator trained with HVDL and SVDL, respectively, are derived under mild assumptions. To evaluate the performance of CcGANs, two new benchmark datasets (RC-49 and Cell-200) are created. A novel evaluation metric (Sliding Fréchet Inception Distance) is also proposed to replace Intra-FID when Intra-FID is not applicable. Our extensive experiments on several benchmark datasets (i.e., RC-49, UTKFace, Cell-200, and Steering Angle with both low and high resolutions) support the following findings: the proposed CcGAN is able to generate diverse, high-quality samples from the image distribution conditional on a given regression label; and CcGAN substantially outperforms cGAN both visually and quantitatively.

8.
Sensors (Basel) ; 23(3)2023 Jan 31.
Article En | MEDLINE | ID: mdl-36772595

This paper tackles a novel and challenging problem-3D hand pose estimation (HPE) from a single RGB image using partial annotation. Most HPE methods ignore the fact that the keypoints could be partially visible (e.g., under occlusions). In contrast, we propose a deep-learning framework, PA-Tran, that jointly estimates the keypoints status and 3D hand pose from a single RGB image with two dependent branches. The regression branch consists of a Transformer encoder which is trained to predict a set of target keypoints, given an input set of status, position, and visual features embedding from a convolutional neural network (CNN); the classification branch adopts a CNN for estimating the keypoints status. One key idea of PA-Tran is a selective mask training (SMT) objective that uses a binary encoding scheme to represent the status of the keypoints as observed or unobserved during training. In addition, by explicitly encoding the label status (observed/unobserved), the proposed PA-Tran can efficiently handle the condition when only partial annotation is available. Investigating the annotation percentage ranging from 50-100%, we show that training with partial annotation is more efficient (e.g., achieving the best 6.0 PA-MPJPE when using about 85% annotations). Moreover, we provide two new datasets. APDM-Hand, is for synthetic hands with APDM sensor accessories, which is designed for a specific hand task. PD-APDM-Hand, is a real hand dataset collected from Parkinson's Disease (PD) patients with partial annotation. The proposed PA-Tran can achieve higher estimation accuracy when evaluated on both proposed datasets and a more general hand dataset.


Hand , Neural Networks, Computer , Humans
9.
Med Image Anal ; 84: 102693, 2023 02.
Article En | MEDLINE | ID: mdl-36462373

Skin cancer is one of the most common types of malignancy, affecting a large population and causing a heavy economic burden worldwide. Over the last few years, computer-aided diagnosis has been rapidly developed and make great progress in healthcare and medical practices due to the advances in artificial intelligence, particularly with the adoption of convolutional neural networks. However, most studies in skin cancer detection keep pursuing high prediction accuracies without considering the limitation of computing resources on portable devices. In this case, the knowledge distillation (KD) method has been proven as an efficient tool to help improve the adaptability of lightweight models under limited resources, meanwhile keeping a high-level representation capability. To bridge the gap, this study specifically proposes a novel method, termed SSD-KD, that unifies diverse knowledge into a generic KD framework for skin disease classification. Our method models an intra-instance relational feature representation and integrates it with existing KD research. A dual relational knowledge distillation architecture is self-supervised trained while the weighted softened outputs are also exploited to enable the student model to capture richer knowledge from the teacher model. To demonstrate the effectiveness of our method, we conduct experiments on ISIC 2019, a large-scale open-accessed benchmark of skin diseases dermoscopic images. Experiments show that our distilled MobileNetV2 can achieve an accuracy as high as 85% for the classification tasks of 8 different skin diseases with minimal parameters and computing requirements. Ablation studies confirm the effectiveness of our intra- and inter-instance relational knowledge integration strategy. Compared with state-of-the-art knowledge distillation techniques, the proposed method demonstrates improved performance. To the best of our knowledge, this is the first deep knowledge distillation application for multi-disease classification on the large-scale dermoscopy database. Our codes and models are available at https://github.com/enkiwang/Portable-Skin-Lesion-Diagnosis.


Melanoma , Skin Diseases , Skin Neoplasms , Humans , Melanoma/diagnosis , Melanoma/pathology , Artificial Intelligence , Dermoscopy/methods , Skin Diseases/diagnostic imaging , Skin Neoplasms/diagnostic imaging , Skin Neoplasms/pathology
10.
Comput Biol Med ; 150: 106078, 2022 11.
Article En | MEDLINE | ID: mdl-36155266

Resting-state Magnetic resonance imaging-based parcellation aims to group the voxels/vertices non-invasively based on their connectivity profiles, which has achieved great success in understanding the fundamental organizational principles of the human brain. Given the substantial inter-individual variability, the increasing number of studies focus on individual parcellation. However, current methods perform individual parcellations independently or are based on the group prior, requiring expensive computational costs, precise parcel alignment, and extra group information. In this work, an efficient and flexible parcellation framework of individual cerebral cortex was proposed based on a region growing algorithm by merging the unassigned and neighbor vertex with the highest-correlated parcel iteratively. It considered both consistency with prior atlases and individualized functional homogeneity of parcels, which can be applied to a single individual without parcel alignment and group information. The proposed framework was leveraged to 100 unrelated subjects for functional homogeneity comparison and individual identification, and 186 patients with Parkison's disease for symptom prediction. Results demonstrated our framework outperformed other methods in functional homogeneity, and the generated parcellations provided 100% individual identification accuracy. Moreover, the default mode network (DMN) exhibited higher functional homogeneity, intra-subject parcel reproducibility and fingerprinting accuracy, while the sensorimotor network did the opposite, reflecting that the DMN is the most representative, stable, and individual-identifiable network in the resting state. The correlation analysis showed that the severity of the disease symptoms was related negatively to the similarity of individual parcellation and the atlases of healthy populations. The disease severity can be correctly predicted using machine learning models based on individual topographic features such as parcel similarity and parcel size. In summary, the proposed framework not only significantly improves the functional homogeneity but also captures individualized and disease-related brain topography, serving as a potential tool to explore brain function and disease in the future.


Brain , Magnetic Resonance Imaging , Humans , Reproducibility of Results , Brain/diagnostic imaging , Magnetic Resonance Imaging/methods , Brain Mapping/methods , Cerebral Cortex/diagnostic imaging
12.
Science ; 376(6594): 754-758, 2022 05 13.
Article En | MEDLINE | ID: mdl-35549420

Insects have evolved sophisticated reflexes to right themselves in mid-air. Their recovery mechanisms involve complex interactions among the physical senses, muscles, body, and wings, and they must obey the laws of flight. We sought to understand the key mechanisms involved in dragonfly righting reflexes and to develop physics-based models for understanding the control strategies of flight maneuvers. Using kinematic analyses, physical modeling, and three-dimensional flight simulations, we found that a dragonfly uses left-right wing pitch asymmetry to roll its body 180 degrees to recover from falling upside down in ~200 milliseconds. Experiments of dragonflies with blocked vision further revealed that this rolling maneuver is initiated by their ocelli and compound eyes. These results suggest a pathway from the dragonfly's visual system to the muscles regulating wing pitch that underly the recovery. The methods developed here offer quantitative tools for inferring insects' internal actions from their acrobatics, and are applicable to a broad class of natural and robotic flying systems.


Flight, Animal , Odonata , Reflex, Righting , Animals , Flight, Animal/physiology , Odonata/physiology , Wings, Animal/physiology
13.
Comput Biol Med ; 137: 104812, 2021 10.
Article En | MEDLINE | ID: mdl-34507158

In recent years, vast developments in Computer-Aided Diagnosis (CAD) for skin diseases have generated much interest from clinicians and other eventual end-users of this technology. Introducing clinical domain knowledge to these machine learning strategies can help dispel the black box nature of these tools, strengthening clinician trust. Clinical domain knowledge also provides new information channels which can improve CAD diagnostic performance. In this paper, we propose a novel framework for malignant melanoma (MM) detection by fusing clinical images and dermoscopic images. The proposed method combines a multi-labeled deep feature extractor and clinically constrained classifier chain (CC). This allows the 7-point checklist, a clinician diagnostic algorithm, to be included in the decision level while maintaining the clinical importance of the major and minor criteria in the checklist. Our proposed framework achieved an average accuracy of 81.3% for detecting all criteria and melanoma when testing on a publicly available 7-point checklist dataset. This is the highest reported results, outperforming state-of-the-art methods in the literature by 6.4% or more. Analyses also show that the proposed system surpasses the single modality system of using either clinical images or dermoscopic images alone and the systems without adopting the approach of multi-label and clinically constrained classifier chain. Our carefully designed system demonstrates a substantial improvement over melanoma detection. By keeping the familiar major and minor criteria of the 7-point checklist and their corresponding weights, the proposed system may be more accepted by physicians as a human-interpretable CAD tool for automated melanoma detection.


Melanoma , Skin Diseases , Skin Neoplasms , Dermoscopy , Diagnosis, Computer-Assisted , Humans , Melanoma/diagnostic imaging , Skin Neoplasms/diagnostic imaging
14.
IEEE Trans Image Process ; 30: 6701-6714, 2021.
Article En | MEDLINE | ID: mdl-34283715

With the great success of convolutional neural networks (CNNs), interpretation of their internal network mechanism has been increasingly critical, while the network decision-making logic is still an open issue. In the bottom-up hierarchical logic of neuroscience, the decision-making process can be deduced from a series of sub-decision-making processes from low to high levels. Inspired by this, we propose the Concept-harmonized HierArchical INference (CHAIN) interpretation scheme. In CHAIN, a network decision-making process from shallow to deep layers is interpreted by the hierarchical backward inference based on visual concepts from high to low semantic levels. Firstly, we learned a general hierarchical visual-concept representation in CNN layered feature space by concept harmonizing model on a large concept dataset. Secondly, for interpreting a specific network decision-making process, we conduct the concept-harmonized hierarchical inference backward from the highest to the lowest semantic level. Specifically, the network learning for a target concept at a deeper layer is disassembled into that for concepts at shallower layers. Finally, a specific network decision-making process is explained as a form of concept-harmonized hierarchical inference, which is intuitively comparable to the bottom-up hierarchical visual recognition way. Quantitative and qualitative experiments demonstrate the effectiveness of the proposed CHAIN at both instance and class levels.

15.
J Healthc Eng ; 2021: 6632394, 2021.
Article En | MEDLINE | ID: mdl-34094040

Background: Activating vestibular afferents via galvanic vestibular stimulation (GVS) has been recently shown to have a number of complex motor effects in Parkinson's disease (PD), but the basis of these improvements is unclear. The evaluation of network-level connectivity changes may provide us with greater insights into the mechanisms of GVS efficacy. Objective: To test the effects of different GVS stimuli on brain subnetwork interactions in both health control (HC) and PD groups using fMRI. Methods: FMRI data were collected for all participants at baseline (resting state) and under noisy, 1 Hz sinusoidal, and 70-200 Hz multisine GVS. All stimuli were given below sensory threshold, blinding subjects to stimulation. The subnetworks of 15 healthy controls and 27 PD subjects (on medication) were identified in their native space, and their subnetwork interactions were estimated by nonnegative canonical correlation analysis. We then determined if the inferred subnetwork interaction changes were affected by disease and stimulus type and if the stimulus-dependent GVS effects were influenced by demographic features. Results: At baseline, interactions with the visual-cerebellar network were significantly decreased in the PD group. Sinusoidal and multisine GVS improved (i.e., made values approaching those seen in HC) subnetwork interactions more effectively than noisy GVS stimuli overall. Worsening disease severity, apathy, depression, impaired cognitive function, and increasing age all limited the beneficial effects of GVS. Conclusions: Vestibular stimulation has widespread system-level brain influences and can improve subnetwork interactions in PD in a stimulus-dependent manner, with the magnitude of such effects associating with demographics and disease status.


Parkinson Disease , Vestibule, Labyrinth , Brain/diagnostic imaging , Electric Stimulation , Humans , Magnetic Resonance Imaging , Parkinson Disease/therapy , Vestibule, Labyrinth/physiology
16.
IEEE J Biomed Health Inform ; 25(9): 3450-3459, 2021 09.
Article En | MEDLINE | ID: mdl-33905339

Automatic liver and tumor segmentation remain a challenging topic, which subjects to the exploration of 2D and 3D contexts in CT volume. Existing methods are either only focus on the 2D context by treating the CT volume as many independent image slices (but ignore the useful temporal information between adjacent slices), or just explore the 3D context lied in many little voxels (but damage the spatial detail in each slice). These factors lead an inadequate context exploration together for automatic liver and tumor segmentation. In this paper, we propose a novel full-context convolution neural network to bridge the gap between 2D and 3D contexts. The proposed network can utilize the temporal information along the Z axis in CT volume while retaining the spatial detail in each slice. Specifically, a 2D spatial network for intra-slice features extraction and a 3D temporal network for inter-slice features extraction are proposed separately and then are guided by the squeeze-and-excitation layer that allows the flow of 2D context and 3D temporal information. To address the severe class imbalance issue in the CT volume and meanwhile improve the segmentation performance, a loss function consisting of weighted cross-entropy and jaccard distance is proposed. During the network training, the 2D and 3D contexts are learned jointly in an end-to-end way. The proposed network achieves competitive results on the Liver Tumor Segmentation Challenge (LiTS) and the 3D-IRCADB datasets. This method should be a new promising paradigm to explore the contexts for liver and tumor segmentation.


Image Processing, Computer-Assisted , Liver Neoplasms , Humans , Liver Neoplasms/diagnostic imaging , Neural Networks, Computer , Tomography, X-Ray Computed
17.
Med Image Anal ; 68: 101847, 2021 02.
Article En | MEDLINE | ID: mdl-33249389

A computer assisted system for automatic retrieval of medical images with similar image contents can serve as an efficient management tool for handling and mining large scale data, and can also be used as a tool in clinical decision support systems. In this paper, we propose a deep community based automated medical image retrieval framework for extracting similar images from a large scale X-ray database. The framework integrates a deep learning-based image feature generation approach and a network community detection technique to extract similar images. When compared with the state-of-the-art medical image retrieval techniques, the proposed approach demonstrated improved performance. We evaluated the performance of the proposed method on two large scale chest X-ray datasets, where given a query image, the proposed approach was able to extract images with similar disease labels with a precision of 85%. To the best of our knowledge, this is the first deep community based image retrieval application on large scale chest X-ray database.


Decision Support Systems, Clinical , Pattern Recognition, Automated , Algorithms , Humans , Information Storage and Retrieval , X-Rays
18.
IEEE Trans Neural Netw Learn Syst ; 32(4): 1486-1496, 2021 Apr.
Article En | MEDLINE | ID: mdl-32356763

Multi-view data can represent objects from different perspectives and thus provide complementary information for data analysis. A topic of great importance in multi-view learning is to locate a low-dimensional latent subspace, where common semantic features are shared by multiple data sets. However, most existing methods ignore uncorrelated items (i.e., view-specific features) and may cause semantic bias during the process of common feature learning. In this article, we propose a non-negative correlated and uncorrelated feature co-learning (CoUFC) method to address this concern. More specifically, view-specific (uncorrelated) features are identified for each view when learning the common (correlated) feature across views in the latent semantic subspace. By eliminating the effects of uncorrelated information, useful inter-view feature correlations can be captured. We design a new objective function in CoUFC and derive an optimization approach to solve the objective with the analysis on its convergence. Experiments on real-world sensor, image, and text data sets demonstrate that the proposed method outperforms the state-of-the-art multiview learning methods.

19.
Front Neurol ; 11: 407, 2020.
Article En | MEDLINE | ID: mdl-32581993

Although functional connectivity has been extensively studied in MS, robust estimates of both stationary (static connectivity at the time) and dynamic (connectivity variation across time) functional connectivity has not been commonly evaluated and neither has its association to cognition. In this study, we focused on interhemispheric connections as previous research has shown links between anatomical homologous connections and cognition. We examined functional interhemispheric connectivity (IC) in MS during resting-state functional MRI using both stationary and dynamic strategies and related connectivity measures to processing speed performance. Twenty-five patients with relapsing-remitting MS and 41 controls were recruited. Stationary functional IC was assessed between homologous Regions of Interest (ROIs) using correlation. For dynamic IC, a sliding window approach was used to quantify changes between homologous ROIs across time. We related IC measures to cognitive performance with correlation and regression. Compared to control subjects, MS demonstrated increased IC across homologous regions, which accurately predicted performance on the symbol digit modalities test (SDMT) (R 2 = 0.96) and paced auditory serial addition test (PASAT) (R 2 = 0.59). Dynamic measures were not different between the 2 groups, but dynamic IC was related to PASAT scores. The associations between stationary/dynamic connectivity and cognitive tests demonstrated that different aspects of functional IC were associated with cognitive processes. Processing speed measured in SDMT was associated with static interhemispheric connections and better PASAT performance, which requires working memory, sustain attention, and processing speed, was more related to rigid IC, underlining the neurophysiological mechanism of cognition in MS.

20.
IEEE Trans Neural Syst Rehabil Eng ; 28(6): 1271-1281, 2020 06.
Article En | MEDLINE | ID: mdl-32305927

High-density surface electromyography (HD-sEMG) can provide rich temporal and spatial information about muscle activation. However, HD-sEMG signals are often contaminated by power line interference (PLI) and white Gaussian noise (WGN). In the literature, independent component analysis (ICA) and canonical correlation analysis (CCA), as two popular used blind source separation techniques, are widely used for noise removal from HD-sEMG signals. In this paper, a novel method to remove PLI and WGN was proposed based on independent vector analysis (IVA). Taking advantage of both ICA and CCA, this method exploits the higher order and second-order statistical information simultaneously. Our proposed method was applied to both simulated and experimental EMG data for performance evaluation, which was at least 37.50% better than ICA and CCA methods in terms of relative root mean squared error and 28.84% better than ICA and CCA methods according to signal to noise ratio. The results demonstrated that our proposed method performed significantly better than either ICA or CCA. Specifically, the mean signal to noise ratio increased considerably. Our proposed method is a promising tool for denoising HD-sEMG signals while leading to a minimal distortion.


Algorithms , Signal Processing, Computer-Assisted , Electromyography , Humans , Muscle, Skeletal , Normal Distribution , Signal-To-Noise Ratio
...