Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 22
Filter
1.
Mar Pollut Bull ; 205: 116644, 2024 Jul 02.
Article in English | MEDLINE | ID: mdl-38959569

ABSTRACT

The cleanup of marine debris is an urgent problem in marine environmental protection. AUVs with visual recognition technology have gradually become a central research issue. However, existing recognition algorithms have slow inference speeds and high computational overhead. They are also affected by blurred images and interference information. To solve these problems, a real-time semantic segmentation network is proposed, called WaterBiSeg-Net. First, we propose the Multi-scale Information Enhancement Module to solve the impact of low-definition and blurred images. Then, to suppress the interference of background information, the Gated Aggregation Layer is proposed. In addition, we propose a method that can extract boundary information directly. Finally, extensive experiments on SUIM and TrashCan datasets show that WaterBiSeg-Net can better complete the task of marine debris segmentation and provide accurate segmentation results for AUVs in real-time. This research offers a low computational cost and real-time solution for AUVs to identify marine debris.

2.
Quant Imaging Med Surg ; 14(6): 4067-4085, 2024 Jun 01.
Article in English | MEDLINE | ID: mdl-38846298

ABSTRACT

Background: The segmentation of prostates from transrectal ultrasound (TRUS) images is a critical step in the diagnosis and treatment of prostate cancer. Nevertheless, the manual segmentation performed by physicians is a time-consuming and laborious task. To address this challenge, there is a pressing need to develop computerized algorithms capable of autonomously segmenting prostates from TRUS images, which sets a direction and form for future development. However, automatic prostate segmentation in TRUS images has always been a challenging problem since prostates in TRUS images have ambiguous boundaries and inhomogeneous intensity distribution. Although many prostate segmentation methods have been proposed, they still need to be improved due to the lack of sensibility to edge information. Consequently, the objective of this study is to devise a highly effective prostate segmentation method that overcomes these limitations and achieves accurate segmentation of prostates in TRUS images. Methods: A three-dimensional (3D) edge-aware attention generative adversarial network (3D EAGAN)-based prostate segmentation method is proposed in this paper, which consists of an edge-aware segmentation network (EASNet) that performs the prostate segmentation and a discriminator network that distinguishes predicted prostates from real prostates. The proposed EASNet is composed of an encoder-decoder-based U-Net backbone network, a detail compensation module (DCM), four 3D spatial and channel attention modules (3D SCAM), an edge enhancement module (EEM), and a global feature extractor (GFE). The DCM is proposed to compensate for the loss of detailed information caused by the down-sampling process of the encoder. The features of the DCM are selectively enhanced by the 3D spatial and channel attention module. Furthermore, an EEM is proposed to guide shallow layers in the EASNet to focus on contour and edge information in prostates. Finally, features from shallow layers and hierarchical features from the decoder module are fused through the GFE to predict the segmentation prostates. Results: The proposed method is evaluated on our TRUS image dataset and the open-source µRegPro dataset. Specifically, experimental results on two datasets show that the proposed method significantly improved the average segmentation Dice score from 85.33% to 90.06%, Jaccard score from 76.09% to 84.11%, Hausdorff distance (HD) score from 8.59 to 4.58 mm, Precision score from 86.48% to 90.58%, and Recall score from 84.79% to 89.24%. Conclusions: A novel 3D EAGAN-based prostate segmentation method is proposed. The proposed method consists of an EASNet and a discriminator network. Experimental results demonstrate that the proposed method has achieved satisfactory results on 3D TRUS image segmentation for prostates.

3.
Comput Biol Med ; 175: 108550, 2024 Jun.
Article in English | MEDLINE | ID: mdl-38701590

ABSTRACT

BACKGROUND AND OBJECTIVE: Complete denture is a common restorative treatment in dental patients and the design of the core components (major connector and retentive mesh) of complete denture metal base (CDMB) is the basis of successful restoration. However, the automated design process of CDMB has become a challenging task primarily due to the complexity of manual interaction, low personalization, and low design accuracy. METHODS: To solve the existing problems, we develop a computer-aided Segmentation Network-driven CDMB design framework, called CDMB-SegNet, to automatically generate personalized digital design boundaries for complete dentures of edentulous patients. Specifically, CDMB-SegNet consists of a novel upright-orientation adjustment module (UO-AM), a dental feature-driven segmentation network, and a specific boundary-optimization design module (BO-DM). UO-AM automatically identifies key points for locating spatial attitude of the three-dimensional dental model with arbitrary posture, while BO-DM can result in smoother and more personalized designs for complete denture. In addition, to achieve efficient and accurate feature extraction and segmentation of 3D edentulous models with irregular gingival tissues, the light-weight backbone network is also incorporated into CDMB-SegNet. RESULTS: Experimental results on a large clinical dataset showed that CDMB-SegNet can achieve superior performance over the state-of-the-art methods. Quantitative evaluation (major connector/retentive mesh) showed improved Accuracy (98.54 ± 0.58 %/97.73 ± 0.92 %) and IoU (87.42 ± 5.48 %/70.42 ± 7.95 %), and reduced Maximum Symmetric Surface Distance (4.54 ± 2.06 mm/4.62 ± 1.68 mm), Average Symmetric Surface Distance (1.45 ± 0.63mm/1.28 ± 0.54 mm), Roughness Rate (6.17 ± 1.40 %/6.80 ± 1.23 %) and Vertices Number (23.22 ± 1.85/43.15 ± 2.72). Moreover, CDMB-SegNet shortened the overall design time to around 4 min, which is one tenth of the comparison methods. CONCLUSIONS: CDMB-SegNet is the first intelligent neural network for automatic CDMB design driven by oral big data and dental features. The designed CDMB is able to couple with patient's personalized dental anatomical morphology, providing higher clinical applicability compared with the state-of-the-art methods.


Subject(s)
Denture, Complete , Humans , Denture Design/methods , Neural Networks, Computer , Computer-Aided Design
4.
Int J Neural Syst ; 34(5): 2450022, 2024 May.
Article in English | MEDLINE | ID: mdl-38487872

ABSTRACT

Deep convolutional neural networks have shown advanced performance in accurately segmenting images. In this paper, an SNP-like convolutional neuron structure is introduced, abstracted from the nonlinear mechanism in nonlinear spiking neural P (NSNP) systems. Then, a U-shaped convolutional neural network named SNP-like parallel-convolutional network, or SPC-Net, is constructed for segmentation tasks. The dual-convolution concatenate (DCC) and dual-convolution addition (DCA) network blocks are designed, respectively, in the encoder and decoder stages. The two blocks employ parallel convolution with different kernel sizes to improve feature representation ability and make full use of spatial detail information. Meanwhile, different feature fusion strategies are used to fuse their features to achieve feature complementarity and augmentation. Furthermore, a dual-scale pooling (DSP) module in the bottleneck is designed to improve the feature extraction capability, which can extract multi-scale contextual information and reduce information loss while extracting salient features. The SPC-Net is applied in medical image segmentation tasks and is compared with several recent segmentation methods on the GlaS and CRAG datasets. The proposed SPC-Net achieves 90.77% DICE coefficient, 83.76% IoU score and 83.93% F1 score, 86.33% ObjDice coefficient, 135.60 Obj-Hausdorff distance, respectively. The experimental results show that the proposed model can achieve good segmentation performance.


Subject(s)
Image Processing, Computer-Assisted , Neural Networks, Computer
5.
Acad Radiol ; 31(2): 617-627, 2024 Feb.
Article in English | MEDLINE | ID: mdl-37330356

ABSTRACT

RATIONALE AND OBJECTIVES: Ki67 proliferation index is associated with more aggressive tumor behavior and recurrence of pituitary adenomas (PAs). Recently, radiomics and deep learning have been introduced into the study of pituitary tumors. The present study aimed to investigate the feasibility of predicting the Ki67 proliferation index of PAs using the deep segmentation network and radiomics analysis based on multiparameter MRI. MATERIALS AND METHODS: First, the cfVB-Net autosegmentation model was trained; subsequently, its performance was evaluated in terms of the dice similarity coefficient (DSC). In the present study, 1214 patients were classified into the high Ki67 expression group (HG) and the low Ki67 expression group (LG). Analyses of three classification models based on radiomics features were performed to distinguish HG from LG. Clinical factors, imaging features, and Radscores were collectively used to create a nomogram in order to effectively predict Ki67 expression. RESULTS: The cfVB-Net segmentation model demonstrated good performance (DSC: 0.723-0.930). Overall, 18, 15, and 11 optimal features in contrast-enhanced (CE) T1WI, T1WI, and T2WI were obtained for differentiating between HG and LG, respectively. Notably, the best results were presented in the bagging decision tree when CE T1WI and T1WI were combined (area under the receiver operating characteristic curve: training set, 0.927; validation set, 0.831; and independent testing set, 0.825). In the nomogram, age, Hardy' grade, and Radscores were identified as risk predictors of high Ki67 expression. CONCLUSION: The deep segmentation network and radiomics analysis based on multiparameter MRI exhibited good performance and clinical application value in predicting the expression of Ki67 in PAs.


Subject(s)
Adenoma , Pituitary Neoplasms , Humans , Pituitary Neoplasms/diagnostic imaging , Radiomics , Ki-67 Antigen , Magnetic Resonance Imaging , Adenoma/diagnostic imaging , Adenoma/surgery , Retrospective Studies
6.
Mar Pollut Bull ; 198: 115874, 2024 Jan.
Article in English | MEDLINE | ID: mdl-38056290

ABSTRACT

The oil spill accidents on the sea surface pose a severe threat to the marine environment and human health. This paper proposes a novel Semantic Segmentation Network (SSN) for processing oil spill images so that low-contrast oil spills on the sea surface can be accurately identified. After the detection accuracy and real-time performance of the current SSNs are compared, the basic network architecture of DeeplabV3+ based target detection is analyzed. The standard convolution is replaced by the Omni-dimensional Dynamic Convolution (ODConv) in the Ghost Module Depth-Wise separable Convolution (DWConv) to further enhance the feature extraction ability of the network. Furthermore, a new DeeplabV3+ based network with ODGhostNetV2 is constructed as the main feature extraction module, and an Adaptive Triplet Attention (ATA) module is deployed in the encoder and decoder at the same time. This not only improves the richness of semantic features but also increases the following receptive fields of the network model. ATA integrates the Adaptively Spatial Feature Fusion (ASFF) module to optimize the weight assignment problem in the feature map fusion process. The ablation experiments are conducted to verify the proposed network which show high accuracy and good real-time performance for the oil spill detection.


Subject(s)
Petroleum Pollution , Humans , Semantics , Oceans and Seas
7.
Med Phys ; 51(4): 2759-2771, 2024 Apr.
Article in English | MEDLINE | ID: mdl-38108587

ABSTRACT

BACKGROUND: Accurate segmentation of lung nodules is of great significance for early screening and diagnosis of lung cancer. PURPOSE: However, the heterogeneity of lung nodules and the similarities between them and other lung tissues make it difficult to accurately segment these nodules. As regards the use of deep learning to segment lung nodules, convolutional neural networks would gradually lead to errors accumulating at the network layer due to the presence of multiple upsampling and downsampling layers, resulting in poor segmentation results. METHODS: In this study, we developed a refined segmentation network (RS-Net) for lung nodule segmentation to solve this problem. Accordingly, the proposed RS-Net was first used to locate the core region of the lung nodules and to gradually refine the segmentation results of the core region. In addition, to solve the problem of misdetection of small-sized nodules owing to the imbalance of positive and negative samples, we devised an average dice-loss function computed on nodule level. By calculating the loss of each nodule sample to measure the overall loss, the network can address the misdetection problem of lung nodules with smaller diameters more efficiently. RESULTS: Our method was evaluated based on 1055 lung nodules from Lung Image Database Consortium data and a set of 120 lung nodules collected from Shanghai Chest Hospital for additional validation. The segmentation dice coefficients of RS-Net on these two datasets were 85.90% and 81.13%, respectively. The analysis of the segmentation effect of different properties and sizes of nodules indicates that RS-Net yields a stable segmentation effect. CONCLUSIONS: The results show that the segmentation strategy based on gradual refinement can considerably improve the segmentation of lung nodules.


Subject(s)
Lung Neoplasms , Tomography, X-Ray Computed , Humans , Tomography, X-Ray Computed/methods , China , Lung Neoplasms/diagnostic imaging , Lung/diagnostic imaging , Neural Networks, Computer , Image Processing, Computer-Assisted/methods
8.
Int J Med Robot ; 19(6): e2569, 2023 Dec.
Article in English | MEDLINE | ID: mdl-37634070

ABSTRACT

During percutaneous coronary intervention, the guiding catheter plays an important role. Tracking the catheter tip placed at the coronary ostium in the X-ray fluoroscopy sequence can obtain image displacement information caused by the heart beating, which can help dynamic coronary roadmap overlap on X-ray fluoroscopy images. Due to a low exposure dose, the X-ray fluoroscopy is noisy and low contrast, which causes some difficulties in tracking. In this paper, we developed a new catheter tip tracking framework. First, a lightweight efficient catheter tip segmentation network is proposed and boosted by a self-distillation training mechanism. Then, the Bayesian filtering post-processing method is used to consider the sequence information to refine the single image segmentation results. By separating the segmentation results into several groups based on connectivity, our framework can track multiple catheter tips. The proposed tracking framework is validated on a clinical X-ray sequence dataset.


Subject(s)
Catheters , Image Processing, Computer-Assisted , Humans , X-Rays , Bayes Theorem , Image Processing, Computer-Assisted/methods , Fluoroscopy/methods
10.
Sensors (Basel) ; 23(3)2023 Jan 29.
Article in English | MEDLINE | ID: mdl-36772541

ABSTRACT

Malaria is a life-threatening disease caused by parasites that are transmitted to humans through the bites of infected mosquitoes. The early diagnosis and treatment of malaria are crucial for reducing morbidity and mortality rates, particularly in developing countries where the disease is prevalent. In this article, we present a novel convolutional neural network (CNN) architecture for detecting malaria from blood samples with a 99.68% accuracy. Our method outperforms the existing approaches in terms of both accuracy and speed, making it a promising tool for malaria diagnosis in resource-limited settings. The CNN was trained on a large dataset of blood smears and was able to accurately classify infected and uninfected samples with high sensitivity and specificity. Additionally, we present an analysis of model performance on different subtypes of malaria and discuss the implications of our findings for the use of deep learning in infectious disease diagnosis.


Subject(s)
Deep Learning , Humans , Neural Networks, Computer , Diagnosis, Computer-Assisted/methods , Early Diagnosis
11.
Comput Biol Med ; 154: 106555, 2023 03.
Article in English | MEDLINE | ID: mdl-36701967

ABSTRACT

Hypopharyngeal cancer (HPC) is a rare disease. Therefore, it is a challenge to automatically segment HPC tumors and metastatic lymph nodes (HPC risk areas) from medical images with the small-scale dataset. Combining low-level details and high-level semantics from feature maps in different scales can improve the accuracy of segmentation. Herein, we propose a Multi-Modality Transfer Learning Network with Hybrid Bilateral Encoder (Twist-Net) for Hypopharyngeal Cancer Segmentation. Specifically, we propose a Bilateral Transition (BT) block and a Bilateral Gather (BG) block to twist (fuse) high-level semantic feature maps and low-level detailed feature maps. We design a block with multi-receptive field extraction capabilities, M Block, to capture multi-scale information. To avoid overfitting caused by the small scale of the dataset, we propose a transfer learning method that can transfer priors experience from large computer vision datasets to multi-modality medical imaging datasets. Compared with other methods, our method outperforms other methods on HPC dataset, achieving the highest Dice of 82.98%. Our method is also superior to other methods on two public medical segmentation datasets, i.e., the CHASE_DB1 dataset and BraTS2018 dataset. On these two datasets, the Dice of our method is 79.83% and 84.87%, respectively. The code is available at: https://github.com/zhongqiu1245/TwistNet.


Subject(s)
Hypopharyngeal Neoplasms , Humans , Hypopharyngeal Neoplasms/diagnostic imaging , Learning , Rare Diseases , Semantics , Machine Learning , Image Processing, Computer-Assisted
12.
Med Phys ; 50(5): 3027-3038, 2023 May.
Article in English | MEDLINE | ID: mdl-36463516

ABSTRACT

BACKGROUND: Manual contouring is very labor-intensive, time-consuming, and subject to intra- and inter-observer variability. An automated deep learning approach to fast and accurate contouring and segmentation is desirable during radiotherapy treatment planning. PURPOSE: This work investigates an efficient deep-learning-based segmentation algorithm in abdomen computed tomography (CT) to facilitate radiation treatment planning. METHODS: In this work, we propose a novel deep-learning model utilizing U-shaped multi-layer perceptron mixer (MLP-Mixer) and convolutional neural network (CNN) for multi-organ segmentation in abdomen CT images. The proposed model has a similar structure to V-net, while a proposed MLP-Convolutional block replaces each convolutional block. The MLP-Convolutional block consists of three components: an early convolutional block for local features extraction and feature resampling, a token-based MLP-Mixer layer for capturing global features with high efficiency, and a token projector for pixel-level detail recovery. We evaluate our proposed network using: (1) an institutional dataset with 60 patient cases and (2) a public dataset (BCTV) with 30 patient cases. The network performance was quantitatively evaluated in three domains: (1) volume similarity between the ground truth contours and the network predictions using the Dice score coefficient (DSC), sensitivity, and precision; (2) surface similarity using Hausdorff distance (HD), mean surface distance (MSD) and residual mean square distance (RMS); and (3) the computational complexity reported by the number of network parameters, training time, and inference time. The performance of the proposed network is compared with other state-of-the-art networks. RESULTS: In the institutional dataset, the proposed network achieved the following volume similarity measures when averaged over all organs: DSC = 0.912, sensitivity = 0.917, precision = 0.917, average surface similarities were HD = 11.95 mm, MSD = 1.90 mm, RMS = 3.86 mm. The proposed network achieved DSC = 0.786 and HD = 9.04 mm on the public dataset. The network also shows statistically significant improvement, which is evaluated by a two-tailed Wilcoxon Mann-Whitney U test, on right lung (MSD where the maximum p-value is 0.001), spinal cord (sensitivity, precision, HD, RMSD where p-value ranges from 0.001 to 0.039), and stomach (DSC where the maximum p-value is 0.01) over all other competing networks. On the public dataset, the network report statistically significant improvement, which is shown by the Wilcoxon Mann-Whitney test, on pancreas (HD where the maximum p-value is 0.006), left (HD where the maximum p-value is 0.022) and right adrenal glands (DSC where the maximum p-value is 0.026). In both datasets, the proposed method can generate contours in less than 5 s. Overall, the proposed MLP-Vnet demonstrates comparable or better performance than competing methods with much lower memory complexity and higher speed. CONCLUSIONS: The proposed MLP-Vnet demonstrates superior segmentation performance, in terms of accuracy and efficiency, relative to state-of-the-art methods. This reliable and efficient method demonstrates potential to streamline clinical workflows in abdominal radiotherapy, which may be especially important for online adaptive treatments.


Subject(s)
Neural Networks, Computer , Tomography, X-Ray Computed , Humans , Abdomen/diagnostic imaging , Algorithms , Lung , Image Processing, Computer-Assisted/methods
13.
Sensors (Basel) ; 22(19)2022 Sep 29.
Article in English | MEDLINE | ID: mdl-36236516

ABSTRACT

Assisted reproductive technology is helping humans by addressing infertility using different medical procedures that help in a successful pregnancy. In vitro fertilization (IVF) is one of those assisted reproduction methods in which the sperm and eggs are combined outside the human body in a specialized environment and kept for growth. Assisted reproductive technology is helping humans by addressing infertility using different medical procedures that help in a successful pregnancy. The morphology of the embryological components is highly related to the success of the assisted reproduction procedure. In approximately 3-5 days, the embryo transforms into the blastocyst. To prevent the multiple-birth risk and to increase the chance of pregnancy the embryologist manually analyzes the blastocyst components and selects valuable embryos to transfer to the women's uterus. The manual microscopic analysis of blastocyst components, such as trophectoderm, zona pellucida, blastocoel, and inner cell mass, is time-consuming and requires keen expertise to select a viable embryo. Artificial intelligence is easing medical procedures by the successful implementation of deep learning algorithms that mimic the medical doctor's knowledge to provide a better diagnostic procedure that helps in reducing the diagnostic burden. The deep learning-based automatic detection of these blastocyst components can help to analyze the morphological properties to select viable embryos. This research presents a deep learning-based embryo component segmentation network (ECS-Net) that accurately detects trophectoderm, zona pellucida, blastocoel, and inner cell mass for embryological analysis. The proposed method (ECS-Net) is based on a shallow deep segmentation network that uses two separate streams produced by a base convolutional block and a depth-wise separable convolutional block. Both streams are densely concatenated in combination with two dense skip paths to produce powerful features before and after upsampling. The proposed ECS-Net is evaluated on a publicly available microscopic blastocyst image dataset, the experimental segmentation results confirm the efficacy of the proposed method. The proposed ECS-Net is providing a mean Jaccard Index (Mean JI) of 85.93% for embryological analysis.


Subject(s)
Artificial Intelligence , Infertility , Female , Fertilization in Vitro/methods , Humans , Male , Pregnancy , Reproduction , Semen
14.
Math Biosci Eng ; 19(4): 4178-4195, 2022 02 17.
Article in English | MEDLINE | ID: mdl-35341293

ABSTRACT

The tire factory mainly inspects tire quality through X-ray images. In this paper, an end-to-end lightweight semantic segmentation network is proposed to realize the error detection of bead toe. In the network, firstly, the texture feature of different regions of tire is extracted by an encoder. Then, we introduce a decoder to fuse the output feature of the encoder. As the dimension of the feature maps is reduced, the positions of bead toe in the X-ray image have been recorded. When evaluating the final segmentation effect, we propose a local mIoU(L-mIoU) index. The segmentation accuracy and reasoning speed of the network are verified on the tire X-ray image set. Specifically, for 512 × 512 input images, we achieve 97.1% mIoU and 92.4% L-mIoU. Alternatively, the bead toe coordinates are calculated using only 1.0 s.


Subject(s)
Image Processing, Computer-Assisted , Semantics , Image Processing, Computer-Assisted/methods , X-Rays
15.
Med Image Anal ; 73: 102196, 2021 10.
Article in English | MEDLINE | ID: mdl-34365142

ABSTRACT

Automatic medical image segmentation plays a crucial role in many medical image analysis applications, such as disease diagnosis and prognosis. Despite the extensive progress of existing deep learning based models for medical image segmentation, they focus on extracting accurate features by designing novel network structures and solely utilize fully connected (FC) layer for pixel-level classification. Considering the insufficient capability of FC layer to encode the extracted diverse feature representations, we propose a Hierarchical Segmentation (HieraSeg) Network for medical image segmentation and devise a Hierarchical Fully Connected (HFC) layer. Specifically, it consists of three classifiers and decouples each category into several subcategories by introducing multiple weight vectors to denote the diverse characteristics in each category. A subcategory-level and a category-level learning schemes are then designed to explicitly enforce the discrepant subcategories and automatically capture the most representative characteristics. Hence, the HFC layer can fit the variant characteristics so as to derive an accurate decision boundary. To enhance the robustness of HieraSeg Network with the variability of lesions, we further propose a Dynamic-Weighting HieraSeg (DW-HieraSeg) Network, which introduces an Image-level Weight Net (IWN) and a Pixel-level Weight Net (PWN) to learn data-driven curriculum. Through progressively incorporating informative images and pixels in an easy-to-hard manner, DW-HieraSeg Network is able to eliminate local optimums and accelerate the training process. Additionally, a class balanced loss is proposed to constrain the PWN for preventing the overfitting problem in minority category. Comprehensive experiments on three benchmark datasets, EndoScene, ISIC and Decathlon, show our newly proposed HieraSeg and DW-HieraSeg Networks achieve state-of-the-art performance, which clearly demonstrates the effectiveness of the proposed approaches for medical image segmentation.


Subject(s)
Image Processing, Computer-Assisted , Humans
16.
PeerJ Comput Sci ; 7: e368, 2021.
Article in English | MEDLINE | ID: mdl-33817018

ABSTRACT

The pandemic of Coronavirus Disease-19 (COVID-19) has spread around the world, causing an existential health crisis. Automated detection of COVID-19 infections in the lungs from Computed Tomography (CT) images offers huge potential in tackling the problem of slow detection and augments the conventional diagnostic procedures. However, segmenting COVID-19 from CT Scans is problematic, due to high variations in the types of infections and low contrast between healthy and infected tissues. While segmenting Lung CT Scans for COVID-19, fast and accurate results are required and furthermore, due to the pandemic, most of the research community has opted for various cloud based servers such as Google Colab, etc. to develop their algorithms. High accuracy can be achieved using Deep Networks but the prediction time would vary as the resources are shared amongst many thus requiring the need to compare different lightweight segmentation model. To address this issue, we aim to analyze the segmentation of COVID-19 using four Convolutional Neural Networks (CNN). The images in our dataset are preprocessed where the motion artifacts are removed. The four networks are UNet, Segmentation Network (Seg Net), High-Resolution Network (HR Net) and VGG UNet. Trained on our dataset of more than 3,000 images, HR Net was found to be the best performing network achieving an accuracy of 96.24% and a Dice score of 0.9127. The analysis shows that lightweight CNN models perform better than other neural net models when to segment infectious tissue due to COVID-19 from CT slices.

17.
Front Neuroinform ; 15: 782262, 2021.
Article in English | MEDLINE | ID: mdl-34975444

ABSTRACT

Convolutional neural networks (CNNs) have brought hope for the medical image auxiliary diagnosis. However, the shortfall of labeled medical image data is the bottleneck that limits the performance improvement of supervised CNN methods. In addition, annotating a large number of labeled medical image data is often expensive and time-consuming. In this study, we propose a co-optimization learning network (COL-Net) for Magnetic Resonance Imaging (MRI) segmentation of ischemic penumbra tissues. COL-Net base on the limited labeled samples and consists of an unsupervised reconstruction network (R), a supervised segmentation network (S), and a transfer block (T). The reconstruction network extracts the robust features from reconstructing pseudo unlabeled samples, which is the auxiliary branch of the segmentation network. The segmentation network is used to segment the target lesions under the limited labeled samples and the auxiliary of the reconstruction network. The transfer block is used to co-optimization the feature maps between the bottlenecks of the reconstruction network and segmentation network. We propose a mix loss function to optimize COL-Net. COL-Net is verified on the public ischemic penumbra segmentation challenge (SPES) with two dozen labeled samples. Results demonstrate that COL-Net has high predictive accuracy and generalization with the Dice coefficient of 0.79. The extended experiment also shows COL-Net outperforms most supervised segmentation methods. COL-Net is a meaningful attempt to alleviate the limited labeled sample problem in medical image segmentation.

18.
Entropy (Basel) ; 22(8)2020 Jul 30.
Article in English | MEDLINE | ID: mdl-33286615

ABSTRACT

Medical image segmentation is an important part of medical image analysis. With the rapid development of convolutional neural networks in image processing, deep learning methods have achieved great success in the field of medical image processing. Deep learning is also used in the field of auxiliary diagnosis of glaucoma, and the effective segmentation of the optic disc area plays an important assistant role in the diagnosis of doctors in the clinical diagnosis of glaucoma. Previously, many U-Net-based optic disc segmentation methods have been proposed. However, the channel dependence of different levels of features is ignored. The performance of fundus image segmentation in small areas is not satisfactory. In this paper, we propose a new aggregation channel attention network to make full use of the influence of context information on semantic segmentation. Different from the existing attention mechanism, we exploit channel dependencies and integrate information of different scales into the attention mechanism. At the same time, we improved the basic classification framework based on cross entropy, combined the dice coefficient and cross entropy, and balanced the contribution of dice coefficients and cross entropy loss to the segmentation task, which enhanced the performance of the network in small area segmentation. The network retains more image features, restores the significant features more accurately, and further improves the segmentation performance of medical images. We apply it to the fundus optic disc segmentation task. We demonstrate the segmentation performance of the model on the Messidor dataset and the RIM-ONE dataset, and evaluate the proposed architecture. Experimental results show that our network architecture improves the prediction performance of the base architectures under different datasets while maintaining the computational efficiency. The results render that the proposed technologies improve the segmentation with 0.0469 overlapping error on Messidor.

19.
Sensors (Basel) ; 20(17)2020 Aug 19.
Article in English | MEDLINE | ID: mdl-32825038

ABSTRACT

To achieve the satisfactory performance of human action recognition, a central task is to address the sub-action sharing problem, especially in similar action classes. Nevertheless, most existing convolutional neural network (CNN)-based action recognition algorithms uniformly divide video into frames and then randomly select the frames as inputs, ignoring the distinct characteristics among different frames. In recent years, depth videos have been increasingly used for action recognition, but most methods merely focus on the spatial information of the different actions without utilizing temporal information. In order to address these issues, a novel energy-guided temporal segmentation method is proposed here, and a multimodal fusion strategy is employed with the proposed segmentation method to construct an energy-guided temporal segmentation network (EGTSN). Specifically, the EGTSN had two parts: energy-guided video segmentation and a multimodal fusion heterogeneous CNN. The proposed solution was evaluated on a public large-scale NTU RGB+D dataset. Comparisons with state-of-the-art methods demonstrate the effectiveness of the proposed network.


Subject(s)
Algorithms , Neural Networks, Computer , Human Activities , Humans
20.
Article in English | MEDLINE | ID: mdl-34286332

ABSTRACT

One of the challenges in developing deep learning algorithms for medical image segmentation is the scarcity of annotated training data. To overcome this limitation, data augmentation and semi-supervised learning (SSL) methods have been developed. However, these methods have limited effectiveness as they either exploit the existing data set only (data augmentation) or risk negative impact by adding poor training examples (SSL). Segmentations are rarely the final product of medical image analysis -they are typically used in downstream tasks to infer higher-order patterns to evaluate diseases. Clinicians take into account a wealth of prior knowledge on biophysics and physiology when evaluating image analysis results. We have used these clinical assessments in previous works to create robust quality-control (QC) classifiers for automated cardiac magnetic resonance (CMR) analysis. In this paper, we propose a novel scheme that uses QC of the downstream task to identify high quality outputs of CMR segmentation networks, that are subsequently utilised for further network training. In essence, this provides quality-aware augmentation of training data in a variant of SSL for segmentation networks (semiQCSeg). We evaluate our approach in two CMR segmentation tasks (aortic and short axis cardiac volume segmentation) using UK Biobank data and two commonly used network architectures (U-net and a Fully Convolutional Network) and compare against supervised and SSL strategies. We show that semiQCSeg improves training of the segmentation networks. It decreases the need for labelled data, while outperforming the other methods in terms of Dice and clinical metrics. SemiQCSeg can be an efficient approach for training segmentation networks for medical image data when labelled datasets are scarce.

SELECTION OF CITATIONS
SEARCH DETAIL
...