Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 1.188
Filtrar
1.
J Imaging Inform Med ; 2024 Aug 08.
Artículo en Inglés | MEDLINE | ID: mdl-39117939

RESUMEN

To propose a deep learning framework "SpineCurve-net" for automated measuring the 3D Cobb angles from computed tomography (CT) images of presurgical scoliosis patients. A total of 116 scoliosis patients were analyzed, divided into a training set of 89 patients (average age 32.4 ± 24.5 years) and a validation set of 27 patients (average age 17.3 ± 5.8 years). Vertebral identification and curve fitting were achieved through U-net and NURBS-net and resulted in a Non-Uniform Rational B-Spline (NURBS) curve of the spine. The 3D Cobb angles were measured in two ways: the predicted 3D Cobb angle (PRED-3D-CA), which is the maximum value in the smoothed angle map derived from the NURBS curve, and the 2D mapping Cobb angle (MAP-2D-CA), which is the maximal angle formed by the tangent vectors along the projected 2D spinal curve. The model segmented spinal masks effectively, capturing easily missed vertebral bodies. Spoke kernel filtering distinguished vertebral regions, centralizing spinal curves. The SpineCurve Network method's Cobb angle (PRED-3D-CA and MAP-2D-CA) measurements correlated strongly with the surgeons' annotated Cobb angle (ground truth, GT) based on 2D radiographs, revealing high Pearson correlation coefficients of 0.983 and 0.934, respectively. This paper proposed an automated technique for calculating the 3D Cobb angle in preoperative scoliosis patients, yielding results that are highly correlated with traditional 2D Cobb angle measurements. Given its capacity to accurately represent the three-dimensional nature of spinal deformities, this method shows potential in aiding physicians to develop more precise surgical strategies in upcoming cases.

2.
Skin Res Technol ; 30(8): e13783, 2024 Aug.
Artículo en Inglés | MEDLINE | ID: mdl-39113617

RESUMEN

BACKGROUND: In recent years, the increasing prevalence of skin cancers, particularly malignant melanoma, has become a major concern for public health. The development of accurate automated segmentation techniques for skin lesions holds immense potential in alleviating the burden on medical professionals. It is of substantial clinical importance for the early identification and intervention of skin cancer. Nevertheless, the irregular shape, uneven color, and noise interference of the skin lesions have presented significant challenges to the precise segmentation. Therefore, it is crucial to develop a high-precision and intelligent skin lesion segmentation framework for clinical treatment. METHODS: A precision-driven segmentation model for skin cancer images is proposed based on the Transformer U-Net, called BiADATU-Net, which integrates the deformable attention Transformer and bidirectional attention blocks into the U-Net. The encoder part utilizes deformable attention Transformer with dual attention block, allowing adaptive learning of global and local features. The decoder part incorporates specifically tailored scSE attention modules within skip connection layers to capture image-specific context information for strong feature fusion. Additionally, deformable convolution is aggregated into two different attention blocks to learn irregular lesion features for high-precision prediction. RESULTS: A series of experiments are conducted on four skin cancer image datasets (i.e., ISIC2016, ISIC2017, ISIC2018, and PH2). The findings show that our model exhibits satisfactory segmentation performance, all achieving an accuracy rate of over 96%. CONCLUSION: Our experiment results validate the proposed BiADATU-Net achieves competitive performance supremacy compared to some state-of-the-art methods. It is potential and valuable in the field of skin lesion segmentation.


Asunto(s)
Melanoma , Neoplasias Cutáneas , Humanos , Neoplasias Cutáneas/diagnóstico por imagen , Neoplasias Cutáneas/patología , Melanoma/diagnóstico por imagen , Melanoma/patología , Algoritmos , Redes Neurales de la Computación , Procesamiento de Imagen Asistido por Computador/métodos , Interpretación de Imagen Asistida por Computador/métodos , Dermoscopía/métodos , Aprendizaje Profundo
3.
Comput Biol Med ; 180: 108947, 2024 Aug 01.
Artículo en Inglés | MEDLINE | ID: mdl-39094324

RESUMEN

Recently, ViT and CNNs based on encoder-decoder architecture have become the dominant model in the field of medical image segmentation. However, there are some deficiencies for each of them: (1) It is difficult for CNNs to capture the interaction between two locations with consideration of the longer distance. (2) ViT cannot acquire the interaction of local context information and carries high computational complexity. To optimize the above deficiencies, we propose a new network for medical image segmentation, which is called FCSU-Net. FCSU-Net uses the proposed collaborative fusion of multi-scale feature block that enables the network to obtain more abundant and more accurate features. In addition, FCSU-Net fuses full-scale feature information through the FFF (Full-scale Feature Fusion) structure instead of simple skip connections, and establishes long-range dependencies on multiple dimensions through the CS (Cross-dimension Self-attention) mechanism. Meantime, every dimension is complementary to each other. Also, CS mechanism has the advantage of convolutions capturing local contextual weights. Finally, FCSU-Net is validated on several datasets, and the results show that FCSU-Net not only has a relatively small number of parameters, but also has a leading segmentation performance.

4.
Artículo en Inglés | MEDLINE | ID: mdl-39094595

RESUMEN

Dynamic 2-[18F] fluoro-2-deoxy-D-glucose positron emission tomography (dFDG-PET) for human brain imaging has considerable clinical potential, yet its utilization remains limited. A key challenge in the quantitative analysis of dFDG-PET is characterizing a patient-specific blood input function, traditionally reliant on invasive arterial blood sampling. This research introduces a novel approach employing non-invasive deep learning model-based computations from the internal carotid arteries (ICA) with partial volume (PV) corrections, thereby eliminating the need for invasive arterial sampling. We present an end-to-end pipeline incorporating a 3D U-Net based ICA-net for ICA segmentation, alongside a Recurrent Neural Network (RNN) based MCIF-net for the derivation of a model-corrected blood input function (MCIF) with PV corrections. The developed 3D U-Net and RNN was trained and validated using a 5-fold cross-validation approach on 50 human brain FDG PET datasets. The ICA-net achieved an average Dice score of 82.18% and an Intersection over Union of 68.54% across all tested scans. Furthermore, the MCIF-net exhibited a minimal root mean squared error of 0.0052. The application of this pipeline to ground truth data for dFDG-PET brain scans resulted in the precise localization of seizure onset regions, which contributed to a successful clinical outcome, with the patient achieving a seizure-free state after treatment. These results underscore the efficacy of the ICA-net and MCIF-net deep learning pipeline in learning the ICA structure's distribution and automating MCIF computation with PV corrections. This advancement marks a significant leap in non-invasive neuroimaging.

5.
Med Phys ; 2024 Aug 01.
Artículo en Inglés | MEDLINE | ID: mdl-39088756

RESUMEN

BACKGROUND: The quality of treatment plans for breast cancer can vary greatly. This variation could be reduced by using dose prediction to automate treatment planning. Our work investigates novel methods for training deep-learning models that are capable of producing high-quality dose predictions for breast cancer treatment planning. PURPOSE: The goal of this work was to compare the performance impact of two novel techniques for deep learning dose prediction models for tangent field treatments for breast cancer. The first technique, a "glowing" mask algorithm, encodes the distance from a contour into each voxel in a mask. The second, a gradient-weighted mean squared error (MSE) loss function, emphasizes the error in high-dose gradient regions in the predicted image. METHODS: Four 3D U-Net deep learning models were trained using the planning CT and contours of the heart, lung, and tumor bed as inputs. The dataset consisted of 305 treatment plans split into 213/46/46 training/validation/test sets using a 70/15/15% split. We compared the impact of novel "glowing" anatomical mask inputs and a novel gradient-weighted MSE loss function to their standard counterparts, binary anatomical masks, and MSE loss, using an ablation study methodology. To assess performance, we examined the mean error and mean absolute error (ME/MAE) in dose across all within-body voxels, the error in mean dose to heart, ipsilateral lung, and tumor bed, dice similarity coefficient (DSC) across isodose volumes defined by 0%-100% prescribed dose thresholds, and gamma analysis (3%/3 mm). RESULTS: The combination of novel glowing masks and gradient weighted loss function yielded the best-performing model in this study. This model resulted in a mean ME of 0.40%, MAE of 2.70%, an error in mean dose to heart and lung of -0.10 and 0.01 Gy, and an error in mean dose to the tumor bed of -0.01%. The median DSC at 50/95/100% isodose levels were 0.91/0.87/0.82. The mean 3D gamma pass rate (3%/3 mm) was 93%. CONCLUSIONS: This study found the combination of novel anatomical mask inputs and loss function for dose prediction resulted in superior performance to their standard counterparts. These results have important implications for the field of radiotherapy dose prediction, as the methods used here can be easily incorporated into many other dose prediction models for other treatment sites. Additionally, this dose prediction model for breast radiotherapy has sufficient performance to be used in an automated planning pipeline for tangent field radiotherapy and has the major benefit of not requiring a PTV for accurate dose prediction.

6.
Comput Biol Med ; 180: 108927, 2024 Aug 02.
Artículo en Inglés | MEDLINE | ID: mdl-39096608

RESUMEN

Rare genetic diseases are difficult to diagnose and this translates in patient's diagnostic odyssey! This is particularly true for more than 900 rare diseases including orodental developmental anomalies such as missing teeth. However, if left untreated, their symptoms can become significant and disabling for the patient. Early detection and rapid management are therefore essential in this context. The i-Dent project aims to supply a pre-diagnostic tool to detect rare diseases with tooth agenesis of varying severity and pattern. To identify missing teeth, image segmentation models (Mask R-CNN, U-Net) have been trained for the automatic detection of teeth on patients' panoramic dental X-rays. Teeth segmentation enables the identification of teeth which are present or missing within the mouth. Furthermore, a dental age assessment is conducted to verify whether the absence of teeth is an anomaly or a characteristic of the patient's age. Due to the small size of our dataset, we developed a new dental age assessment technique based on the tooth eruption rate. Information about missing teeth is then used by a final algorithm based on the agenesis probabilities to propose a pre-diagnosis of a rare disease. The results obtained in detecting three types of genes (PAX9, WNT10A and EDA) by our system are very promising, providing a pre-diagnosis with an average accuracy of 72 %.

7.
Med Image Anal ; 97: 103280, 2024 Jul 22.
Artículo en Inglés | MEDLINE | ID: mdl-39096845

RESUMEN

Medical image segmentation is crucial for healthcare, yet convolution-based methods like U-Net face limitations in modeling long-range dependencies. To address this, Transformers designed for sequence-to-sequence predictions have been integrated into medical image segmentation. However, a comprehensive understanding of Transformers' self-attention in U-Net components is lacking. TransUNet, first introduced in 2021, is widely recognized as one of the first models to integrate Transformer into medical image analysis. In this study, we present the versatile framework of TransUNet that encapsulates Transformers' self-attention into two key modules: (1) a Transformer encoder tokenizing image patches from a convolution neural network (CNN) feature map, facilitating global context extraction, and (2) a Transformer decoder refining candidate regions through cross-attention between proposals and U-Net features. These modules can be flexibly inserted into the U-Net backbone, resulting in three configurations: Encoder-only, Decoder-only, and Encoder+Decoder. TransUNet provides a library encompassing both 2D and 3D implementations, enabling users to easily tailor the chosen architecture. Our findings highlight the encoder's efficacy in modeling interactions among multiple abdominal organs and the decoder's strength in handling small targets like tumors. It excels in diverse medical applications, such as multi-organ segmentation, pancreatic tumor segmentation, and hepatic vessel segmentation. Notably, our TransUNet achieves a significant average Dice improvement of 1.06% and 4.30% for multi-organ segmentation and pancreatic tumor segmentation, respectively, when compared to the highly competitive nn-UNet, and surpasses the top-1 solution in the BrasTS2021 challenge. 2D/3D Code and models are available at https://github.com/Beckschen/TransUNet and https://github.com/Beckschen/TransUNet-3D, respectively.

8.
J Imaging Inform Med ; 2024 Aug 05.
Artículo en Inglés | MEDLINE | ID: mdl-39103563

RESUMEN

Obstructive sleep apnea is characterized by a decrease or cessation of breathing due to repetitive closure of the upper airway during sleep, leading to a decrease in blood oxygen saturation. In this study, employing a U-Net model, we utilized drug-induced sleep endoscopy images to segment the major causes of airway obstruction, including the epiglottis, oropharynx lateral walls, and tongue base. The evaluation metrics included sensitivity, specificity, accuracy, and Dice score, with airway sensitivity at 0.93 (± 0.06), specificity at 0.96 (± 0.01), accuracy at 0.95 (± 0.01), and Dice score at 0.84 (± 0.03), indicating overall high performance. The results indicate the potential for artificial intelligence (AI)-driven automatic interpretation of sleep disorder diagnosis, with implications for standardizing medical procedures and improving healthcare services. The study suggests that advancements in AI technology hold promise for enhancing diagnostic accuracy and treatment efficacy in sleep and respiratory disorders, fostering competitiveness in the medical AI market.

9.
J Imaging Inform Med ; 2024 Aug 05.
Artículo en Inglés | MEDLINE | ID: mdl-39103564

RESUMEN

Retinal vessel segmentation is crucial for the diagnosis of ophthalmic and cardiovascular diseases. However, retinal vessels are densely and irregularly distributed, with many capillaries blending into the background, and exhibit low contrast. Moreover, the encoder-decoder-based network for retinal vessel segmentation suffers from irreversible loss of detailed features due to multiple encoding and decoding, leading to incorrect segmentation of the vessels. Meanwhile, the single-dimensional attention mechanisms possess limitations, neglecting the importance of multidimensional features. To solve these issues, in this paper, we propose a detail-enhanced attention feature fusion network (DEAF-Net) for retinal vessel segmentation. First, the detail-enhanced residual block (DERB) module is proposed to strengthen the capacity for detailed representation, ensuring that intricate features are efficiently maintained during the segmentation of delicate vessels. Second, the multidimensional collaborative attention encoder (MCAE) module is proposed to optimize the extraction of multidimensional information. Then, the dynamic decoder (DYD) module is introduced to preserve spatial information during the decoding process and reduce the information loss caused by upsampling operations. Finally, the proposed detail-enhanced feature fusion (DEFF) module composed of DERB, MCAE and DYD modules fuses feature maps from both encoding and decoding and achieves effective aggregation of multi-scale contextual information. The experiments conducted on the datasets of DRIVE, CHASEDB1, and STARE, achieving Sen of 0.8305, 0.8784, and 0.8654, and AUC of 0.9886, 0.9913, and 0.9911 on DRIVE, CHASEDB1, and STARE, respectively, demonstrate the performance of our proposed network, particularly in the segmentation of fine retinal vessels.

10.
Sci Rep ; 14(1): 18895, 2024 08 14.
Artículo en Inglés | MEDLINE | ID: mdl-39143126

RESUMEN

To develop a deep learning-based model capable of segmenting the left ventricular (LV) myocardium on native T1 maps from cardiac MRI in both long-axis and short-axis orientations. Models were trained on native myocardial T1 maps from 50 healthy volunteers and 75 patients using manual segmentation as the reference standard. Based on a U-Net architecture, we systematically optimized the model design using two different training metrics (Sørensen-Dice coefficient = DSC and Intersection-over-Union = IOU), two different activation functions (ReLU and LeakyReLU) and various numbers of training epochs. Training with DSC metric and a ReLU activation function over 35 epochs achieved the highest overall performance (mean error in T1 10.6 ± 17.9 ms, mean DSC 0.88 ± 0.07). Limits of agreement between model results and ground truth were from -35.5 to + 36.1 ms. This was superior to the agreement between two human raters (-34.7 to + 59.1 ms). Segmentation was as accurate for long-axis views (mean error T1: 6.77 ± 8.3 ms, mean DSC: 0.89 ± 0.03) as for short-axis images (mean error ΔT1: 11.6 ± 19.7 ms, mean DSC: 0.88 ± 0.08). Fully automated segmentation and quantitative analysis of native myocardial T1 maps is possible in both long-axis and short-axis orientations with very high accuracy.


Asunto(s)
Aprendizaje Profundo , Imagen por Resonancia Magnética , Humanos , Imagen por Resonancia Magnética/métodos , Masculino , Femenino , Adulto , Persona de Mediana Edad , Procesamiento de Imagen Asistido por Computador/métodos , Miocardio , Ventrículos Cardíacos/diagnóstico por imagen , Corazón/diagnóstico por imagen
11.
Br J Radiol ; 2024 Aug 14.
Artículo en Inglés | MEDLINE | ID: mdl-39141433

RESUMEN

OBJECTIVES: This study aims to develop an automated approach for estimating the vertical rotation of the thorax, which can be used to assess the technical adequacy of chest X-ray radiographs (CXRs). METHODS: Total 800 chest radiographs were used to train and establish segmentation networks for outlining the lungs and spine regions in chest X-ray images. By measuring the widths of the left and right lungs between the central line of segmented spine and the lateral sides of the segmented lungs, the quantification of thoracic vertical rotation was achieved. Additionally, a life-size, full body anthropomorphic phantom was employed to collect chest radiographic images under various specified rotation angles for assessing the accuracy of the proposed approach. RESULTS: The deep learning networks effectively segmented the anatomical structures of the lungs and spine. The proposed approach demonstrated a mean estimation error of less than 2° for thoracic rotation, surpassing existing techniques and indicating its superiority. CONCLUSIONS: The proposed approach offers a robust assessment of thoracic rotation and presents new possibilities for automated image quality control in chest X-ray examinations. ADVANCES IN KNOWLEDGE: This study presents a novel deep learning-based approach for the automated estimation of vertical thoracic rotation in chest X-ray radiographs. The proposed method enables a quantitative assessment of the technical adequacy of CXR examinations and opens up new possibilities for automated screening and quality control of radiographs.

12.
Artículo en Inglés | MEDLINE | ID: mdl-39142295

RESUMEN

With the advancement of computer-aided diagnosis, the automatic segmentation of COVID-19 infection areas holds great promise for assisting in the timely diagnosis and recovery of patients in clinical practice. Currently, methods relying on U-Net face challenges in effectively utilizing fine-grained semantic information from input images and bridging the semantic gap between the encoder and decoder. To address these issues, we propose an FMD-UNet dual-decoder U-Net network for COVID-19 infection segmentation, which integrates a Fine-grained Feature Squeezing (FGFS) decoder and a Multi-scale Dilated Semantic Aggregation (MDSA) decoder. The FGFS decoder produces fine feature maps through the compression of fine-grained features and a weighted attention mechanism, guiding the model to capture detailed semantic information. The MDSA decoder consists of three hierarchical MDSA modules designed for different stages of input information. These modules progressively fuse different scales of dilated convolutions to process the shallow and deep semantic information from the encoder, and use the extracted feature information to bridge the semantic gaps at various stages, this design captures extensive contextual information while decoding and predicting segmentation, thereby suppressing the increase in model parameters. To better validate the robustness and generalizability of the FMD-UNet, we conducted comprehensive performance evaluations and ablation experiments on three public datasets, and achieved leading Dice Similarity Coefficient (DSC) scores of 84.76, 78.56 and 61.99% in COVID-19 infection segmentation, respectively. Compared to previous methods, the FMD-UNet has fewer parameters and shorter inference time, which also demonstrates its competitiveness.

13.
Diagnostics (Basel) ; 14(15)2024 Jul 24.
Artículo en Inglés | MEDLINE | ID: mdl-39125472

RESUMEN

BACKGROUND: Vision-based pulmonary diagnostics present a unique approach for tracking and measuring natural breathing behaviors through remote imaging. While many existing methods correlate chest and diaphragm movements to respiratory behavior, we look at how the direct visualization of thermal CO2 exhale flow patterns can be tracked to directly measure expiratory flow. METHODS: In this work, we present a novel method for isolating and extracting turbulent exhale flow signals from thermal image sequences through flow-field prediction and optical flow measurement. The objective of this work is to introduce a respiratory diagnostic tool that can be used to capture and quantify natural breathing, to identify and measure respiratory metrics such as breathing rate, flow, and volume. One of the primary contributions of this work is a method for capturing and measuring natural exhale behaviors that describe individualized pulmonary traits. By monitoring subtle individualized respiratory traits, we can perform secondary analysis to identify unique personalized signatures and abnormalities to gain insight into pulmonary function. In our study, we perform data acquisition within a clinical setting to train an inference model (FieldNet) that predicts flow-fields to quantify observed exhale behaviors over time. RESULTS: Expiratory flow measurements capturing individualized flow signatures from our initial cohort demonstrate how the proposed flow field model can be used to isolate and analyze turbulent exhale behaviors and measure anomalous behavior. CONCLUSIONS: Our results illustrate that detailed spatial flow analysis can contribute to unique signatures for identifying patient specific natural breathing behaviors and abnormality detection. This provides the first-step towards a non-contact respiratory technology that directly captures effort-independent behaviors based on the direct measurement of imaged CO2 exhaled airflow patterns.

14.
Comput Biol Med ; 180: 108975, 2024 Aug 16.
Artículo en Inglés | MEDLINE | ID: mdl-39153395

RESUMEN

Skin surface imaging has been used to examine skin lesions with a microscope for over a century and is commonly known as epiluminescence microscopy, dermatoscopy, or dermoscopy. Skin surface microscopy has been recommended to reduce the necessity of biopsy. This imaging technique could improve the clinical diagnostic performance of pigmented skin lesions. Different imaging techniques are employed in dermatology to find diseases. Segmentation and classification are the two main steps in the examination. The classification performance is influenced by the algorithm employed in the segmentation procedure. The most difficult aspect of segmentation is getting rid of the unwanted artifacts. Many deep-learning models are being created to segment skin lesions. In this paper, an analysis of common artifacts is proposed to investigate the segmentation performance of deep learning models with skin surface microscopic images. The most prevalent artifacts in skin images are hair and dark corners. These artifacts can be observed in the majority of dermoscopy images captured through various imaging techniques. While hair detection and removal methods are common, the introduction of dark corner detection and removal represents a novel approach to skin lesion segmentation. A comprehensive analysis of this segmentation performance is assessed using the surface density of artifacts. Assessment of the PH2, ISIC 2017, and ISIC 2018 datasets demonstrates significant enhancements, as reflected by Dice coefficients rising to 93.49 (86.81), 85.86 (79.91), and 75.38 (51.28) respectively, upon artifact removal. These results underscore the pivotal significance of artifact removal techniques in amplifying the efficacy of deep-learning models for skin lesion segmentation.

15.
Sci Rep ; 14(1): 19067, 2024 Aug 17.
Artículo en Inglés | MEDLINE | ID: mdl-39154127

RESUMEN

Building extraction aims to extract building pixels from remote sensing imagery, which plays a significant role in urban planning, dynamic urban monitoring, and many other applications. UNet3+ is widely applied in building extraction from remote sensing images. However, it still faces issues such as low segmentation accuracy, imprecise boundary delineation, and the complexity of network models. Therefore, based on the UNet3+ model, this paper proposes a 3D Joint Attention (3DJA) module that effectively enhances the correlation between local and global features, obtaining more accurate object semantic information and enhancing feature representation. The 3DJA module models semantic interdependence in the vertical and horizontal dimensions to obtain feature map spatial encoding information, as well as in the channel dimensions to increase the correlation between dependent channel graphs. In addition, a bottleneck module is constructed to reduce the number of network parameters and improve model training efficiency. Many experiments are conducted on publicly accessible WHU,INRIA and Massachusetts building dataset, and the benchmarks, BOMSC-Net, CVNet, SCA-Net, SPCL-Net, ACMFNet, MFCF-Net models are selected for comparison with the 3DJA-UNet3+ model proposed in this paper. The experimental results show that 3DJA-UNet3+ achieves competitive results in three evaluation indicators: overall accuracy, mean intersection over union, and F1-score. The code will be available at https://github.com/EnjiLi/3DJA-UNet3Plus .

16.
PeerJ Comput Sci ; 10: e2238, 2024.
Artículo en Inglés | MEDLINE | ID: mdl-39145244

RESUMEN

The abdomen houses multiple vital organs, which are associated with various diseases posing significant risks to human health. Early detection of abdominal organ conditions allows for timely intervention and treatment, preventing deterioration of patients' health. Segmenting abdominal organs aids physicians in more accurately diagnosing organ lesions. However, the anatomical structures of abdominal organs are relatively complex, with organs overlapping each other, sharing similar features, thereby presenting challenges for segmentation tasks. In real medical scenarios, models must demonstrate real-time and low-latency features, necessitating an improvement in segmentation accuracy while minimizing the number of parameters. Researchers have developed various methods for abdominal organ segmentation, ranging from convolutional neural networks (CNNs) to Transformers. However, these methods often encounter difficulties in accurately identifying organ segmentation boundaries. MetaFormer abstracts the framework of Transformers, excluding the multi-head Self-Attention, offering a new perspective for solving computer vision problems and overcoming the limitations of Vision Transformers and CNN backbone networks. To further enhance segmentation effectiveness, we propose a U-shaped network, integrating SEFormer and depthwise cascaded upsampling (dCUP) as the encoder and decoder, respectively, into the UNet structure, named SEF-UNet. SEFormer combines Squeeze-and-Excitation modules with depthwise separable convolutions, instantiating the MetaFormer framework, enhancing the capture of local details and texture information, thereby improving edge segmentation accuracy. dCUP further integrates shallow and deep information layers during the upsampling process. Our model significantly improves segmentation accuracy while reducing the parameter count and exhibits superior performance in segmenting organ edges that overlap each other, thereby offering potential deployment in real medical scenarios.

17.
Sci Rep ; 14(1): 17809, 2024 08 01.
Artículo en Inglés | MEDLINE | ID: mdl-39090263

RESUMEN

Skin microvasculature is vital for human cardiovascular health and thermoregulation, but its imaging and analysis presents significant challenges. Statistical methods such as speckle decorrelation in optical coherence tomography angiography (OCTA) often require multiple co-located B-scans, leading to lengthy acquisitions prone to motion artefacts. Deep learning has shown promise in enhancing accuracy and reducing measurement time by leveraging local information. However, both statistical and deep learning methods typically focus solely on processing individual 2D B-scans, neglecting contextual information from neighbouring B-scans. This limitation compromises spatial context and disregards the 3D features within tissue, potentially affecting OCTA image accuracy. In this study, we propose a novel approach utilising 3D convolutional neural networks (CNNs) to address this limitation. By considering the 3D spatial context, these 3D CNNs mitigate information loss, preserving fine details and boundaries in OCTA images. Our method reduces the required number of B-scans while enhancing accuracy, thereby increasing clinical applicability. This advancement holds promise for improving clinical practices and understanding skin microvascular dynamics crucial for cardiovascular health and thermoregulation.


Asunto(s)
Imagenología Tridimensional , Microvasos , Redes Neurales de la Computación , Piel , Tomografía de Coherencia Óptica , Tomografía de Coherencia Óptica/métodos , Humanos , Microvasos/diagnóstico por imagen , Microvasos/fisiología , Piel/diagnóstico por imagen , Piel/irrigación sanguínea , Imagenología Tridimensional/métodos , Procesamiento de Imagen Asistido por Computador/métodos , Aprendizaje Profundo
18.
Sensors (Basel) ; 24(15)2024 Jul 23.
Artículo en Inglés | MEDLINE | ID: mdl-39123824

RESUMEN

In this work, we investigate the impact of annotation quality and domain expertise on the performance of Convolutional Neural Networks (CNNs) for semantic segmentation of wear on titanium nitride (TiN) and titanium carbonitride (TiCN) coated end mills. Using an innovative measurement system and customized CNN architecture, we found that domain expertise significantly affects model performance. Annotator 1 achieved maximum mIoU scores of 0.8153 for abnormal wear and 0.7120 for normal wear on TiN datasets, whereas Annotator 3 with the lowest expertise achieved significantly lower scores. Sensitivity to annotation inconsistencies and model hyperparameters were examined, revealing that models for TiCN datasets showed a higher coefficient of variation (CV) of 16.32% compared to 8.6% for TiN due to the subtle wear characteristics, highlighting the need for optimized annotation policies and high-quality images to improve wear segmentation.

19.
Comput Biol Med ; 180: 109000, 2024 Aug 11.
Artículo en Inglés | MEDLINE | ID: mdl-39133952

RESUMEN

The fetus's health is evaluated with the biometric parameters obtained from the low-resolution ultrasound images. The accuracy of biometric parameters in existing protocols typically depends on conventional image processing approaches and hence, is prone to error. This study introduces the Attention Gate Double U-Net with Guided Decoder (ADU-GD) model specifically crafted for fetal biometric parameter prediction. The attention network and guided decoder are specifically designed to dynamically merge local features with their global dependencies, enhancing the precision of parameter estimation. The ADU-GD displays superior performance with Mean Absolute Error of 0.99 mm and segmentation accuracy of 99.1 % when benchmarked against the well-established models. The proposed model consistently achieved a high Dice index score of about 99.1 ± 0.8, with a minimal Hausdorff distance of about 1.01 ± 1.07 and a low Average Symmetric Surface Distance of about 0.25 ± 0.21, demonstrating the model's excellence. In a comprehensive evaluation, ADU-GD emerged as a frontrunner, outperforming existing deep-learning models such as Double U-Net, DeepLabv3, FCN-32s, PSPNet, SegNet, Trans U-Net, Swin U-Net, Mask-R2CNN, and RDHCformer models in terms of Mean Absolute Error for crucial fetal dimensions, including Head Circumference, Abdomen Circumference, Femur Length, and BiParietal Diameter. It achieved superior accuracy with MAE values of 2.2 mm, 2.6 mm, 0.6 mm, and 1.2 mm, respectively.

20.
Med Phys ; 2024 Aug 12.
Artículo en Inglés | MEDLINE | ID: mdl-39134025

RESUMEN

BACKGROUND: The landscape of prostate cancer (PCa) segmentation within multiparametric magnetic resonance imaging (MP-MRI) was fragmented, with a noticeable lack of consensus on incorporating background details, culminating in inconsistent segmentation outputs. Given the complex and heterogeneous nature of PCa, conventional imaging segmentation algorithms frequently fell short, prompting the need for specialized research and refinement. PURPOSE: This study sought to dissect and compare various segmentation methods, emphasizing the role of background information and gland masks in achieving superior PCa segmentation. The goal was to systematically refine segmentation networks to ascertain the most efficacious approach. METHODS: A cohort of 232 patients (ages 61-73 years old, prostate-specific antigen: 3.4-45.6 ng/mL), who had undergone MP-MRI followed by prostate biopsies, was analyzed. An advanced segmentation model, namely Attention-Unet, which combines U-Net with attention gates, was employed for training and validation. The model was further enhanced through a multiscale module and a composite loss function, culminating in the development of Matt-Unet. Performance metrics included Dice Similarity Coefficient (DSC) and accuracy (ACC). RESULTS: The Matt-Unet model, which integrated background information and gland masks, outperformed the baseline U-Net model using raw images, yielding significant gains (DSC: 0.7215 vs. 0.6592; ACC: 0.8899 vs. 0.8601, p < 0.001). CONCLUSION: A targeted and practical PCa segmentation method was designed, which could significantly improve PCa segmentation on MP-MRI by combining background information and gland masks. The Matt-Unet model showcased promising capabilities for effectively delineating PCa, enhancing the precision of MP-MRI analysis.

SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA
...