Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 92
Filtrar
1.
Comput Methods Programs Biomed ; 255: 108367, 2024 Aug 08.
Artigo em Inglês | MEDLINE | ID: mdl-39141962

RESUMO

Medical image segmentation has made remarkable progress with advances in deep learning technology, depending on the quality and quantity of labeled data. Although various deep learning model structures and training methods have been proposed and high performance has been published, limitations such as inter-class accuracy bias exist in actual clinical applications, especially due to the significant lack of small object performance in multi-organ segmentation tasks. In this paper, we propose an uncertainty-based contrastive learning technique, namely UncerNCE, with an optimal hybrid architecture for high classification and segmentation performance of small organs. Our backbone architecture adopts a hybrid network that employs both convolutional and transformer layers, which have demonstrated remarkable performance in recent years. The key proposal of this study addresses the multi-class accuracy bias and resolves a common tradeoff in existing studies between segmenting regions of small objects and reducing overall noise (i.e., false positives). Uncertainty based contrastive learning based on the proposed hybrid network performs spotlight learning on selected regions based on uncertainty and achieved accurate segmentation for all classes while suppressing noise. Comparison with state-of-the-art techniques demonstrates the superiority of our results on BTCV and 1K data.

2.
PeerJ Comput Sci ; 10: e2238, 2024.
Artigo em Inglês | MEDLINE | ID: mdl-39145244

RESUMO

The abdomen houses multiple vital organs, which are associated with various diseases posing significant risks to human health. Early detection of abdominal organ conditions allows for timely intervention and treatment, preventing deterioration of patients' health. Segmenting abdominal organs aids physicians in more accurately diagnosing organ lesions. However, the anatomical structures of abdominal organs are relatively complex, with organs overlapping each other, sharing similar features, thereby presenting challenges for segmentation tasks. In real medical scenarios, models must demonstrate real-time and low-latency features, necessitating an improvement in segmentation accuracy while minimizing the number of parameters. Researchers have developed various methods for abdominal organ segmentation, ranging from convolutional neural networks (CNNs) to Transformers. However, these methods often encounter difficulties in accurately identifying organ segmentation boundaries. MetaFormer abstracts the framework of Transformers, excluding the multi-head Self-Attention, offering a new perspective for solving computer vision problems and overcoming the limitations of Vision Transformers and CNN backbone networks. To further enhance segmentation effectiveness, we propose a U-shaped network, integrating SEFormer and depthwise cascaded upsampling (dCUP) as the encoder and decoder, respectively, into the UNet structure, named SEF-UNet. SEFormer combines Squeeze-and-Excitation modules with depthwise separable convolutions, instantiating the MetaFormer framework, enhancing the capture of local details and texture information, thereby improving edge segmentation accuracy. dCUP further integrates shallow and deep information layers during the upsampling process. Our model significantly improves segmentation accuracy while reducing the parameter count and exhibits superior performance in segmenting organ edges that overlap each other, thereby offering potential deployment in real medical scenarios.

3.
Artigo em Inglês | MEDLINE | ID: mdl-38957182

RESUMO

Organ segmentation is a fundamental requirement in medical image analysis. Many methods have been proposed over the past 6 decades for segmentation. A unique feature of medical images is the anatomical information hidden within the image itself. To bring natural intelligence (NI) in the form of anatomical information accumulated over centuries into deep learning (DL) AI methods effectively, we have recently introduced the idea of hybrid intelligence (HI) that combines NI and AI and a system based on HI to perform medical image segmentation. This HI system has shown remarkable robustness to image artifacts, pathology, deformations, etc. in segmenting organs in the Thorax body region in a multicenter clinical study. The HI system utilizes an anatomy modeling strategy to encode NI and to identify a rough container region in the shape of each object via a non-DL-based approach so that DL training and execution are applied only to the fuzzy container region. In this paper, we introduce several advances related to modeling of the NI component so that it becomes substantially more efficient computationally, and at the same time, is well integrated with the DL portion (AI component) of the system. We demonstrate a 9-40 fold computational improvement in the auto-segmentation task for radiation therapy (RT) planning via clinical studies obtained from 4 different RT centers, while retaining state-of-the-art accuracy of the previous system in segmenting 11 objects in the Thorax body region.

4.
Artigo em Inglês | MEDLINE | ID: mdl-38957740

RESUMO

Organ segmentation is a crucial task in various medical imaging applications. Many deep learning models have been developed to do this, but they are slow and require a lot of computational resources. To solve this problem, attention mechanisms are used which can locate important objects of interest within medical images, allowing the model to segment them accurately even when there is noise or artifact. By paying attention to specific anatomical regions, the model becomes better at segmentation. Medical images have unique features in the form of anatomical information, which makes them different from natural images. Unfortunately, most deep learning methods either ignore this information or do not use it effectively and explicitly. Combined natural intelligence with artificial intelligence, known as hybrid intelligence, has shown promising results in medical image segmentation, making models more robust and able to perform well in challenging situations. In this paper, we propose several methods and models to find attention regions in medical images for deep learning-based segmentation via non-deep-learning methods. We developed these models and trained them using hybrid intelligence concepts. To evaluate their performance, we tested the models on unique test data and analyzed metrics including false negatives quotient and false positives quotient. Our findings demonstrate that object shape and layout variations can be explicitly learned to create computational models that are suitable for each anatomic object. This work opens new possibilities for advancements in medical image segmentation and analysis.

5.
Biomed Eng Online ; 23(1): 52, 2024 Jun 08.
Artigo em Inglês | MEDLINE | ID: mdl-38851691

RESUMO

Accurate segmentation of multiple organs in the head, neck, chest, and abdomen from medical images is an essential step in computer-aided diagnosis, surgical navigation, and radiation therapy. In the past few years, with a data-driven feature extraction approach and end-to-end training, automatic deep learning-based multi-organ segmentation methods have far outperformed traditional methods and become a new research topic. This review systematically summarizes the latest research in this field. We searched Google Scholar for papers published from January 1, 2016 to December 31, 2023, using keywords "multi-organ segmentation" and "deep learning", resulting in 327 papers. We followed the PRISMA guidelines for paper selection, and 195 studies were deemed to be within the scope of this review. We summarized the two main aspects involved in multi-organ segmentation: datasets and methods. Regarding datasets, we provided an overview of existing public datasets and conducted an in-depth analysis. Concerning methods, we categorized existing approaches into three major classes: fully supervised, weakly supervised and semi-supervised, based on whether they require complete label information. We summarized the achievements of these methods in terms of segmentation accuracy. In the discussion and conclusion section, we outlined and summarized the current trends in multi-organ segmentation.


Assuntos
Aprendizado Profundo , Processamento de Imagem Assistida por Computador , Humanos , Processamento de Imagem Assistida por Computador/métodos , Automação
6.
Comput Biol Med ; 177: 108659, 2024 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-38823366

RESUMO

Automatic abdominal organ segmentation is an essential prerequisite for accurate volumetric analysis, disease diagnosis, and tracking by medical practitioners. However, the deformable shapes, variable locations, overlapping with nearby organs, and similar contrast make the segmentation challenging. Moreover, the requirement of a large manually labeled dataset makes it harder. Hence, a semi-supervised contrastive learning approach is utilized to perform the automatic abdominal organ segmentation. Existing 3D deep learning models based on contrastive learning are not able to capture the 3D context of medical volumetric data along three planes/views: axial, sagittal, and coronal views. In this work, a semi-supervised view-adaptive unified model (VAU-model) is proposed to make the 3D deep learning model as view-adaptive to learn 3D context along each view in a unified manner. This method utilizes the novel optimization function that assists the 3D model to learn the 3D context of volumetric medical data along each view in a single model. The effectiveness of the proposed approach is validated on the three types of datasets: BTCV, NIH, and MSD quantitatively and qualitatively. The results demonstrate that the VAU model achieves an average Dice score of 81.61% which is a 3.89% improvement compared to the previous best results for pancreas segmentation in multi-organ dataset BTCV. It also achieves an average Dice score of 77.76% and 76.76% for the pancreas under the single organ non-pathological NIH dataset, and pathological MSD dataset.


Assuntos
Imageamento Tridimensional , Humanos , Imageamento Tridimensional/métodos , Aprendizado Profundo , Abdome/diagnóstico por imagem , Abdome/anatomia & histologia , Tomografia Computadorizada por Raios X/métodos , Pâncreas/diagnóstico por imagem , Pâncreas/anatomia & histologia , Bases de Dados Factuais
7.
ArXiv ; 2024 May 29.
Artigo em Inglês | MEDLINE | ID: mdl-38855547

RESUMO

Image-guided mouse irradiation is essential to understand interventions involving radiation prior to human studies. Our objective is to employ Swin UNEt Transformers (Swin UNETR) to segment native micro-CT and contrast-enhanced micro-CT scans and benchmark the results against 3D no-new-Net (nnU-Net). Swin UNETR reformulates mouse organ segmentation as a sequence-to-sequence prediction task, using a hierarchical Swin Transformer encoder to extract features at 5 resolution levels, and connects to a Fully Convolutional Neural Network (FCNN)-based decoder via skip connections. The models were trained and evaluated on open datasets, with data separation based on individual mice. Further evaluation on an external mouse dataset acquired on a different micro-CT with lower kVp and higher imaging noise was also employed to assess model robustness and generalizability. Results indicate that Swin UNETR consistently outperforms nnU-Net and AIMOS in terms of average dice similarity coefficient (DSC) and Hausdorff distance (HD95p), except in two mice of intestine contouring. This superior performance is especially evident in the external dataset, confirming the model's robustness to variations in imaging conditions, including noise and quality, thereby positioning Swin UNETR as a highly generalizable and efficient tool for automated contouring in pre-clinical workflows.

8.
Med Image Anal ; 97: 103226, 2024 Jun 04.
Artigo em Inglês | MEDLINE | ID: mdl-38852215

RESUMO

The advancement of artificial intelligence (AI) for organ segmentation and tumor detection is propelled by the growing availability of computed tomography (CT) datasets with detailed, per-voxel annotations. However, these AI models often struggle with flexibility for partially annotated datasets and extensibility for new classes due to limitations in the one-hot encoding, architectural design, and learning scheme. To overcome these limitations, we propose a universal, extensible framework enabling a single model, termed Universal Model, to deal with multiple public datasets and adapt to new classes (e.g., organs/tumors). Firstly, we introduce a novel language-driven parameter generator that leverages language embeddings from large language models, enriching semantic encoding compared with one-hot encoding. Secondly, the conventional output layers are replaced with lightweight, class-specific heads, allowing Universal Model to simultaneously segment 25 organs and six types of tumors and ease the addition of new classes. We train our Universal Model on 3410 CT volumes assembled from 14 publicly available datasets and then test it on 6173 CT volumes from four external datasets. Universal Model achieves first place on six CT tasks in the Medical Segmentation Decathlon (MSD) public leaderboard and leading performance on the Beyond The Cranial Vault (BTCV) dataset. In summary, Universal Model exhibits remarkable computational efficiency (6× faster than other dataset-specific models), demonstrates strong generalization across different hospitals, transfers well to numerous downstream tasks, and more importantly, facilitates the extensibility to new classes while alleviating the catastrophic forgetting of previously learned classes. Codes, models, and datasets are available at https://github.com/ljwztc/CLIP-Driven-Universal-Model.

9.
Comput Methods Programs Biomed ; 254: 108280, 2024 Sep.
Artigo em Inglês | MEDLINE | ID: mdl-38878361

RESUMO

BACKGROUND AND OBJECTIVE: Transformer, which is notable for its ability of global context modeling, has been used to remedy the shortcomings of Convolutional neural networks (CNN) and break its dominance in medical image segmentation. However, the self-attention module is both memory and computational inefficient, so many methods have to build their Transformer branch upon largely downsampled feature maps or adopt the tokenized image patches to fit their model into accessible GPUs. This patch-wise operation restricts the network in extracting pixel-level intrinsic structural or dependencies inside each patch, hurting the performance of pixel-level classification tasks. METHODS: To tackle these issues, we propose a memory- and computation-efficient self-attention module to enable reasoning on relatively high-resolution features, promoting the efficiency of learning global information while effective grasping fine spatial details. Furthermore, we design a novel Multi-Branch Transformer (MultiTrans) architecture to provide hierarchical features for handling objects with variable shapes and sizes in medical images. By building four parallel Transformer branches on different levels of CNN, our hybrid network aggregates both multi-scale global contexts and multi-scale local features. RESULTS: MultiTrans achieves the highest segmentation accuracy on three medical image datasets with different modalities: Synapse, ACDC and M&Ms. Compared to the Standard Self-Attention (SSA), the proposed Efficient Self-Attention (ESA) can largely reduce the training memory and computational complexity while even slightly improve the accuracy. Specifically, the training memory cost, FLOPs and Params of our ESA are 18.77%, 20.68% and 74.07% of the SSA. CONCLUSIONS: Experiments on three medical image datasets demonstrate the generality and robustness of the designed network. The ablation study shows the efficiency and effectiveness of our proposed ESA. Code is available at: https://github.com/Yanhua-Zhang/MultiTrans-extension.


Assuntos
Processamento de Imagem Assistida por Computador , Redes Neurais de Computação , Humanos , Processamento de Imagem Assistida por Computador/métodos , Algoritmos , Diagnóstico por Imagem , Bases de Dados Factuais
10.
Phys Med ; 122: 103385, 2024 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-38810392

RESUMO

PURPOSE: The segmentation of abdominal organs in magnetic resonance imaging (MRI) plays a pivotal role in various therapeutic applications. Nevertheless, the application of deep-learning methods to abdominal organ segmentation encounters numerous challenges, especially in addressing blurred boundaries and regions characterized by low-contrast. METHODS: In this study, a multi-scale visual attention-guided network (VAG-Net) was proposed for abdominal multi-organ segmentation based on unpaired multi-sequence MRI. A new visual attention-guided (VAG) mechanism was designed to enhance the extraction of contextual information, particularly at the edge of organs. Furthermore, a new loss function inspired by knowledge distillation was introduced to minimize the semantic disparity between different MRI sequences. RESULTS: The proposed method was evaluated on the CHAOS 2019 Challenge dataset and compared with six state-of-the-art methods. The results demonstrated that our model outperformed these methods, achieving DSC values of 91.83 ± 0.24% and 94.09 ± 0.66% for abdominal multi-organ segmentation in T1-DUAL and T2-SPIR modality, respectively. CONCLUSION: The experimental results show that our proposed method has superior performance in abdominal multi-organ segmentation, especially in the case of small organs such as the kidneys.


Assuntos
Abdome , Processamento de Imagem Assistida por Computador , Imageamento por Ressonância Magnética , Humanos , Processamento de Imagem Assistida por Computador/métodos , Abdome/diagnóstico por imagem , Aprendizado Profundo , Redes Neurais de Computação
11.
Artigo em Inglês | MEDLINE | ID: mdl-38720159

RESUMO

PURPOSE: This paper considers a new problem setting for multi-organ segmentation based on the following observations. In reality, (1) collecting a large-scale dataset from various institutes is usually impeded due to privacy issues; (2) many images are not labeled since the slice-by-slice annotation is costly; and (3) datasets may exhibit inconsistent, partial annotations across different institutes. Learning a federated model from these distributed, partially labeled, and unlabeled samples is an unexplored problem. METHODS: To simulate this multi-organ segmentation problem, several distributed clients and a central server are maintained. The central server coordinates with clients to learn a global model using distributed private datasets, which comprise a small part of partially labeled images and a large part of unlabeled images. To address this problem, a practical framework that unifies partially supervised learning (PSL), semi-supervised learning (SSL), and federated learning (FL) paradigms with PSL, SSL, and FL modules is proposed. The PSL module manages to learn from partially labeled samples. The SSL module extracts valuable information from unlabeled data. Besides, the FL module aggregates local information from distributed clients to generate a global statistical model. With the collaboration of three modules, the presented scheme could take advantage of these distributed imperfect datasets to train a generalizable model. RESULTS: The proposed method was extensively evaluated with multiple abdominal CT datasets, achieving an average result of 84.83% in Dice and 41.62 mm in 95HD for multi-organ (liver, spleen, and stomach) segmentation. Moreover, its efficacy in transfer learning further demonstrated its good generalization ability for downstream segmentation tasks. CONCLUSION: This study considers a novel problem of multi-organ segmentation, which aims to develop a generalizable model using distributed, partially labeled, and unlabeled CT images. A practical framework is presented, which, through extensive validation, has proved to be an effective solution, demonstrating strong potential in addressing this challenging problem.

12.
Phys Eng Sci Med ; 2024 Apr 24.
Artigo em Inglês | MEDLINE | ID: mdl-38656437

RESUMO

Cervical cancer is a common cancer in women globally, with treatment usually involving radiation therapy (RT). Accurate segmentation for the tumour site and organ-at-risks (OARs) could assist in the reduction of treatment side effects and improve treatment planning efficiency. Cervical cancer Magnetic Resonance Imaging (MRI) segmentation is challenging due to a limited amount of training data available and large inter- and intra- patient shape variation for OARs. The proposed Masked-Net consists of a masked encoder within the 3D U-Net to account for the large shape variation within the dataset, with additional dilated layers added to improve segmentation performance. A new loss function was introduced to consider the bounding box loss during training with the proposed Masked-Net. Transfer learning from a male pelvis MRI data with a similar field of view was included. The approaches were compared to the 3D U-Net which was widely used in MRI image segmentation. The data used consisted of 52 volumes obtained from 23 patients with stage IB to IVB cervical cancer across a maximum of 7 weeks of RT with manually contoured labels including the bladder, cervix, gross tumour volume, uterus and rectum. The model was trained and tested with a 5-fold cross validation. Outcomes were evaluated based on the Dice Similarity Coefficients (DSC), the Hausdorff Distance (HD) and the Mean Surface Distance (MSD). The proposed method accounted for the small dataset, large variations in OAR shape and tumour sizes with an average DSC, HD and MSD for all anatomical structures of 0.790, 30.19mm and 3.15mm respectively.

13.
Med Image Anal ; 95: 103156, 2024 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-38603844

RESUMO

The state-of-the-art multi-organ CT segmentation relies on deep learning models, which only generalize when trained on large samples of carefully curated data. However, it is challenging to train a single model that can segment all organs and types of tumors since most large datasets are partially labeled or are acquired across multiple institutes that may differ in their acquisitions. A possible solution is Federated learning, which is often used to train models on multi-institutional datasets where the data is not shared across sites. However, predictions of federated learning can be unreliable after the model is locally updated at sites due to 'catastrophic forgetting'. Here, we address this issue by using knowledge distillation (KD) so that the local training is regularized with the knowledge of a global model and pre-trained organ-specific segmentation models. We implement the models in a multi-head U-Net architecture that learns a shared embedding space for different organ segmentation, thereby obtaining multi-organ predictions without repeated processes. We evaluate the proposed method using 8 publicly available abdominal CT datasets of 7 different organs. Of those datasets, 889 CTs were used for training, 233 for internal testing, and 30 volumes for external testing. Experimental results verified that our proposed method substantially outperforms other state-of-the-art methods in terms of accuracy, inference time, and the number of parameters.


Assuntos
Aprendizado Profundo , Tomografia Computadorizada por Raios X , Humanos , Conjuntos de Dados como Assunto , Bases de Dados Factuais
14.
Sci Rep ; 14(1): 9784, 2024 04 29.
Artigo em Inglês | MEDLINE | ID: mdl-38684904

RESUMO

Accurate multi-organ segmentation in 3D CT images is imperative for enhancing computer-aided diagnosis and radiotherapy planning. However, current deep learning-based methods for 3D multi-organ segmentation face challenges such as the need for labor-intensive manual pixel-level annotations and high hardware resource demands, especially regarding GPU resources. To address these issues, we propose a 3D proxy-bridged region-growing framework specifically designed for the segmentation of the liver and spleen. Specifically, a key slice is selected from each 3D volume according to the corresponding intensity histogram. Subsequently, a deep learning model is employed to pinpoint the semantic central patch on this key slice, to calculate the growing seed. To counteract the impact of noise, segmentation of the liver and spleen is conducted on superpixel images created through proxy-bridging strategy. The segmentation process is then extended to adjacent slices by applying the same methodology iteratively, culminating in the comprehensive segmentation results. Experimental results demonstrate that the proposed framework accomplishes segmentation of the liver and spleen with an average Dice Similarity Coefficient of approximately 0.93 and a Jaccard Similarity Coefficient of around 0.88. These outcomes substantiate the framework's capability to achieve performance on par with that of deep learning methods, albeit requiring less guidance information and lower GPU resources.


Assuntos
Aprendizado Profundo , Imageamento Tridimensional , Fígado , Baço , Tomografia Computadorizada por Raios X , Fígado/diagnóstico por imagem , Baço/diagnóstico por imagem , Baço/anatomia & histologia , Humanos , Imageamento Tridimensional/métodos , Tomografia Computadorizada por Raios X/métodos , Processamento de Imagem Assistida por Computador/métodos , Algoritmos
15.
Bioengineering (Basel) ; 11(4)2024 Mar 27.
Artigo em Inglês | MEDLINE | ID: mdl-38671742

RESUMO

Organ segmentation from CT images is critical in the early diagnosis of diseases, progress monitoring, pre-operative planning, radiation therapy planning, and CT dose estimation. However, data limitation remains one of the main challenges in medical image segmentation tasks. This challenge is particularly huge in pediatric CT segmentation due to children's heightened sensitivity to radiation. In order to address this issue, we propose a novel segmentation framework with a built-in auxiliary classifier generative adversarial network (ACGAN) that conditions age, simultaneously generating additional features during training. The proposed conditional feature generation segmentation network (CFG-SegNet) was trained on a single loss function and used 2.5D segmentation batches. Our experiment was performed on a dataset with 359 subjects (180 male and 179 female) aged from 5 days to 16 years and a mean age of 7 years. CFG-SegNet achieved an average segmentation accuracy of 0.681 dice similarity coefficient (DSC) on the prostate, 0.619 DSC on the uterus, 0.912 DSC on the liver, and 0.832 DSC on the heart with four-fold cross-validation. We compared the segmentation accuracy of our proposed method with previously published U-Net results, and our network improved the segmentation accuracy by 2.7%, 2.6%, 2.8%, and 3.4% for the prostate, uterus, liver, and heart, respectively. The results indicate that our high-performing segmentation framework can more precisely segment organs when limited training images are available.

16.
Comput Biol Med ; 172: 108261, 2024 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-38508056

RESUMO

Whole heart segmentation (WHS) has significant clinical value for cardiac anatomy, modeling, and analysis of cardiac function. This study aims to address the WHS accuracy on cardiac CT images, as well as the fast inference speed and low graphics processing unit (GPU) memory consumption required by practical clinical applications. Thus, we propose a multi-residual two-dimensional (2D) network integrating spatial correlation for WHS. The network performs slice-by-slice segmentation on three-dimensional cardiac CT images in a 2D encoder-decoder manner. In the network, a convolutional long short-term memory skip connection module is designed to perform spatial correlation feature extraction on the feature maps at different resolutions extracted by the sub-modules of the pre-trained ResNet-based encoder. Moreover, a decoder based on the multi-residual module is designed to analyze the extracted features from the perspectives of multi-scale and channel attention, thereby accurately delineating the various substructures of the heart. The proposed method is verified on a dataset of the multi-modality WHS challenge, an in-house WHS dataset, and a dataset of the abdominal organ segmentation challenge. The dice, Jaccard, average symmetric surface distance, Hausdorff distance, inference time, and maximum GPU memory of the WHS are 0.914, 0.843, 1.066 mm, 15.778 mm, 9.535 s, and 1905 MB, respectively. The proposed network has high accuracy, fast inference speed, minimal GPU memory consumption, strong robustness, and good generalization. It can be deployed to clinical practical applications for WHS and can be effectively extended and applied to other multi-organ segmentation fields. The source code is publicly available at https://github.com/nancy1984yan/MultiResNet-SC.


Assuntos
Coração , Software , Coração/diagnóstico por imagem , Tomografia Computadorizada por Raios X
17.
Phys Med Biol ; 69(11)2024 May 14.
Artigo em Inglês | MEDLINE | ID: mdl-38479023

RESUMO

Precise delineation of multiple organs or abnormal regions in the human body from medical images plays an essential role in computer-aided diagnosis, surgical simulation, image-guided interventions, and especially in radiotherapy treatment planning. Thus, it is of great significance to explore automatic segmentation approaches, among which deep learning-based approaches have evolved rapidly and witnessed remarkable progress in multi-organ segmentation. However, obtaining an appropriately sized and fine-grained annotated dataset of multiple organs is extremely hard and expensive. Such scarce annotation limits the development of high-performance multi-organ segmentation models but promotes many annotation-efficient learning paradigms. Among these, studies on transfer learning leveraging external datasets, semi-supervised learning including unannotated datasets and partially-supervised learning integrating partially-labeled datasets have led the dominant way to break such dilemmas in multi-organ segmentation. We first review the fully supervised method, then present a comprehensive and systematic elaboration of the 3 abovementioned learning paradigms in the context of multi-organ segmentation from both technical and methodological perspectives, and finally summarize their challenges and future trends.


Assuntos
Processamento de Imagem Assistida por Computador , Humanos , Processamento de Imagem Assistida por Computador/métodos , Aprendizado Profundo , Aprendizado de Máquina
18.
Radiography (Lond) ; 30(2): 673-680, 2024 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-38364707

RESUMO

INTRODUCTION: This paper presents a novel approach to automate the segmentation of Organ-at-Risk (OAR) in Head and Neck cancer patients using Deep Learning models combined with Ensemble Learning techniques. The study aims to improve the accuracy and efficiency of OAR segmentation, essential for radiotherapy treatment planning. METHODS: The dataset comprised computed tomography (CT) scans of 182 patients in DICOM format, obtained from an institutional image bank. Experienced Radiation Oncologists manually segmented seven OARs for each scan. Two models, 3D U-Net and 3D DenseNet-FCN, were trained on reduced CT scans (192 × 192 x 128) due to memory limitations. Ensemble Learning techniques were employed to enhance accuracy and segmentation metrics. Testing was conducted on 78 patients from the institutional dataset and an open-source dataset (TCGA-HNSC and Head-Neck Cetuximab) consisting of 31 patient scans. RESULTS: Using the Ensemble Learning technique, the average dice similarity coefficient for OARs ranged from 0.990 to 0.994, indicating high segmentation accuracy. The 95% Hausdorff distance (mm) ranged from 1.3 to 2.1, demonstrating precise segmentation boundaries. CONCLUSION: The proposed automated segmentation method achieved efficient and accurate OAR segmentation, surpassing human expert performance in terms of time and accuracy. IMPLICATIONS FOR PRACTICE: This approach has implications for improving treatment planning and patient care in radiotherapy. By reducing manual segmentation reliance, the proposed method offers significant time savings and potential improvements in treatment planning efficiency and precision for head and neck cancer patients.


Assuntos
Neoplasias de Cabeça e Pescoço , Órgãos em Risco , Humanos , Órgãos em Risco/diagnóstico por imagem , Neoplasias de Cabeça e Pescoço/diagnóstico por imagem , Neoplasias de Cabeça e Pescoço/radioterapia , Tomografia Computadorizada por Raios X , Planejamento da Radioterapia Assistida por Computador/métodos , Aprendizado de Máquina
19.
Angiology ; : 33197231225286, 2024 Jan 02.
Artigo em Inglês | MEDLINE | ID: mdl-38166442

RESUMO

To evaluate deep learning-based calcium segmentation and quantification on ECG-gated cardiac CT scans compared with manual evaluation. Automated calcium quantification was performed using a neural network based on mask regions with convolutional neural networks (R-CNNs) for multi-organ segmentation. Manual evaluation of calcium was carried out using proprietary software. This is a retrospective study of archived data. This study used 40 patients to train the segmentation model and 110 patients were used for the validation of the algorithm. The Pearson correlation coefficient between the reference actual and the computed predictive scores shows high level of correlation (0.84; P < .001) and high limits of agreement (±1.96 SD; -2000, 2000) in Bland-Altman plot analysis. The proposed method correctly classifies the risk group in 75.2% and classifies the subjects in the same group. In total, 81% of the predictive scores lie in the same categories and only seven patients out of 110 were more than one category off. For the presence/absence of coronary artery calcifications, the deep learning model achieved a sensitivity of 90% and a specificity of 94%. Fully automated model shows good correlation compared with reference standards. Automating process reduces evaluation time and optimizes clinical calcium scoring without additional resources.

20.
Nan Fang Yi Ke Da Xue Xue Bao ; 44(1): 83-92, 2024 Jan 20.
Artigo em Chinês | MEDLINE | ID: mdl-38293979

RESUMO

OBJECTIVE: To propose a method for abdominal multi-organ segmentation assisted by multi-phase CT synthesis. METHODS: Multi-phase CT synthesis for synthesizing high-quality CT images was used to increase the information details for image segmentation. A transformer block was introduced to help to capture long-range semantic information in cooperation with perceptual loss to minimize the differences between the real image and synthesized image. RESULTS: The model was trained using multi-phase CT dataset of 526 total cases from Nanfang Hospital. The mean maximum absolute error (MAE) of the synthesized non-contrast CT, venous phase contrast- enhanced CT (CECT), and delay phase CECT images from arterial phase CECT was 19.192±3.381, 20.140±2.676 and 22.538±2.874, respectively, which were better than those of images synthesized using other methods. Validation of the multi-phase CT synthesis-assisted abdominal multi-organ segmentation method showed an average dice coefficient of 0.847 for the internal validation set and 0.823 for the external validation set. CONCLUSION: The propose method is capable of synthesizing high-quality multi-phase CT images to effectively reduce the errors in registration between different phase CT images and improve the performance for segmentation of 13 abdominal organs.


Assuntos
Processamento de Imagem Assistida por Computador , Tomografia Computadorizada por Raios X , Tomografia Computadorizada por Raios X/métodos , Processamento de Imagem Assistida por Computador/métodos
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA