Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 12 de 12
Filtrar
1.
Comput Assist Surg (Abingdon) ; 29(1): 2329675, 2024 12.
Artigo em Inglês | MEDLINE | ID: mdl-38504595

RESUMO

The real-time requirement for image segmentation in laparoscopic surgical assistance systems is extremely high. Although traditional deep learning models can ensure high segmentation accuracy, they suffer from a large computational burden. In the practical setting of most hospitals, where powerful computing resources are lacking, these models cannot meet the real-time computational demands. We propose a novel network SwinD-Net based on Skip connections, incorporating Depthwise separable convolutions and Swin Transformer Blocks. To reduce computational overhead, we eliminate the skip connection in the first layer and reduce the number of channels in shallow feature maps. Additionally, we introduce Swin Transformer Blocks, which have a larger computational and parameter footprint, to extract global information and capture high-level semantic features. Through these modifications, our network achieves desirable performance while maintaining a lightweight design. We conduct experiments on the CholecSeg8k dataset to validate the effectiveness of our approach. Compared to other models, our approach achieves high accuracy while significantly reducing computational and parameter overhead. Specifically, our model requires only 98.82 M floating-point operations (FLOPs) and 0.52 M parameters, with an inference time of 47.49 ms per image on a CPU. Compared to the recently proposed lightweight segmentation network UNeXt, our model not only outperforms it in terms of the Dice metric but also has only 1/3 of the parameters and 1/22 of the FLOPs. In addition, our model achieves a 2.4 times faster inference speed than UNeXt, demonstrating comprehensive improvements in both accuracy and speed. Our model effectively reduces parameter count and computational complexity, improving the inference speed while maintaining comparable accuracy. The source code will be available at https://github.com/ouyangshuiming/SwinDNet.


Assuntos
Laparoscopia , Fígado , Humanos , Fígado/diagnóstico por imagem , Fígado/cirurgia , Software
2.
IEEE J Biomed Health Inform ; 27(10): 4983-4994, 2023 10.
Artigo em Inglês | MEDLINE | ID: mdl-37498758

RESUMO

Surgical action triplet recognition plays a significant role in helping surgeons facilitate scene analysis and decision-making in computer-assisted surgeries. Compared to traditional context-aware tasks such as phase recognition, surgical action triplets, comprising the instrument, verb, and target, can offer more comprehensive and detailed information. However, current triplet recognition methods fall short in distinguishing the fine-grained subclasses and disregard temporal correlation in action triplets. In this article, we propose a multi-task fine-grained spatial-temporal framework for surgical action triplet recognition named MT-FiST. The proposed method utilizes a multi-label mutual channel loss, which consists of diversity and discriminative components. This loss function decouples global task features into class-aligned features, enabling the learning of more local details from the surgical scene. The proposed framework utilizes partial shared-parameters LSTM units to capture temporal correlations between adjacent frames. We conducted experiments on the CholecT50 dataset proposed in the MICCAI 2021 Surgical Action Triplet Recognition Challenge. Our framework is evaluated on the private test set of the challenge to ensure fair comparisons. Our model apparently outperformed state-of-the-art models in instrument, verb, target, and action triplet recognition tasks, with mAPs of 82.1% (+4.6%), 51.5% (+4.0%), 45.50% (+7.8%), and 35.8% (+3.1%), respectively. The proposed MT-FiST boosts the recognition of surgical action triplets in a context-aware surgical assistant system, further solving multi-task recognition by effective temporal aggregation and fine-grained features.


Assuntos
Cirurgia Assistida por Computador , Humanos
3.
Med Phys ; 50(10): 6243-6258, 2023 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-36975007

RESUMO

BACKGROUND: The fusion of computed tomography (CT) and ultrasound (US) image can enhance lesion detection ability and improve the success rate of liver interventional radiology. The image-based fusion methods encounter the challenge of registration initialization due to the random scanning pose and limited field of view of US. Existing automatic methods those used vessel geometric information and intensity-based metric are sensitive to parameters and have low success rate. The learning-based methods require a large number of registered datasets for training. PURPOSE: The aim of this study is to provide a fully automatic and robust US-3D CT registration method without registered training data and user-specified parameters assisted by the revolutionary deep learning-based segmentation, which can further be used for preparing training samples for the study of learning-based methods. METHODS: We propose a fully automatic CT-3D US registration method by two improved registration metrics. We propose to use 3D U-Net-based multi-organ segmentation of US and CT to assist the conventional registration. The rigid transform is searched in the space of any paired vessel bifurcation planes where the best transform is decided by a segmentation overlap metric, which is more related to the segmentation precision than Dice coefficient. In nonrigid registration phase, we propose a hybrid context and edge based image similarity metric with a simple mask that can remove most noisy US voxels to guide the B-spline transform registration. We evaluate our method on 42 paired CT-3D US datasets scanned with two different US devices from two hospitals. We compared our methods with other exsiting methods with both quantitative measures of target registration error (TRE) and the Jacobian determinent with paired t-test and qualitative registration imaging results. RESULTS: The results show that our method achieves fully automatic rigid registration TRE of 4.895 mm, deformable registration TRE of 2.995 mm in average, which outperforms state-of-the-art automatic linear methods and nonlinear registration metrics with paired t-test's p value less than 0.05. The proposed overlap metric achieves better results than self similarity description (SSD), edge matching (EM), and block matching (BM) with p values of 1.624E-10, 4.235E-9, and 0.002, respectively. The proposed hybrid edge and context-based metric outperforms context-only, edge-only, and intensity statistics-only-based metrics with p values of 0.023, 3.81E-5, and 1.38E-15, respectively. The 3D US segmentation has achieved mean Dice similarity coefficient (DSC) of 0.799, 0.724, 0.788, and precision of 0.871, 0.769, 0.862 for gallbladder, vessel, and branch vessel, respectively. CONCLUSIONS: The deep learning-based US segmentation can achieve satisfied result to assist robust conventional rigid registration. The Dice similarity coefficient-based metrics, hybrid context, and edge image similarity metric contribute to robust and accurate registration.


Assuntos
Imageamento Tridimensional , Fígado , Imageamento Tridimensional/métodos , Ultrassonografia/métodos , Fígado/diagnóstico por imagem , Tomografia Computadorizada por Raios X/métodos , Processamento de Imagem Assistida por Computador/métodos
4.
Int J Comput Assist Radiol Surg ; 18(8): 1521-1531, 2023 Aug.
Artigo em Inglês | MEDLINE | ID: mdl-36787037

RESUMO

PURPOSE: Laparoscopic liver resection is a minimal invasive surgery. Augmented reality can map preoperative anatomy information extracted from computed tomography to the intraoperative liver surface reconstructed from stereo 3D laparoscopy. However, liver surface registration is particularly challenging as the intraoperative surface is only partially visible and suffers from large liver deformations due to pneumoperitoneum. This study proposes a deep learning-based robust point cloud registration network. METHODS: This study proposed a low overlap liver surface registration algorithm combining local mixed features and global features of point clouds. A learned overlap mask is used to filter the non-overlapping region of the point cloud, and a network is used to predict the overlapping region threshold to regulate the training process. RESULTS: We validated the algorithm on the DePoLL (the Deformable Porcine Laparoscopic Liver) dataset. Compared with the baseline method and other state-of-the-art registration methods, our method achieves minimum target registration error (TRE) of 19.9 ± 2.7 mm. CONCLUSION: The proposed point cloud registration method uses the learned overlapping mask to filter the non-overlapping areas in the point cloud, then the extracted overlapping area point cloud is registered according to the mixed features and global features, and this method is robust and efficient in low-overlap liver surface registration.


Assuntos
Laparoscopia , Cirurgia Assistida por Computador , Animais , Algoritmos , Laparoscopia/métodos , Fígado/diagnóstico por imagem , Fígado/cirurgia , Cirurgia Assistida por Computador/métodos , Suínos , Tomografia Computadorizada por Raios X/métodos
5.
Comput Biol Med ; 140: 105109, 2021 Dec 03.
Artigo em Inglês | MEDLINE | ID: mdl-34891097

RESUMO

BACKGROUND: Learning-based methods have achieved remarkable performances on depth estimation. However, the premise of most self-learning and unsupervised learning methods is built on rigorous, geometrically-aligned stereo rectification. The performances of these methods degrade when the rectification is not accurate. Therefore, we explore an approach for unsupervised depth estimation from stereo images that can handle imperfect camera parameters. METHODS: We propose an unsupervised deep convolutional network that takes rectified stereo image pairs as input and outputs corresponding dense disparity maps. First, a new vertical correction module is designed for predicting a correction map to compensate for the imperfect geometry alignment. Second, the left and right images, which are reconstructed based on the input image pair and corresponding disparities as well as the vertical correction maps, are regarded as the outputs of the generative term of the generative adversarial network (GAN). Then, the discriminator term of the GAN is used to distinguish the reconstructed images from the original inputs to force the generator to output increasingly realistic images. In addition, a residual mask is introduced to exclude pixels that conflict with the appearance of the original image in the loss calculation. RESULTS: The proposed model is validated on the publicly available Stereo Correspondence and Reconstruction of Endoscopic Data (SCARED) dataset and the average MAE is 3.054 mm. CONCLUSION: Our model can effectively handle imperfect rectified stereo images for depth estimation.

6.
BMC Med Imaging ; 21(1): 178, 2021 11 24.
Artigo em Inglês | MEDLINE | ID: mdl-34819022

RESUMO

BACKGROUND: Most existing algorithms have been focused on the segmentation from several public Liver CT datasets scanned regularly (no pneumoperitoneum and horizontal supine position). This study primarily segmented datasets with unconventional liver shapes and intensities deduced by contrast phases, irregular scanning conditions, different scanning objects of pigs and patients with large pathological tumors, which formed the multiple heterogeneity of datasets used in this study. METHODS: The multiple heterogeneous datasets used in this paper includes: (1) One public contrast-enhanced CT dataset and one public non-contrast CT dataset; (2) A contrast-enhanced dataset that has abnormal liver shape with very long left liver lobes and large-sized liver tumors with abnormal presets deduced by microvascular invasion; (3) One artificial pneumoperitoneum dataset under the pneumoperitoneum and three scanning profiles (horizontal/left/right recumbent position); (4) Two porcine datasets of Bama type and domestic type that contains pneumoperitoneum cases but with large anatomy discrepancy with humans. The study aimed to investigate the segmentation performances of 3D U-Net in: (1) generalization ability between multiple heterogeneous datasets by cross-testing experiments; (2) the compatibility when hybrid training all datasets in different sampling and encoder layer sharing schema. We further investigated the compatibility of encoder level by setting separate level for each dataset (i.e., dataset-wise convolutions) while sharing the decoder. RESULTS: Model trained on different datasets has different segmentation performance. The prediction accuracy between LiTS dataset and Zhujiang dataset was about 0.955 and 0.958 which shows their good generalization ability due to that they were all contrast-enhanced clinical patient datasets scanned regularly. For the datasets scanned under pneumoperitoneum, their corresponding datasets scanned without pneumoperitoneum showed good generalization ability. Dataset-wise convolution module in high-level can improve the dataset unbalance problem. The experimental results will facilitate researchers making solutions when segmenting those special datasets. CONCLUSIONS: (1) Regularly scanned datasets is well generalized to irregularly ones. (2) The hybrid training is beneficial but the dataset imbalance problem always exits due to the multi-domain homogeneity. The higher levels encoded more domain specific information than lower levels and thus were less compatible in terms of our datasets.


Assuntos
Imageamento Tridimensional , Neoplasias Hepáticas/diagnóstico por imagem , Fígado/diagnóstico por imagem , Aprendizado de Máquina , Interpretação de Imagem Radiográfica Assistida por Computador/métodos , Tomografia Computadorizada por Raios X/métodos , Animais , Meios de Contraste , Conjuntos de Dados como Assunto , Humanos , Pneumoperitônio/diagnóstico por imagem , Suínos
7.
Surg Laparosc Endosc Percutan Tech ; 31(6): 679-684, 2021 Aug 23.
Artigo em Inglês | MEDLINE | ID: mdl-34420005

RESUMO

BACKGROUND: Clinically, the total and residual liver volume must be accurately calculated before major hepatectomy. However, liver volume might be influenced by pneumoperitoneum during surgery. Changes in liver volume change also affect the accuracy of simulation and augmented reality navigation systems, which are commonly first validated in animal models. In this study, the morphologic changes in porcine livers in vivo under 13 mm Hg pneumoperitoneum pressure were investigated. MATERIALS AND METHODS: Twenty male pigs were scanned with contrast-enhanced computed tomography without pneumoperitoneum and with 13 mm Hg pneumoperitoneum pressure. RESULTS: The surface area and volume of the liver and the vascular diameter of the aortic lumen, inferior vena cava lumen, and portal vein lumen were measured. There were statistically significant differences in the surface area and volume of the liver (P=0.000), transverse diameter of the portal vein (P=0.038), longitudinal diameter of the inferior vena cava (P=0.033), longitudinal diameter of the portal vein (P=0.036), vascular cross-sectional area of the inferior vena cava (P=0.028), and portal vein (P=0.038) before and after 13 mm Hg pneumoperitoneum pressure. CONCLUSIONS: This study indicated that the creation of pneumoperitoneum at 13 mm Hg pressure in a porcine causes liver morphologic alterations affecting the area and volume, as well as the diameter of a blood vessel.


Assuntos
Pneumoperitônio , Abdome , Animais , Fígado/diagnóstico por imagem , Masculino , Veia Porta/diagnóstico por imagem , Suínos , Veia Cava Inferior/diagnóstico por imagem
8.
Comput Methods Programs Biomed ; 187: 105099, 2020 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-31601442

RESUMO

OBJECTIVE: Understanding the three-dimensional (3D) spatial position and orientation of vessels and tumor(s) is vital in laparoscopic liver resection procedures. Augmented reality (AR) techniques can help surgeons see the patient's internal anatomy in conjunction with laparoscopic video images. METHOD: In this paper, we present an AR-assisted navigation system for liver resection based on a rigid stereoscopic laparoscope. The stereo image pairs from the laparoscope are used by an unsupervised convolutional network (CNN) framework to estimate depth and generate an intraoperative 3D liver surface. Meanwhile, 3D models of the patient's surgical field are segmented from preoperative CT images using V-Net architecture for volumetric image data in an end-to-end predictive style. A globally optimal iterative closest point (Go-ICP) algorithm is adopted to register the pre- and intraoperative models into a unified coordinate space; then, the preoperative 3D models are superimposed on the live laparoscopic images to provide the surgeon with detailed information about the subsurface of the patient's anatomy, including tumors, their resection margins and vessels. RESULTS: The proposed navigation system is tested on four laboratory ex vivo porcine livers and five operating theatre in vivo porcine experiments to validate its accuracy. The ex vivo and in vivo reprojection errors (RPE) are 6.04 ±â€¯1.85 mm and 8.73 ±â€¯2.43 mm, respectively. CONCLUSION AND SIGNIFICANCE: Both the qualitative and quantitative results indicate that our AR-assisted navigation system shows promise and has the potential to be highly useful in clinical practice.


Assuntos
Realidade Aumentada , Laparoscopia/métodos , Fígado/diagnóstico por imagem , Fígado/cirurgia , Algoritmos , Animais , Aprendizado Profundo , Percepção de Profundidade , Modelos Animais de Doenças , Processamento de Imagem Assistida por Computador , Imageamento Tridimensional , Laparoscópios , Neoplasias/diagnóstico por imagem , Reprodutibilidade dos Testes , Software , Cirurgia Assistida por Computador , Suínos , Tomografia Computadorizada por Raios X , Gravação em Vídeo
9.
Surg Endosc ; 34(8): 3449-3459, 2020 08.
Artigo em Inglês | MEDLINE | ID: mdl-31705286

RESUMO

BACKGROUND: Understanding the internal anatomy of the liver remains a major challenge in anatomical liver resection. Although virtual hepatectomy and indocyanine green (ICG) fluorescence imaging techniques have been widely used in hepatobiliary surgery, limitations in their application for real-time navigation persist. OBJECTIVE: The aim of the present study was to evaluate the feasibility and clinical utility of the novel laparoscopic hepatectomy navigation system (LHNS), which fuses preoperative three-dimensional (3D) models with ICG fluorescence imaging to achieve real-time surgical navigation. METHODS: We conducted a retrospective review of clinical outcome for 64 patients who underwent laparoscopic hepatectomy from January 2018 to December 2018, including 30 patients who underwent the procedure using the LHNS (LHNS group) and 34 patients who underwent the procedure without LHNS guidance (Non-LHNS group). RESULTS: There was no significant difference in preoperative characteristics between the two groups. The LHNS group had a significantly less blood loss (285.0 ± 163.0 mL vs. 391.1 ± 242.0 mL; P = 0.047), less intraoperative blood transfusion rate (13.3% vs. 38.2%; P = 0.045), and shorter postoperative hospital stay (7.8 ± 2.1 days vs. 10.6 ± 3.8 days; P < 0.001) than the Non-LHNS group. There was no statistical difference in operative time and the overall complication rate between the two groups. The liver transection line was clearly delineated by the LHNS in 27 patients; however, the projection of boundary was unclear in 2 cases, and in 1 case, the boundary was not clearly displayed by ICG fluorescence imaging. CONCLUSIONS: We developed the LHNS to address limitations of current intraoperative imaging systems. The LHNS is hopefully to become a promising real-time navigation system for laparoscopic hepatectomy.


Assuntos
Carcinoma Hepatocelular/diagnóstico por imagem , Hepatectomia/métodos , Laparoscopia/métodos , Neoplasias Hepáticas/diagnóstico por imagem , Imagem Óptica/métodos , Cirurgia Assistida por Computador/métodos , Sistemas de Navegação Cirúrgica , Adulto , Idoso , Perda Sanguínea Cirúrgica , Transfusão de Sangue , Carcinoma Hepatocelular/cirurgia , Estudos de Viabilidade , Feminino , Fluorescência , Humanos , Imageamento Tridimensional , Verde de Indocianina/uso terapêutico , Tempo de Internação , Neoplasias Hepáticas/cirurgia , Masculino , Pessoa de Meia-Idade , Duração da Cirurgia , Complicações Pós-Operatórias , Estudos Retrospectivos
10.
Healthc Technol Lett ; 6(6): 154-158, 2019 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-32038849

RESUMO

Depth estimation plays an important role in vision-based laparoscope surgical navigation systems. Most learning-based depth estimation methods require ground truth depth or disparity images for training; however, these data are difficult to obtain in laparoscopy. The authors present an unsupervised learning depth estimation approach by fusing traditional stereo knowledge. The traditional stereo method is used to generate proxy disparity labels, in which unreliable depth measurements are removed via a confidence measure to improve stereo accuracy. The disparity images are generated by training a dual encoder-decoder convolutional neural network from rectified stereo images coupled with proxy labels generated by the traditional stereo method. A principled mask is computed to exclude the pixels, which are not seen in one of views due to parallax effects from the calculation of loss function. Moreover, the neighbourhood smoothness term is employed to constrain neighbouring pixels with similar appearances to generate a smooth depth surface. This approach can make the depth of the projected point cloud closer to the real surgical site and preserve realistic details. The authors demonstrate the performance of the method by training and evaluation with a partial nephrectomy da Vinci surgery dataset and heart phantom data from the Hamlyn Centre.

11.
Abdom Radiol (NY) ; 42(7): 1993-2000, 2017 07.
Artigo em Inglês | MEDLINE | ID: mdl-28217826

RESUMO

PURPOSE: To compare the accuracy of a Kinect-Optical navigation system with an electromagnetic (EM) navigation system for percutaneous liver needle intervention. MATERIALS AND METHODS: Five beagles with nine artificial tumors were used for validation. The Veran IG4 EM navigation system and a custom-made Kinect-Optical navigation system were used. Needle insertions into each tumor were conducted with these two guidance methods. The target positioning error (TPE) and the time cost of the puncture procedures were evaluated. RESULTS: A total of 18 needle insertions were performed to evaluate the navigation accuracy of the two guidance approaches. The targeting error was 6.78 ± 3.22 mm and 8.72 ± 3.5 mm for the Kinect-Optical navigation system and the EM navigation system, respectively. There is no statistically significant difference in the TPE between the Kinect-Optical navigation system and the EM navigation system (p = 0.229). The processing time with the Kinect-Optical system (10 min) is similar to that of the Veran IG4 system (12 min). CONCLUSIONS: The accuracy of the Kinect-Optical navigation system is comparable to that of the EM navigation system.


Assuntos
Técnicas de Ablação/instrumentação , Neoplasias Hepáticas Experimentais/cirurgia , Imagem Óptica/instrumentação , Tomografia Computadorizada por Raios X , Animais , Cães , Campos Eletromagnéticos , Agulhas , Punções , Cirurgia Assistida por Computador
12.
Phys Med Biol ; 61(15): 5687-705, 2016 08 07.
Artigo em Inglês | MEDLINE | ID: mdl-27401131

RESUMO

Percutaneous abdominal puncture is a popular interventional method for the management of abdominal tumors. Image-guided puncture can help interventional radiologists improve targeting accuracy. The second generation of Kinect(™) was released recently, we developed an optical navigation system to investigate its feasibility for guiding percutaneous abdominal puncture, and compare its performance on needle insertion guidance with that of the first-generation Kinect(™). For physical-to-image registration in this system, two surfaces extracted from preoperative CT and intraoperative Kinect(™) depth images were matched using an iterative closest point (ICP) algorithm. A 2D shape image-based correspondence searching algorithm was proposed for generating a close initial position before ICP matching. Evaluation experiments were conducted on an abdominal phantom and six beagles in vivo. For phantom study, a two-factor experiment was designed to evaluate the effect of the operator's skill and trajectory on target positioning error (TPE). A total of 36 needle punctures were tested on a Kinect(™) for Windows version 2 (Kinect(™) V2). The target registration error (TRE), user error, and TPE are 4.26 ± 1.94 mm, 2.92 ± 1.67 mm, and 5.23 ± 2.29 mm, respectively. No statistically significant differences in TPE regarding operator's skill and trajectory are observed. Additionally, a Kinect(™) for Windows version 1 (Kinect(™) V1) was tested with 12 insertions, and the TRE evaluated with the Kinect(™) V1 is statistically significantly larger than that with the Kinect(™) V2. For the animal experiment, fifteen artificial liver tumors were inserted guided by the navigation system. The TPE was evaluated as 6.40 ± 2.72 mm, and its lateral and longitudinal component were 4.30 ± 2.51 mm and 3.80 ± 3.11 mm, respectively. This study demonstrates that the navigation accuracy of the proposed system is acceptable, and that the second generation Kinect(™)-based navigation is superior to the first-generation Kinect(™), and has potential of clinical application in percutaneous abdominal puncture.


Assuntos
Neoplasias Abdominais/radioterapia , Dispositivos Ópticos , Punções/métodos , Radiocirurgia/instrumentação , Algoritmos , Animais , Cães , Humanos , Imagens de Fantasmas , Radiocirurgia/métodos
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA