Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 20 de 33
Filtrar
Mais filtros










Base de dados
Intervalo de ano de publicação
1.
Artigo em Inglês | MEDLINE | ID: mdl-38051611

RESUMO

Emotion is a complex physiological and psychological activity, accompanied by subjective physiological sensations and objective physiological changes. The body sensation map describes the changes in body sensation associated with emotion in a topographic manner, but it relies on subjective evaluations from participants. Physiological signals are a more reliable measure of emotion, but most research focuses on the central nervous system, neglecting the importance of the peripheral nervous system. In this study, a body surface potential mapping (BSPM) system was constructed, and an experiment was designed to induce emotions and obtain high-density body surface potential information under negative and non-negative emotions. Then, by constructing and analyzing the functional connectivity network of BSPs, the high-density electrophysiological characteristics are obtained and visualized as bodily emotion maps. The results showed that the functional connectivity network of BSPs under negative emotions had denser connections, and emotion maps based on local clustering coefficient (LCC) are consistent with BSMs under negative emotions. in addition, our features can classify negative and non-negative emotions with the highest classification accuracy of 80.77%. In conclusion, this study constructs an emotion map based on high-density BSPs, which offers a novel approach to psychophysiological computing.

2.
Artigo em Inglês | MEDLINE | ID: mdl-38117619

RESUMO

Minimally invasive surgery, which relies on surgical robots and microscopes, demands precise image segmentation to ensure safe and efficient procedures. Nevertheless, achieving accurate segmentation of surgical instruments remains challenging due to the complexity of the surgical environment. To tackle this issue, this paper introduces a novel multiscale dual-encoding segmentation network, termed MSDE-Net, designed to automatically and precisely segment surgical instruments. The proposed MSDE-Net leverages a dual-branch encoder comprising a convolutional neural network (CNN) branch and a transformer branch to effectively extract both local and global features. Moreover, an attention fusion block (AFB) is introduced to ensure effective information complementarity between the dual-branch encoding paths. Additionally, a multilayer context fusion block (MCF) is proposed to enhance the network's capacity to simultaneously extract global and local features. Finally, to extend the scope of global feature information under larger receptive fields, a multi-receptive field fusion (MRF) block is incorporated. Through comprehensive experimental evaluations on two publicly available datasets for surgical instrument segmentation, the proposed MSDE-Net demonstrates superior performance compared to existing methods.

3.
Med Image Anal ; 86: 102803, 2023 05.
Artigo em Inglês | MEDLINE | ID: mdl-37004378

RESUMO

Context-aware decision support in the operating room can foster surgical safety and efficiency by leveraging real-time feedback from surgical workflow analysis. Most existing works recognize surgical activities at a coarse-grained level, such as phases, steps or events, leaving out fine-grained interaction details about the surgical activity; yet those are needed for more helpful AI assistance in the operating room. Recognizing surgical actions as triplets of combination delivers more comprehensive details about the activities taking place in surgical videos. This paper presents CholecTriplet2021: an endoscopic vision challenge organized at MICCAI 2021 for the recognition of surgical action triplets in laparoscopic videos. The challenge granted private access to the large-scale CholecT50 dataset, which is annotated with action triplet information. In this paper, we present the challenge setup and the assessment of the state-of-the-art deep learning methods proposed by the participants during the challenge. A total of 4 baseline methods from the challenge organizers and 19 new deep learning algorithms from the competing teams are presented to recognize surgical action triplets directly from surgical videos, achieving mean average precision (mAP) ranging from 4.2% to 38.1%. This study also analyzes the significance of the results obtained by the presented approaches, performs a thorough methodological comparison between them, in-depth result analysis, and proposes a novel ensemble method for enhanced recognition. Our analysis shows that surgical workflow analysis is not yet solved, and also highlights interesting directions for future research on fine-grained surgical activity recognition which is of utmost importance for the development of AI in surgery.


Assuntos
Benchmarking , Laparoscopia , Humanos , Algoritmos , Salas Cirúrgicas , Fluxo de Trabalho , Aprendizado Profundo
4.
IEEE Trans Med Imaging ; 42(10): 2924-2935, 2023 10.
Artigo em Inglês | MEDLINE | ID: mdl-37079409

RESUMO

In recent intelligent-robot-assisted surgery studies, an urgent issue is how to detect the motion of instruments and soft tissue accurately from intra-operative images. Although optical flow technology from computer vision is a powerful solution to the motion-tracking problem, it has difficulty obtaining the pixel-wise optical flow ground truth of real surgery videos for supervised learning. Thus, unsupervised learning methods are critical. However, current unsupervised methods face the challenge of heavy occlusion in the surgical scene. This paper proposes a novel unsupervised learning framework to estimate the motion from surgical images under occlusion. The framework consists of a Motion Decoupling Network to estimate the tissue and the instrument motion with different constraints. Notably, the network integrates a segmentation subnet that estimates the segmentation map of instruments in an unsupervised manner to obtain the occlusion region and improve the dual motion estimation. Additionally, a hybrid self-supervised strategy with occlusion completion is introduced to recover realistic vision clues. Extensive experiments on two surgical datasets show that the proposed method achieves accurate motion estimation for intra-operative scenes and outperforms other unsupervised methods, with a margin of 15% in accuracy. The average estimation error for tissue is less than 2.2 pixels on average for both surgical datasets.


Assuntos
Procedimentos Cirúrgicos Robóticos , Cirurgia Assistida por Computador , Algoritmos , Movimento (Física) , Cirurgia Assistida por Computador/métodos
5.
Bioengineering (Basel) ; 10(2)2023 Feb 07.
Artigo em Inglês | MEDLINE | ID: mdl-36829720

RESUMO

BACKGROUND: Medical image processing tasks represented by multi-object segmentation are of great significance for surgical planning, robot-assisted surgery, and surgical safety. However, the exceptionally low contrast among tissues and limited available annotated data makes developing an automatic segmentation algorithm for pelvic CT challenging. METHODS: A bi-direction constrained dual-task consistency model named PICT is proposed to improve segmentation quality by leveraging free unlabeled data. First, to learn more unmarked data features, it encourages the model prediction of the interpolated image to be consistent with the interpolation of the model prediction at the pixel, model, and data levels. Moreover, to constrain the error prediction of interpolation interference, PICT designs an auxiliary pseudo-supervision task that focuses on the underlying information of non-interpolation data. Finally, an effective loss algorithm for both consistency tasks is designed to ensure the complementary manner and produce more reliable predictions. RESULTS: Quantitative experiments show that the proposed PICT achieves 87.18%, 96.42%, and 79.41% mean DSC score on ACDC, CTPelvic1k, and the individual Multi-tissue Pelvis dataset with gains of around 0.8%, 0.5%, and 1% compared to the state-of-the-art semi-supervised method. Compared to the baseline supervised method, the PICT brings over 3-9% improvements. CONCLUSIONS: The developed PICT model can effectively leverage unlabeled data to improve segmentation quality of low contrast medical images. The segmentation result could improve the precision of surgical path planning and provide input for robot-assisted surgery.

6.
Comput Biol Med ; 153: 106531, 2023 02.
Artigo em Inglês | MEDLINE | ID: mdl-36638619

RESUMO

Surgical scene segmentation provides critical information for guidance in micro-neurosurgery. Segmentation of instruments and critical tissues contributes further to robot assisted surgery and surgical evaluation. However, due to the lack of relevant scene segmentation dataset, scale variation and local similarity, micro-neurosurgical segmentation faces many challenges. To address these issues, a high correlative non-local network (HCNNet), is proposed to aggregate multi-scale feature by optimized non-local mechanism. HCNNet adopts two-branch design to generate features of different scale efficiently, while the two branches share common weights in shallow layers. Several short-term dense concatenate (STDC) modules are combined as the backbone to capture both semantic and spatial information. Besides, a high correlative non-local module (HCNM) is designed to guide the upsampling process of the high-level feature by modeling global context generated from the low-level feature. It filters out confused pixels of different classes in the non-local correlation map. Meanwhile, a large segmentation dataset named NeuroSeg is constructed, which contains 15 types of instruments and 3 types of tissues that appear in meningioma resection surgery. The proposed HCNNet achieves the state-of-the-art performance on NeuroSeg, it reaches an inference speed of 54.85 FPS with the highest accuracy of 59.62% mIoU, 74.7% Dice, 70.55% mAcc and 87.12% aAcc.


Assuntos
Procedimentos Cirúrgicos Robóticos , Processamento de Imagem Assistida por Computador , Semântica
7.
IEEE Trans Neural Netw Learn Syst ; 34(12): 9727-9741, 2023 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-35333726

RESUMO

Percutaneous coronary intervention (PCI) has increasingly become the main treatment for coronary artery disease. The procedure requires high experienced skills and dexterous manipulations. However, there are few techniques to model PCI skill so far. In this study, a learning framework with local and ensemble learning is proposed to learn skill characteristics of different skill-level subjects from their PCI manipulations. Ten interventional cardiologists (four experts and six novices) were recruited to deliver a medical guidewire to two target arteries on a porcine model for in vivo studies. Simultaneously, translation and twist manipulations of thumb, forefinger, and wrist are acquired with electromagnetic (EM) and fiber-optic bend (FOB) sensors, respectively. These behavior data are then processed with wavelet packet decomposition (WPD) under 1-10 levels for feature extraction. The feature vectors are further fed into three candidate individual classifiers in the local learning layer. Furthermore, the local learning results from different manipulation behaviors are fused in the ensemble learning layer with three rule-based ensemble learning algorithms. In subject-dependent skill characteristics learning, the ensemble learning can achieve 100% accuracy, significantly outperforming the best local result (90%). Furthermore, ensemble learning can also maintain 73% accuracy in subject-independent schemes. These promising results demonstrate the great potential of the proposed method to facilitate skill learning in surgical robotics and skill assessment in clinical practice.


Assuntos
Intervenção Coronária Percutânea , Robótica , Humanos , Animais , Suínos , Redes Neurais de Computação , Algoritmos , Aprendizagem
8.
Comput Biol Med ; 152: 106341, 2023 01.
Artigo em Inglês | MEDLINE | ID: mdl-36463794

RESUMO

Accurate segmentation of retinal vessels from fundus images is fundamental for the diagnosis of numerous diseases of eye, and an automated vessel segmentation method can effectively help clinicians to make accurate diagnosis for the patients and provide the appropriate treatment schemes. It is important to note that both thick and thin vessels play the key role for disease judgements. Because of complex factors, the precise segmentation of thin vessels is still a great challenge, such as the presence of various lesions, image noise, complex backgrounds and poor contrast in the fundus images. Recently, because of the advantage of context feature representation learning capabilities, deep learning has shown a remarkable segmentation performance on retinal vessels. However, it still has some shortcomings on high-precision retinal vessel extraction due to some factors, such as semantic information loss caused by pooling operations, limited receptive field, etc. To address these problems, this paper proposes a new lightweight segmentation network for precise retinal vessel segmentation, which is called as Wave-Net model on account of the whole shape. To alleviate the influence of semantic information loss problem to thin vessels, to acquire more contexts about micro structures and details, a detail enhancement and denoising block (DED) is proposed to improve the segmentation precision on thin vessels, which replaces the simple skip connections of original U-Net. On the other hand, it could well alleviate the influence of the semantic gap problem. Further, faced with limited receptive field, for multi-scale vessel detection, a multi-scale feature fusion block (MFF) is proposed to fuse cross-scale contexts to achieve higher segmentation accuracy and realize effective processing of local feature maps. Experiments indicate that proposed Wave-Net achieves an excellent performance on retinal vessel segmentation while maintaining a lightweight network design compared to other advanced segmentation methods, and it also has shown a better segmentation ability to thin vessels.


Assuntos
Algoritmos , Vasos Retinianos , Humanos , Vasos Retinianos/diagnóstico por imagem , Fundo de Olho , Processamento de Imagem Assistida por Computador/métodos
9.
Int J Med Robot ; 19(2): e2483, 2023 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-36409623

RESUMO

BACKGROUND: Robot-assisted pelvic fracture closed reduction (RPFCR) positively contributes to patient treatment. However, the current path planning suffers from incomplete obstacle avoidance and long paths. METHOD: A collision detection method is proposed for applications in the pelvic environment to improve the safety of RPFCR surgery. Meanwhile, a defined orientation planning strategy (OPS) and linear sampling search (LSS) are coupled into the A* algorithm to optimise the reduction path. Subsequently, pelvic in vitro experimental platform is built to verify the augmented A*algorithm's feasibility. RESULTS: The augmented A* algorithm planned the shortest path for the same fracture model, and the paths planned by the A* algorithm and experience-based increased by 56.12% and 89.02%, respectively. CONCLUSIONS: The augmented A* algorithm effectively improves surgical safety and shortens the path length, which can be adopted as an effective model for developing RPFCR path planning.


Assuntos
Fraturas Ósseas , Procedimentos de Cirurgia Plástica , Robótica , Humanos , Redução Fechada , Fraturas Ósseas/cirurgia , Pelve/cirurgia
10.
Comput Biol Med ; 151(Pt A): 106216, 2022 12.
Artigo em Inglês | MEDLINE | ID: mdl-36356389

RESUMO

Accurate surgical instrument segmentation can provide the precise location and pose information to the surgeons, assisting the surgeon to accurately judge the follow-up operation during the robot-assisted surgery procedures. Due to strong context extraction ability, there have been significant advances in research of automatic surgical instrument segmentation, especially U-Net and its variant networks. However, there are still some problems to affect segmentation accuracy, like insufficient processing of local features, class imbalance issue, etc. To deal with these problems, with the typical encoder-decoder structure, an effective surgical instrument segmentation network is proposed for providing an end-to-end detection scheme. Specifically, aimed at the problem of insufficient processing of local features, the residual path is introduced for the full feature extraction to strengthen the backward propagation of low-level features. Further, to achieve feature enhancement of local feature maps, a non-local attention block is introduced to insert into the bottleneck layer to acquire global contexts. Besides, to highlight the pixel areas of the surgical instruments, a dual-attention module (DAM) is introduced to make full use of the high-level features extracted from decoder unit and the low-level features delivered by the encoder unit to acquire the attention features and suppress the irrelevant features. To prove the effectiveness and superiority of the proposed segmentation model, experiments are conducted on two public surgical instrument segmentation data sets, including Kvasir-instrument set and Endovis2017 set, which could acquire a 95.77% Dice score and 92.13% mIOU value on Kvasir-instrument set, and simultaneously reach 95.60% Dice score and 92.74% mIOU value on Endovis2017 set respectively. Experimental results show that the proposed segmentation model realizes a superior performance on surgical instruments in comparison to other advanced models, which could provide a good reference for further development of intelligent surgical robots. The source code is provided at https://github.com/lyangucas92/Surg_Net.


Assuntos
Procedimentos Cirúrgicos Robóticos , Cirurgiões , Humanos , Endoscopia , Instrumentos Cirúrgicos , Atenção , Processamento de Imagem Assistida por Computador
11.
Annu Int Conf IEEE Eng Med Biol Soc ; 2022: 1663-1666, 2022 07.
Artigo em Inglês | MEDLINE | ID: mdl-36086459

RESUMO

Automatic surgical phase recognition plays a key role in surgical workflow analysis and overall optimization in clinical work. In the complicated surgical procedures, similar inter-class appearance and drastic variability in phase duration make this still a challenging task. In this paper, a spatio-temporal transformer is proposed for online surgical phase recognition with different granularity. To extract rich spatial information, a spatial transformer is used to model global spatial dependencies of each time index. To overcome the variability in phase duration, a temporal transformer captures the multi-scale temporal context of different time indexes with a dual pyramid pattern. Our method is thoroughly validated on the public Cholec80 dataset with 7 coarse-grained phases and the CATARACTS2020 dataset with 19 fine-grained phases, outperforming state-of-the-art approaches with 91.4% and 84.2% accuracy, taking only 24.5M parameters.


Assuntos
Algoritmos , Fluxo de Trabalho
12.
Front Robot AI ; 9: 913930, 2022.
Artigo em Inglês | MEDLINE | ID: mdl-35991847

RESUMO

Retinal vein injection guided by microscopic image is an innovative procedure for treating retinal vein occlusion. However, the retina organization is complex, fine, and weak, and the operation scale and force are small. Surgeons' limited operation and force-sensing accuracy make it difficult to perform precise and stable drug injection operations on the retina in a magnified field of image vision. In this paper, a 3-DOF automatic drug injection mechanism was designed for microscopic image guiding robot-assisted needle delivery and automatic drug injection. Additionally, the robot-assisted real-time three-dimensional micro-force-sensing method for retinal vein injection was proposed. Based on the layout of three FBG sensors on the hollow outer wall of the nested needle tube in a circular array of nickel-titanium alloys, the real-time sensing of the contact force between the intraoperative instrument and the blood vessel was realized. The experimental data of 15 groups of porcine eyeball retinal veins with diameters of 100-200 µm showed that the piercing force of surgical instruments and blood vessels is 5.95∼12.97 mN, with an average value of 9.98 mN. Furthermore, 20 groups of experimental measurements on chicken embryo blood vessels with diameters of 150-500 µm showed that the piercing force was 4.02∼23.4 mN, with an average value of 12.05 mN.

13.
Int J Med Robot ; 18(4): e2401, 2022 Aug.
Artigo em Inglês | MEDLINE | ID: mdl-35368143

RESUMO

BACKGROUND: Recognition of markers in the augmented reality system can reduce the additional cost of a guide plate required for the removal of benign tumours in oral and maxillofacial surgery, but the use of markers often has complex problems. METHOD: In order to avoid the complex problem of using markers, an augmented reality system based on a marker-free registration method was proposed to track the contour of the mandible edge. Use the computer to perform preoperative planning on the jaw model, select and mark the path of maxillofacial lesion resection. RESULTS: This method has an error of the surface matching was 0.6453 ± 0.2826 $0.6453\pm 0.2826$ mm, and an error of the surgical resection was 0.4858 ± 0.3712 $0.4858\pm 0.3712$ mm. CONCLUSIONS: We demonstrate that the system can accurately enhance the display of the surgical path and provide guidance for the tradition of maxillofacial surgery.


Assuntos
Realidade Aumentada , Cirurgia Assistida por Computador , Cirurgia Bucal , Estudos de Viabilidade , Humanos , Imageamento Tridimensional , Mandíbula/cirurgia
14.
IEEE J Biomed Health Inform ; 26(7): 3209-3217, 2022 07.
Artigo em Inglês | MEDLINE | ID: mdl-35226612

RESUMO

Surgical image segmentation is critical for surgical robot control and computer-assisted surgery. In the surgical scene, the local features of objects are highly similar, and the illumination interference is strong, which makes surgical image segmentation challenging. To address the above issues, a bilinear squeeze reasoning network is proposed for surgical image segmentation. In it, the space squeeze reasoning module is proposed, which adopts height pooling and width pooling to squeeze global contexts in the vertical and horizontal directions, respectively. The similarity between each horizontal position and each vertical position is calculated to encode long-range semantic dependencies and establish the affinity matrix. The feature maps are also squeezed from both the vertical and horizontal directions to model channel relations. Guided by channel relations, the affinity matrix is expanded to the same size as the input features. It captures long-range semantic dependencies from different directions, helping address the local similarity issue. Besides, a low-rank bilinear fusion module is proposed to enhance the model's ability to recognize similar features. This module is based on the low-rank bilinear model to capture the inter-layer feature relations. It integrates the location details from low-level features and semantic information from high-level features. Various semantics can be represented more accurately, which effectively improves feature representation. The proposed network achieves state-of-the-art performance on cataract image segmentation dataset CataSeg and robotic image segmentation dataset EndoVis 2018.


Assuntos
Processamento de Imagem Assistida por Computador , Cirurgia Assistida por Computador , Humanos , Processamento de Imagem Assistida por Computador/métodos , Redes Neurais de Computação , Semântica
15.
IEEE Trans Cybern ; 52(4): 2565-2577, 2022 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-32697730

RESUMO

The clinical success of the percutaneous coronary intervention (PCI) is highly dependent on endovascular manipulation skills and dexterous manipulation strategies of interventionalists. However, the analysis of endovascular manipulations and related discussion for technical skill assessment are limited. In this study, a multilayer and multimodal-fusion architecture is proposed to recognize six typical endovascular manipulations. The synchronously acquired multimodal motion signals from ten subjects are used as the inputs of the architecture independently. Six classification-based and two rule-based fusion algorithms are evaluated for performance comparisons. The recognition metrics under the determined architecture are further used to assess technical skills. The experimental results indicate that the proposed architecture can achieve the overall accuracy of 96.41%, much higher than that of a single-layer recognition architecture (92.85%). In addition, the multimodal fusion brings significant performance improvement in comparison with single-modal schemes. Furthermore, the K -means-based skill assessment can obtain an accuracy of 95% to cluster the attempts made by different skill-level groups. These hopeful results indicate the great possibility of the architecture to facilitate clinical skill assessment and skill learning.


Assuntos
Intervenção Coronária Percutânea , Algoritmos , Competência Clínica , Humanos , Aprendizagem
16.
Med Image Anal ; 76: 102310, 2022 02.
Artigo em Inglês | MEDLINE | ID: mdl-34954623

RESUMO

Surgical instrument segmentation plays a promising role in robot-assisted surgery. However, illumination issues often appear in surgical scenes, altering the color and texture of surgical instruments. Changes in visual features make surgical instrument segmentation difficult. To address illumination issues, the SurgiNet is proposed to learn pyramid attention features. The double attention module is designed to capture the semantic dependencies between locations and channels. Based on semantic dependencies, the semantic features in the disturbed area can be inferred for addressing illumination issues. Pyramid attention is aggregated to capture multi-scale features and make predictions more accurate. To perform model compression, class-wise self-distillation is proposed to enhance the representation learning of the network, which performs feature distillation within the class to eliminate interference from other classes. Top-down and multi-stage knowledge distillation is designed to distill class probability maps. By inter-layer supervision, high-level probability maps are applied to calibrate the probability distribution of low-level probability maps. Since class-wise distillation enhances the self-learning of the network, the network can get excellent performance with a lightweight backbone. The proposed network achieves the state-of-the-art performance of 89.14% mIoU on CataIS with only 1.66 GFlops and 2.05 M parameters. It also takes first place on EndoVis 2017 with 66.30% mIoU.


Assuntos
Processamento de Imagem Assistida por Computador , Humanos , Atenção , Semântica , Instrumentos Cirúrgicos
17.
IEEE Trans Biomed Eng ; 69(4): 1406-1416, 2022 04.
Artigo em Inglês | MEDLINE | ID: mdl-34613905

RESUMO

OBJECTIVE: In this paper, Keypoint Localization Region-based CNN (KL R-CNN) is proposed, which can simultaneously accomplish the guidewire detection and endpoint localization in a unified model. METHODS: KL R-CNN modifies Mask R-CNN by replacing the mask branch with a novel keypoint localization branch. Besides, some settings of Mask R-CNN are also modified to generate the keypoint localization results at a higher detail level. At the same time, based on the existing metrics of Average Precision (AP) and Percentage of Correct Keypoints (PCK), a new metric named APPCK is proposed to evaluate the overall performance on the multi-guidewire endpoint localization task. Compared with existing metrics, APPCK is easy to use and its results are more intuitive. RESULTS: Compared with existing methods, KL R-CNN has better performance when the threshold is loose, reaching a mean APPCK of 90.65% when the threshold is 9 pixels. CONCLUSION: KL R-CNN achieves the state-of-the-art performance on the multi-guidewire endpoint localization task and has application potentials. SIGNIFICANCE: KL R-CNN can achieve the localization of guidewire endpoints in fluoroscopy images, which is a prerequisite for computer-assisted percutaneous coronary intervention. KL R-CNN can also be extended to other multi-instrument localization tasks.


Assuntos
Processamento de Imagem Assistida por Computador , Intervenção Coronária Percutânea , Cateterismo , Fluoroscopia , Processamento de Imagem Assistida por Computador/métodos
18.
IEEE Trans Med Imaging ; 40(8): 2002-2014, 2021 08.
Artigo em Inglês | MEDLINE | ID: mdl-33788685

RESUMO

The real-time localization of the guidewire endpoints is a stepping stone to computer-assisted percutaneous coronary intervention (PCI). However, methods for multi-guidewire endpoint localization in fluoroscopy images are still scarce. In this paper, we introduce a framework for real-time multi-guidewire endpoint localization in fluoroscopy images. The framework consists of two stages, first detecting all guidewire instances in the fluoroscopy image, and then locating the endpoints of each single guidewire instance. In the first stage, a YOLOv3 detector is used for guidewire detection, and a post-processing algorithm is proposed to refine the guidewire detection results. In the second stage, a Segmentation Attention-hourglass (SA-hourglass) network is proposed to predict the endpoint locations of each single guidewire instance. The SA-hourglass network can be generalized to the keypoint localization of other surgical instruments. In our experiments, the SA-hourglass network is applied not only on a guidewire dataset but also on a retinal microsurgery dataset, reaching the mean pixel error (MPE) of 2.20 pixels on the guidewire dataset and the MPE of 5.30 pixels on the retinal microsurgery dataset, both achieving the state-of-the-art localization results. Besides, the inference rate of our framework is at least 20FPS, which meets the real-time requirement of fluoroscopy images (6-12FPS).


Assuntos
Intervenção Coronária Percutânea , Algoritmos , Cateterismo , Fluoroscopia , Humanos
19.
Med Image Anal ; 70: 101920, 2021 05.
Artigo em Inglês | MEDLINE | ID: mdl-33676097

RESUMO

Intraoperative tracking of laparoscopic instruments is often a prerequisite for computer and robotic-assisted interventions. While numerous methods for detecting, segmenting and tracking of medical instruments based on endoscopic video images have been proposed in the literature, key limitations remain to be addressed: Firstly, robustness, that is, the reliable performance of state-of-the-art methods when run on challenging images (e.g. in the presence of blood, smoke or motion artifacts). Secondly, generalization; algorithms trained for a specific intervention in a specific hospital should generalize to other interventions or institutions. In an effort to promote solutions for these limitations, we organized the Robust Medical Instrument Segmentation (ROBUST-MIS) challenge as an international benchmarking competition with a specific focus on the robustness and generalization capabilities of algorithms. For the first time in the field of endoscopic image processing, our challenge included a task on binary segmentation and also addressed multi-instance detection and segmentation. The challenge was based on a surgical data set comprising 10,040 annotated images acquired from a total of 30 surgical procedures from three different types of surgery. The validation of the competing methods for the three tasks (binary segmentation, multi-instance detection and multi-instance segmentation) was performed in three different stages with an increasing domain gap between the training and the test data. The results confirm the initial hypothesis, namely that algorithm performance degrades with an increasing domain gap. While the average detection and segmentation quality of the best-performing algorithms is high, future research should concentrate on detection and segmentation of small, crossing, moving and transparent instrument(s) (parts).


Assuntos
Processamento de Imagem Assistida por Computador , Laparoscopia , Algoritmos , Artefatos
20.
Ann Transl Med ; 8(14): 872, 2020 Jul.
Artigo em Inglês | MEDLINE | ID: mdl-32793716

RESUMO

BACKGROUND: Cataract surgery has been recently developed from sight rehabilitating surgery to accurate refractive surgery. The precise concentration of intraocular lens (IOL) is crucial for postoperative high visual quanlity. The three-dimentional (3D) images of ocular anterior segment captured by optial coherence tomography (OCT) make it possible to evaluate the IOL position in 3D space, which provide insights into factors relavant to the visual quanlity and better design of new functional IOL. The deep learning algorithm potentially quantify the IOL position in an objective and efficient way. METHODS: The region-based fully convolutional network (R-FCN) was used to recogonize and delineate the IOL configuration in 3D OCT images. Scleral spur was identified automatically. Then the tilt angle of the IOL relative to the scleral spur plane along with its decentration with respect to the pupil were calculated. Repeatability and reliability of the method was evaluated by the intraclass correlation coefficient. RESULTS: After improvement, the R-FCN network recognition efficiency of IOL configuration reached 0.910. The ICC of reliability and repeatability of the method is 0.867 and 0.901. The average tilt angle of the IOL relative to scleral spur is located in 1.65±1.00 degrees. The offsets dx and dy occurring in the early X and Y directions of the IOL are 0.29±0.22 and 0.33±0.24 mm, respectively. The IOL offset distance is 0.44±0.33 mm. CONCLUSIONS: We proposed a practical method to quantify the IOL postion in 3D space based on OCT images and assisted by an algorithm.

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA
...