Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 11 de 11
Filtrar
1.
Sensors (Basel) ; 24(13)2024 Jun 22.
Artigo em Inglês | MEDLINE | ID: mdl-39000853

RESUMO

Hyperspectral images (HSIs) possess an inherent three-order structure, prompting increased interest in extracting 3D features. Tensor analysis and low-rank representations, notably truncated higher-order SVD (T-HOSVD), have gained prominence for this purpose. However, determining the optimal order and addressing sensitivity to changes in data distribution remain challenging. To tackle these issues, this paper introduces an unsupervised Superpixelwise Multiscale Adaptive T-HOSVD (SmaT-HOSVD) method. Leveraging superpixel segmentation, the algorithm identifies homogeneous regions, facilitating the extraction of local features to enhance spatial contextual information within the image. Subsequently, T-HOSVD is adaptively applied to the obtained superpixel blocks for feature extraction and fusion across different scales. SmaT-HOSVD harnesses superpixel blocks and low-rank representations to extract 3D features, effectively capturing both spectral and spatial information of HSIs. By integrating optimal-rank estimation and multiscale fusion strategies, it acquires more comprehensive low-rank information and mitigates sensitivity to data variations. Notably, when trained on subsets comprising 2%, 1%, and 1% of the Indian Pines, University of Pavia, and Salinas datasets, respectively, SmaT-HOSVD achieves impressive overall accuracies of 93.31%, 97.21%, and 99.25%, while maintaining excellent efficiency. Future research will explore SmaT-HOSVD's applicability in deep-sea HSI classification and pursue additional avenues for advancing the field.

2.
Sensors (Basel) ; 23(8)2023 Apr 10.
Artigo em Inglês | MEDLINE | ID: mdl-37112209

RESUMO

There are some irregular and disordered noise points in large-scale point clouds, and the accuracy of existing large-scale point cloud classification methods still needs further improvement. This paper proposes a network named MFTR-Net, which considers the local point cloud's eigenvalue calculation. The eigenvalues of 3D point cloud data and the 2D eigenvalues of projected point clouds on different planes are calculated to express the local feature relationship between adjacent point clouds. A regular point cloud feature image is constructed and inputs into the designed convolutional neural network. The network adds TargetDrop to be more robust. The experimental result shows that our methods can learn more high-dimensional feature information, further improving point cloud classification, and our approach can achieve 98.0% accuracy with the Oakland 3D dataset.

3.
Sensors (Basel) ; 22(19)2022 Oct 02.
Artigo em Inglês | MEDLINE | ID: mdl-36236576

RESUMO

With the increase in the amount of 3D point cloud data and the wide application of point cloud registration in various fields, the question of whether it is possible to quickly extract the key points of registration and perform accurate coarse registration has become a question to be urgently answered. In this paper, we proposed a novel semantic segmentation algorithm that enables the extracted feature point cloud to have a clustering effect for fast registration. First of all, an adaptive technique was proposed to determine the domain radius of a local point. Secondly, the feature intensity of the point is scored through the regional fluctuation coefficient and stationary coefficient calculated by the normal vector, and the high feature region to be registered is preliminarily determined. In the end, FPFH is used to describe the geometric features of the extracted semantic feature point cloud, so as to realize the coarse registration from the local point cloud to the overall point cloud. The results show that the point cloud can be roughly segmented based on the uniqueness of semantic features. The use of a semantic feature point cloud can make the point cloud have a very fast response speed based on the accuracy of coarse registration, almost equal to that of using the original point cloud, which is conducive to the rapid determination of the initial attitude.

4.
Sensors (Basel) ; 15(8): 18587-612, 2015 Jul 29.
Artigo em Inglês | MEDLINE | ID: mdl-26230701

RESUMO

Camera-based 3D reconstruction of physical objects is one of the most popular computer vision trends in recent years. Many systems have been built to model different real-world subjects, but there is lack of a completely robust system for plants. This paper presents a full 3D reconstruction system that incorporates both hardware structures (including the proposed structured light system to enhance textures on object surfaces) and software algorithms (including the proposed 3D point cloud registration and plant feature measurement). This paper demonstrates the ability to produce 3D models of whole plants created from multiple pairs of stereo images taken at different viewing angles, without the need to destructively cut away any parts of a plant. The ability to accurately predict phenotyping features, such as the number of leaves, plant height, leaf size and internode distances, is also demonstrated. Experimental results show that, for plants having a range of leaf sizes and a distance between leaves appropriate for the hardware design, the algorithms successfully predict phenotyping features in the target crops, with a recall of 0.97 and a precision of 0.89 for leaf detection and less than a 13-mm error for plant size, leaf size and internode distance.


Assuntos
Imageamento Tridimensional/métodos , Luz , Plantas/anatomia & histologia , Algoritmos , Brassica/anatomia & histologia , Cucumis sativus/anatomia & histologia , Solanum lycopersicum/anatomia & histologia , Tamanho do Órgão , Fenótipo , Folhas de Planta/anatomia & histologia , Solo
5.
J Imaging Inform Med ; 37(1): 280-296, 2024 Feb.
Artigo em Inglês | MEDLINE | ID: mdl-38343216

RESUMO

Cervical cancer is a significant health problem worldwide, and early detection and treatment are critical to improving patient outcomes. To address this challenge, a deep learning (DL)-based cervical classification system is proposed using 3D convolutional neural network and Vision Transformer (ViT) module. The proposed model leverages the capability of 3D CNN to extract spatiotemporal features from cervical images and employs the ViT model to capture and learn complex feature representations. The model consists of an input layer that receives cervical images, followed by a 3D convolution block, which extracts features from the images. The feature maps generated are down-sampled using max-pooling block to eliminate redundant information and preserve important features. Four Vision Transformer models are employed to extract efficient feature maps of different levels of abstraction. The output of each Vision Transformer model is an efficient set of feature maps that captures spatiotemporal information at a specific level of abstraction. The feature maps generated by the Vision Transformer models are then supplied into the 3D feature pyramid network (FPN) module for feature concatenation. The 3D squeeze-and-excitation (SE) block is employed to obtain efficient feature maps that recalibrate the feature responses of the network based on the interdependencies between different feature maps, thereby improving the discriminative power of the model. At last, dimension minimization of feature maps is executed using 3D average pooling layer. Its output is then fed into a kernel extreme learning machine (KELM) for classification into one of the five classes. The KELM uses radial basis kernel function (RBF) for mapping features in high-dimensional feature space and classifying the input samples. The superiority of the proposed model is known using simulation results, achieving an accuracy of 98.6%, demonstrating its potential as an effective tool for cervical cancer classification. Also, it can be used as a diagnostic supportive tool to assist medical experts in accurately identifying cervical cancer in patients.

6.
Eur Heart J Cardiovasc Imaging ; 24(4): 503-511, 2023 03 21.
Artigo em Inglês | MEDLINE | ID: mdl-35793360

RESUMO

AIMS: Cardiovascular magnetic resonance (CMR) is valuable for the detection of cardiac involvement in neuromuscular diseases (NMDs). We explored the value of 2D- and 3D-left ventricular (LV) myocardial strain analysis using feature-tracking (FT)-CMR to detect subclinical cardiac involvement in NMD. METHODS AND RESULTS: The study included retrospective analysis of 111 patients with NMD; mitochondrial cytopathies (n = 14), Friedreich's ataxia (FA, n = 27), myotonic dystrophy (n = 27), Becker/Duchenne's muscular dystrophy (BMD/DMD, n = 15), Duchenne's carriers (n = 6), or other (n = 22) and 57 age- and sex-matched healthy volunteers. Biventricular volumes, myocardial late gadolinium enhancement (LGE), and LV myocardial deformation were assessed by FT-CMR, including 2D and 3D global circumferential strain (GCS), global radial strain (GRS), global longitudinal strain (GLS), and torsion. Compared with the healthy volunteers, patients with NMD had impaired 2D-GCS (P < 0.001) and 2D-GRS (in the short-axis, P < 0.001), but no significant differences in 2D-GRS long-axis (P = 0.101), 2D-GLS (P = 0.069), or torsion (P = 0.122). 3D-GRS, 3D-GCS, and 3D-GLS values were all significantly different to the control group (P < 0.0001 for all). Especially, even NMD patients without overt cardiac involvement (i.e. LV dilation/hypertrophy, reduced LVEF, or LGE presence) had significantly impaired 3D-GRS, GCS, and GLS vs. the control group (P < 0.0001). 3D-GRS and GCS values were significantly associated with the LGE presence and pattern, being most impaired in patients with transmural LGE. CONCLUSIONS: 3D-FT CMR detects subclinical cardiac muscle disease in patients with NMD even before the development of replacement fibrosis or ventricular remodelling which may be a useful imaging biomarker for early detection of cardiac involvement.


Assuntos
Doenças Neuromusculares , Função Ventricular Esquerda , Humanos , Estudos Retrospectivos , Função Ventricular Esquerda/fisiologia , Meios de Contraste , Imagem Cinética por Ressonância Magnética/métodos , Gadolínio , Miocárdio , Hipertrofia Ventricular Esquerda , Espectroscopia de Ressonância Magnética , Doenças Neuromusculares/complicações , Doenças Neuromusculares/diagnóstico por imagem , Valor Preditivo dos Testes
7.
Int J Comput Assist Radiol Surg ; 18(6): 1025-1032, 2023 Jun.
Artigo em Inglês | MEDLINE | ID: mdl-37079248

RESUMO

PURPOSE: In laparoscopic liver surgery, preoperative information can be overlaid onto the intra-operative scene by registering a 3D preoperative model to the intra-operative partial surface reconstructed from the laparoscopic video. To assist with this task, we explore the use of learning-based feature descriptors, which, to our best knowledge, have not been explored for use in laparoscopic liver registration. Furthermore, a dataset to train and evaluate the use of learning-based descriptors does not exist. METHODS: We present the LiverMatch dataset consisting of 16 preoperative models and their simulated intra-operative 3D surfaces. We also propose the LiverMatch network designed for this task, which outputs per-point feature descriptors, visibility scores, and matched points. RESULTS: We compare the proposed LiverMatch network with a network closest to LiverMatch and a histogram-based 3D descriptor on the testing split of the LiverMatch dataset, which includes two unseen preoperative models and 1400 intra-operative surfaces. Results suggest that our LiverMatch network can predict more accurate and dense matches than the other two methods and can be seamlessly integrated with a RANSAC-ICP-based registration algorithm to achieve an accurate initial alignment. CONCLUSION: The use of learning-based feature descriptors in laparoscopic liver registration (LLR) is promising, as it can help achieve an accurate initial rigid alignment, which, in turn, serves as an initialization for subsequent non-rigid registration.


Assuntos
Laparoscopia , Fígado , Humanos , Fígado/diagnóstico por imagem , Fígado/cirurgia , Laparoscopia/métodos , Algoritmos
8.
Front Neurosci ; 17: 1330077, 2023.
Artigo em Inglês | MEDLINE | ID: mdl-38268710

RESUMO

Introduction: Multimodal emotion recognition has become a hot topic in human-computer interaction and intelligent healthcare fields. However, combining information from different human different modalities for emotion computation is still challenging. Methods: In this paper, we propose a three-dimensional convolutional recurrent neural network model (referred to as 3FACRNN network) based on multimodal fusion and attention mechanism. The 3FACRNN network model consists of a visual network and an EEG network. The visual network is composed of a cascaded convolutional neural network-time convolutional network (CNN-TCN). In the EEG network, the 3D feature building module was added to integrate band information, spatial information and temporal information of the EEG signal, and the band attention and self-attention modules were added to the convolutional recurrent neural network (CRNN). The former explores the effect of different frequency bands on network recognition performance, while the latter is to obtain the intrinsic similarity of different EEG samples. Results: To investigate the effect of different frequency bands on the experiment, we obtained the average attention mask for all subjects in different frequency bands. The distribution of the attention masks across the different frequency bands suggests that signals more relevant to human emotions may be active in the high frequency bands γ (31-50 Hz). Finally, we try to use the multi-task loss function Lc to force the approximation of the intermediate feature vectors of the visual and EEG modalities, with the aim of using the knowledge of the visual modalities to improve the performance of the EEG network model. The mean recognition accuracy and standard deviation of the proposed method on the two multimodal sentiment datasets DEAP and MAHNOB-HCI (arousal, valence) were 96.75 ± 1.75, 96.86 ± 1.33; 97.55 ± 1.51, 98.37 ± 1.07, better than those of the state-of-the-art multimodal recognition approaches. Discussion: The experimental results show that starting from the multimodal information, the facial video frames and electroencephalogram (EEG) signals of the subjects are used as inputs to the emotion recognition network, which can enhance the stability of the emotion network and improve the recognition accuracy of the emotion network. In addition, in future work, we will try to utilize sparse matrix methods and deep convolutional networks to improve the performance of multimodal emotion networks.

9.
Sensors (Basel) ; 8(8): 4505-4528, 2008 Aug 04.
Artigo em Inglês | MEDLINE | ID: mdl-27873771

RESUMO

Airborne laser scanning (ALS) is a remote sensing technique well-suited for 3D vegetation mapping and structure characterization because the emitted laser pulses are able to penetrate small gaps in the vegetation canopy. The backscattered echoes from the foliage, woody vegetation, the terrain, and other objects are detected, leading to a cloud of points. Higher echo densities (> 20 echoes/m2) and additional classification variables from full-waveform (FWF) ALS data, namely echo amplitude, echo width and information on multiple echoes from one shot, offer new possibilities in classifying the ALS point cloud. Currently FWF sensor information is hardly used for classification purposes. This contribution presents an object-based point cloud analysis (OBPA) approach, combining segmentation and classification of the 3D FWF ALS points designed to detect tall vegetation in urban environments. The definition tall vegetation includes trees and shrubs, but excludes grassland and herbage. In the applied procedure FWF ALS echoes are segmented by a seeded region growing procedure. All echoes sorted descending by their surface roughness are used as seed points. Segments are grown based on echo width homogeneity. Next, segment statistics (mean, standard deviation, and coefficient of variation) are calculated by aggregating echo features such as amplitude and surface roughness. For classification a rule base is derived automatically from a training area using a statistical classification tree. To demonstrate our method we present data of three sites with around 500,000 echoes each. The accuracy of the classified vegetation segments is evaluated for two independent validation sites. In a point-wise error assessment, where the classification is compared with manually classified 3D points, completeness and correctness better than 90% are reached for the validation sites. In comparison to many other algorithms the proposed 3D point classification works on the original measurements directly, i.e. the acquired points. Gridding of the data is not necessary, a process which is inherently coupled to loss of data and precision. The 3D properties provide especially a good separability of buildings and terrain points respectively, if they are occluded by vegetation.

10.
Comput Biol Med ; 92: 64-72, 2018 01 01.
Artigo em Inglês | MEDLINE | ID: mdl-29154123

RESUMO

Pulmonary nodule detection has a significant impact on early diagnosis of lung cancer. To effectively detect pulmonary nodules from interferential vessels in chest CT datasets, this paper proposes a novel 3D skeletonization feature, named as voxels remove rate. Based on this feature, a computer-aided detection system is constructed to validate its performance. The system mainly consists of five stages. Firstly, the lung tissues are segmented by a global optimal active contour model, which can extract all structures (including juxta-pleural nodules) in the lung region. Secondly, thresholding, 3D binary morphological operations, and 3D connected components labeling are utilized to extract candidates of pulmonary nodules. Thirdly, combining the voxels remove rate with other nine existing 3D features (including gray features and shape features), the extracted candidates are characterized. Then, prior anatomical knowledge is utilized for preliminary screening of numerous invalid nodule candidates. Finally, false positives are reduced by support vector machine. Our system is evaluated on early stage lung cancer subjects obtained from the publicly available LIDC-IDRI database. The result shows the proposed 3D skeletonization feature is a useful indicator that efficiently differentiates lung nodules from the other suspicious structures. The computer-aided detection system based on this feature can detect various types of nodules, including solitary, juxta-pleural and juxta-vascular nodules.


Assuntos
Imageamento Tridimensional/métodos , Neoplasias Pulmonares/diagnóstico por imagem , Interpretação de Imagem Radiográfica Assistida por Computador/métodos , Nódulo Pulmonar Solitário/diagnóstico por imagem , Tomografia Computadorizada por Raios X/métodos , Bases de Dados Factuais , Humanos
11.
Comput Assist Surg (Abingdon) ; 22(sup1): 319-325, 2017 12.
Artigo em Inglês | MEDLINE | ID: mdl-29094615

RESUMO

BACKGROUND: Compared with the traditional point-based registration in the image-guided neurosurgery system, surface-based registration is preferable because it does not use fiducial markers before image scanning and does not require image acquisition dedicated for navigation purposes. However, most existing surface-based registration methods must include a manual step for coarse registration, which increases the registration time and elicits some inconvenience and uncertainty. METHODS: A new automatic surface-based registration method is proposed, which applies 3D surface feature description and matching algorithm to obtain point correspondences for coarse registration and uses the iterative closest point (ICP) algorithm in the last step to obtain an image-to-patient registration. RESULTS: Both phantom and clinical data were used to execute automatic registrations and target registration error (TRE) calculated to verify the practicality and robustness of the proposed method. In phantom experiments, the registration accuracy was stable across different downsampling resolutions (18-26 mm) and different support radii (2-6 mm). In clinical experiments, the mean TREs of two patients by registering full head surfaces were 1.30 mm and 1.85 mm. CONCLUSION: This study introduced a new robust automatic surface-based registration method based on 3D feature matching. The method achieved sufficient registration accuracy with different real-world surface regions in phantom and clinical experiments.


Assuntos
Imageamento Tridimensional , Neuronavegação , Procedimentos Neurocirúrgicos/métodos , Imagens de Fantasmas , Cirurgia Assistida por Computador/métodos , Estudos de Viabilidade , Marcadores Fiduciais , Humanos
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA