Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 3 de 3
Filtrar
Mais filtros

Bases de dados
Tipo de documento
País de afiliação
Intervalo de ano de publicação
1.
IEEE Trans Med Imaging ; 38(11): 2556-2568, 2019 11.
Artigo em Inglês | MEDLINE | ID: mdl-30908194

RESUMO

Quantification of cerebral white matter hyperintensities (WMH) of presumed vascular origin is of key importance in many neurological research studies. Currently, measurements are often still obtained from manual segmentations on brain MR images, which is a laborious procedure. The automatic WMH segmentation methods exist, but a standardized comparison of the performance of such methods is lacking. We organized a scientific challenge, in which developers could evaluate their methods on a standardized multi-center/-scanner image dataset, giving an objective comparison: the WMH Segmentation Challenge. Sixty T1 + FLAIR images from three MR scanners were released with the manual WMH segmentations for training. A test set of 110 images from five MR scanners was used for evaluation. The segmentation methods had to be containerized and submitted to the challenge organizers. Five evaluation metrics were used to rank the methods: 1) Dice similarity coefficient; 2) modified Hausdorff distance (95th percentile); 3) absolute log-transformed volume difference; 4) sensitivity for detecting individual lesions; and 5) F1-score for individual lesions. In addition, the methods were ranked on their inter-scanner robustness; 20 participants submitted their methods for evaluation. This paper provides a detailed analysis of the results. In brief, there is a cluster of four methods that rank significantly better than the other methods, with one clear winner. The inter-scanner robustness ranking shows that not all the methods generalize to unseen scanners. The challenge remains open for future submissions and provides a public platform for method evaluation.


Assuntos
Processamento de Imagem Assistida por Computador/métodos , Imageamento por Ressonância Magnética/métodos , Substância Branca/diagnóstico por imagem , Idoso , Algoritmos , Feminino , Humanos , Masculino , Pessoa de Meia-Idade
2.
Pattern Recognit Lett ; 76: 32-40, 2016 Jun 01.
Artigo em Inglês | MEDLINE | ID: mdl-27175043

RESUMO

Conventional curve skeletonization algorithms using the principle of Blum's transform, often, produce unwanted spurious branches due to boundary irregularities, digital effects, and other artifacts. This paper presents a new robust and efficient curve skeletonization algorithm for three-dimensional (3-D) elongated fuzzy objects using a minimum cost path approach, which avoids spurious branches without requiring post-pruning. Starting from a root voxel, the method iteratively expands the skeleton by adding new branches in each iteration that connects the farthest quench voxel to the current skeleton using a minimum cost path. The path-cost function is formulated using a novel measure of local significance factor defined by the fuzzy distance transform field, which forces the path to stick to the centerline of an object. The algorithm terminates when dilated skeletal branches fill the entire object volume or the current farthest quench voxel fails to generate a meaningful skeletal branch. Accuracy of the algorithm has been evaluated using computer-generated phantoms with known skeletons. Performance of the method in terms of false and missing skeletal branches, as defined by human experts, has been examined using in vivo CT imaging of human intrathoracic airways. Results from both experiments have established the superiority of the new method as compared to the existing methods in terms of accuracy as well as robustness in detecting true and false skeletal branches. The new algorithm makes a significant reduction in computation complexity by enabling detection of multiple new skeletal branches in one iteration. Specifically, this algorithm reduces the number of iterations from the number of terminal tree branches to the worst case performance of tree depth. In fact, experimental results suggest that, on an average, the order of computation complexity is reduced to the logarithm of the number of terminal branches of a tree-like object.

3.
Proc IAPR Int Conf Pattern Recogn ; 2014: 942-947, 2014 Aug.
Artigo em Inglês | MEDLINE | ID: mdl-25621320

RESUMO

Traditional arc skeletonization algorithms using the principle of Blum's transform, often, produce unwanted spurious branches due to boundary irregularities and digital effects on objects and other artifacts. This paper presents a new robust approach of extracting arc skeletons for three-dimensional (3-D) elongated fuzzy objects, which avoids spurious branches without requiring post-pruning. Starting from a root voxel, the method iteratively expands the skeleton by adding a new branch in each iteration that connects the farthest voxel to the current skeleton using a minimum-cost geodesic path. The path-cost function is formulated using a novel measure of local significance factor defined by fuzzy distance transform field, which forces the path to stick to the centerline of the object. The algorithm terminates when dilated skeletal branches fill the entire object volume or the current farthest voxel fails to generate a meaningful branch. Accuracy of the algorithm has been evaluated using computer-generated blurred and noisy phantoms with known skeletons. Performance of the method in terms of false and missing skeletal branches, as defined by human expert, has been examined using in vivo CT imaging of human intrathoracic airways. Experimental results from both experiments have established the superiority of the new method as compared to a widely used conventional method in terms of accuracy of medialness as well as robustness of true and false skeletal branches.

SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA