Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 15 de 15
Filtrar
1.
Sensors (Basel) ; 22(5)2022 Feb 28.
Artigo em Inglês | MEDLINE | ID: mdl-35271051

RESUMO

Various types of motion blur are frequently observed in the images captured by sensors based on thermal and photon detectors. The difference in mechanisms between thermal and photon detectors directly results in different patterns of motion blur. Motivated by this observation, we propose a novel method to synthesize blurry images from sharp images by analyzing the mechanisms of the thermal detector. Further, we propose a novel blur kernel rendering method, which combines our proposed motion blur model with the inertial sensor in the thermal image domain. The accuracy of the blur kernel rendering method is evaluated by the task of thermal image deblurring. We construct a synthetic blurry image dataset based on acquired thermal images using an infrared camera for evaluation. This dataset is the first blurry thermal image dataset with ground-truth images in the thermal image domain. Qualitative and quantitative experiments are extensively carried out on our dataset, which show that our proposed method outperforms state-of-the-art methods.

2.
Cytometry A ; 85(8): 709-18, 2014 Aug.
Artigo em Inglês | MEDLINE | ID: mdl-24677732

RESUMO

Automatic segmentation of cell nuclei clusters is a key building block in systems for quantitative analysis of microscopy cell images. For that reason, it has received a great attention over the last decade, and diverse automatic approaches to segment clustered nuclei with varying levels of performance under different test conditions have been proposed in literature. To the best of our knowledge, however, so far there is no comparative study on the methods. This study is a first attempt to fill this research gap. More precisely, the purpose of this study is to present an objective performance comparison of existing state-of-the-art segmentation methods. Particularly, the impact of their accuracy on classification of thyroid follicular lesions is also investigated "quantitatively" under the same experimental condition, to evaluate the applicability of the methods. Thirteen different segmentation approaches are compared in terms of not only errors in nuclei segmentation and delineation, but also their impact on the performance of system to classify thyroid follicular lesions using different metrics (e.g., diagnostic accuracy, sensitivity, specificity, etc.). Extensive experiments have been conducted on a total of 204 digitized thyroid biopsy specimens. Our study demonstrates that significant diagnostic errors can be avoided using more advanced segmentation approaches. We believe that this comprehensive comparative study serves as a reference point and guide for developers and practitioners in choosing an appropriate automatic segmentation technique adopted for building automated systems for specifically classifying follicular thyroid lesions.


Assuntos
Adenocarcinoma Folicular/classificação , Adenocarcinoma Folicular/patologia , Automação , Núcleo Celular/patologia , Processamento de Imagem Assistida por Computador , Humanos
3.
IEEE Trans Pattern Anal Mach Intell ; 46(9): 6326-6340, 2024 Sep.
Artigo em Inglês | MEDLINE | ID: mdl-38502625

RESUMO

Creating novel views from a single image has achieved tremendous strides with advanced autoregressive models, as unseen regions have to be inferred from the visible scene contents. Although recent methods generate high-quality novel views, synthesizing with only one explicit or implicit 3D geometry has a trade-off between two objectives that we call the "seesaw" problem: 1) preserving reprojected contents and 2) completing realistic out-of-view regions. Also, autoregressive models require a considerable computational cost. In this paper, we propose a single-image view synthesis framework for mitigating the seesaw problem while utilizing an efficient non-autoregressive model. Motivated by the characteristics that explicit methods well preserve reprojected pixels and implicit methods complete realistic out-of-view regions, we introduce a loss function to complement two renderers. Our loss function promotes that explicit features improve the reprojected area of implicit features and implicit features improve the out-of-view area of explicit features. With the proposed architecture and loss function, we can alleviate the seesaw problem, outperforming autoregressive-based state-of-the-art methods and generating an image ≈ 100 times faster. We validate the efficiency and effectiveness of our method with experiments on RealEstate10 K and ACID datasets.

4.
IEEE Trans Pattern Anal Mach Intell ; 45(5): 5918-5934, 2023 May.
Artigo em Inglês | MEDLINE | ID: mdl-36070277

RESUMO

Online action detection, which aims to identify an ongoing action from a streaming video, is an important subject in real-world applications. For this task, previous methods use recurrent neural networks for modeling temporal relations in an input sequence. However, these methods overlook the fact that the input image sequence includes not only the action of interest but background and irrelevant actions. This would induce recurrent units to accumulate unnecessary information for encoding features on the action of interest. To overcome this problem, we propose a novel recurrent unit, named Information Discrimination Unit (IDU), which explicitly discriminates the information relevancy between an ongoing action and others to decide whether to accumulate the input information. This enables learning more discriminative representations for identifying an ongoing action. In this paper, we further present a new recurrent unit, called Information Integration Unit (IIU), for action anticipation. Our IIU exploits the outputs from IDN as pseudo action labels as well as RGB frames to learn enriched features of observed actions effectively. In experiments on TVSeries and THUMOS-14, the proposed methods outperform state-of-the-art methods by a significant margin in online action detection and action anticipation. Moreover, we demonstrate the effectiveness of the proposed units by conducting comprehensive ablation studies.

5.
Opt Lett ; 37(9): 1550-2, 2012 May 01.
Artigo em Inglês | MEDLINE | ID: mdl-22555734

RESUMO

We present a new approach for visual saliency detection from various natural images. It is inspired by our careful observation that the human visual system (HVS) responds sensitively and quickly to high textural contrast, derived from the discriminative directional pattern from its surroundings as well as the noticeable luminance difference, for understanding a given scene. By formulating such textural contrast within a multiscale framework, we construct a more reliable saliency map even without color information when compared to most previous approaches still suffering from the complex and cluttered background. The proposed method has been extensively tested on a wide range of natural images, and experimental results show that the proposed scheme is effective in detecting visual saliency compared to various state-of-the-art methods.


Assuntos
Biomimética/métodos , Fenômenos Ópticos , Humanos , Curva ROC , Fatores de Tempo , Percepção Visual
6.
Opt Lett ; 36(22): 4428-30, 2011 Nov 15.
Artigo em Inglês | MEDLINE | ID: mdl-22089586

RESUMO

We present a novel (to our best knowledge) optical recognition technique for detecting shadows from a single image. Most prior approaches definitely depend on explicit physical computational models, but physics-based approaches have the critical problem that they may fail severely even with slight perturbations. Unlike traditional approaches, our method does not rely on any explicit physical models. This breakthrough originates from a discovery of a new modeling mechanism, derived from a biological vision principle, the so-called retinex theory, which is well suited for single-image shadow detection. Experimental results demonstrate that the proposed method outperforms the previous optical recognition techniques and gives robust results even in real-world complex scenes.

7.
IEEE Trans Image Process ; 18(2): 401-11, 2009 Feb.
Artigo em Inglês | MEDLINE | ID: mdl-19095537

RESUMO

Overlay text brings important semantic clues in video content analysis such as video information retrieval and summarization, since the content of the scene or the editor's intention can be well represented by using inserted text. Most of the previous approaches to extracting overlay text from videos are based on low-level features, such as edge, color, and texture information. However, existing methods experience difficulties in handling texts with various contrasts or inserted in a complex background. In this paper, we propose a novel framework to detect and extract the overlay text from the video scene. Based on our observation that there exist transient colors between inserted text and its adjacent background, a transition map is first generated. Then candidate regions are extracted by a reshaping method and the overlay text regions are determined based on the occurrence of overlay text in each candidate. The detected overlay text regions are localized accurately using the projection of overlay text pixels in the transition map and the text extraction is finally conducted. The proposed method is robust to different character size, position, contrast, and color. It is also language independent. Overlay text region update between frames is also employed to reduce the processing time. Experiments are performed on diverse videos to confirm the efficiency of the proposed method.


Assuntos
Algoritmos , Documentação/métodos , Processamento Eletrônico de Dados/métodos , Interpretação de Imagem Assistida por Computador/métodos , Armazenamento e Recuperação da Informação/métodos , Reconhecimento Automatizado de Padrão/métodos , Gravação em Vídeo/métodos , Aumento da Imagem/métodos , Leitura , Reprodutibilidade dos Testes , Sensibilidade e Especificidade , Processamento de Sinais Assistido por Computador , Técnica de Subtração
8.
Artigo em Inglês | MEDLINE | ID: mdl-31425032

RESUMO

In this paper, we propose a blind text image deblurring algorithm by using a text-specific hybrid dictionary. After careful analysis, we find that the text-specific hybrid dictionary has the great ability of providing powerful contextual information for text image deblurring. Here, it is worth noting that our proposed method is inspired by our observation that an intermediate latent image contains not only sharp regions, but also multiple types of small blurred regions. Based upon our discovery, we propose a prior for text images based on sparse representation, which models the relationship between an intermediate latent image and a desired sharp image. To this end, we carefully collect three different image patch pairs, which are 1) Gaussian blur-sharp, 2) motion blur-sharp, and 3) sharp-sharp, in order to construct the text-specific hybrid dictionary. We also propose a new optimization framework suitable for the task of text image deblurring in this paper. Extensive experiments have been conducted on a challenging dataset of synthetic and real-world text images. Our results demonstrate that the proposed method outperforms the state-of-the-art image deblurring methods both quantitatively and qualitatively.

9.
IEEE Trans Vis Comput Graph ; 25(12): 3202-3215, 2019 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-30130231

RESUMO

In this paper, we present a novel grid encoding model for content-aware image retargeting. In contrast to previous approaches such as vertex-based and axis-aligned grid encoding models, our approach takes each horizontal/vertical distance between two adjacent vertices as an optimization variable. Upon this difference-based encoding scheme, every vertex position of a target grid is subsequently determined after optimizing the one-dimensional values. Our quad edge-based grid model has two major advantages for image retargeting. First, the model enables a grid optimization problem to be developed in a simple quadratic program while ensuring the global convexity of objective functions. Second, due to the independency of variables, spatial regularizations can be applied in a locally adaptive manner to preserve structural components. Based on this model, we propose three quadratic objective functions. Note that, in our work, their linear combination guides a grid deformation process to obtain a visually comfortable retargeting result by preserving salient regions and structural components of an input image. Comparative evaluations have been conducted with ten existing state-of-the-art image retargeting methods, and the results show that our method built upon the quad edge-based model consistently outperforms other previous methods both on qualitative and quantitative perspectives.

10.
Comput Methods Programs Biomed ; 165: 215-224, 2018 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-30337076

RESUMO

BACKGROUND AND OBJECTIVE: In pulmonary nodule detection, the first stage, candidate detection, aims to detect suspicious pulmonary nodules. However, detected candidates include many false positives and thus in the following stage, false positive reduction, such false positives are reliably reduced. Note that this task is challenging due to 1) the imbalance between the numbers of nodules and non-nodules and 2) the intra-class diversity of non-nodules. Although techniques using 3D convolutional neural networks (CNNs) have shown promising performance, they suffer from high computational complexity which hinders constructing deep networks. To efficiently address these problems, we propose a novel framework using the ensemble of 2D CNNs using single views, which outperforms existing 3D CNN-based methods. METHODS: Our ensemble of 2D CNNs utilizes single-view 2D patches to improve both computational and memory efficiency compared to previous techniques exploiting 3D CNNs. We first categorize non-nodules on the basis of features encoded by an autoencoder. Then, all 2D CNNs are trained by using the same nodule samples, but with different types of non-nodules. By extending the learning capability, this training scheme resolves difficulties of extracting representative features from non-nodules with large appearance variations. Note that, instead of manual categorization requiring the heavy workload of radiologists, we propose to automatically categorize non-nodules based on the autoencoder and k-means clustering. RESULTS: We performed extensive experiments to validate the effectiveness of our framework based on the database of the lung nodule analysis 2016 challenge. The superiority of our framework is demonstrated through comparing the performance of five frameworks trained with differently constructed training sets. Our proposed framework achieved state-of-the-art performance (0.922 of the competition performance metric score) with low computational demands (789K of parameters and 1024M of floating point operations per second). CONCLUSION: We presented a novel false positive reduction framework, the ensemble of single-view 2D CNNs with fully automatic non-nodule categorization, for pulmonary nodule detection. Unlike previous 3D CNN-based frameworks, we utilized 2D CNNs using 2D single views to improve computational efficiency. Also, our training scheme using categorized non-nodules, extends the learning capability of representative features of different non-nodules. Our framework achieved state-of-the-art performance with low computational complexity.


Assuntos
Diagnóstico por Computador/métodos , Neoplasias Pulmonares/diagnóstico por imagem , Redes Neurais de Computação , Nódulo Pulmonar Solitário/diagnóstico por imagem , Bases de Dados Factuais/estatística & dados numéricos , Diagnóstico por Computador/estatística & dados numéricos , Reações Falso-Positivas , Humanos , Neoplasias Pulmonares/classificação , Interpretação de Imagem Radiográfica Assistida por Computador/métodos , Interpretação de Imagem Radiográfica Assistida por Computador/estatística & dados numéricos , Nódulo Pulmonar Solitário/classificação , Tomografia Computadorizada por Raios X/estatística & dados numéricos
11.
IEEE Trans Image Process ; 14(10): 1503-11, 2005 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-16238056

RESUMO

We propose a novel algorithm to partition an image with low depth-of-field (DOF) into focused object-of-interest (OOI) and defocused background. The proposed algorithm unfolds into three steps. In the first step, we transform the low-DOF image into an appropriate feature space, in which the spatial distribution of the high-frequency components is represented. This is conducted by computing higher order statistics (HOS) for all pixels in the low-DOF image. Next, the obtained feature space, which is called HOS map in this paper, is simplified by removing small dark holes and bright patches using a morphological filter by reconstruction. Finally, the OOI is extracted by applying region merging to the simplified image and by thresholding. Unlike the previous methods that rely on sharp details of OOI only, the proposed algorithm complements the limitation of them by using morphological filters, which also allows perfect preservation of the contour information. Compared with the previous methods, the proposed method yields more accurate segmentation results, supporting faster processing.


Assuntos
Algoritmos , Filtração/métodos , Interpretação de Imagem Assistida por Computador/métodos , Imageamento Tridimensional/métodos , Reconhecimento Automatizado de Padrão/métodos , Técnica de Subtração , Análise por Conglomerados , Aumento da Imagem/métodos , Armazenamento e Recuperação da Informação/métodos , Análise Numérica Assistida por Computador , Processamento de Sinais Assistido por Computador
12.
IEEE Trans Image Process ; 22(4): 1667-73, 2013 Apr.
Artigo em Inglês | MEDLINE | ID: mdl-23221826

RESUMO

In this brief, we present a new indicator, i.e., salient edge energy, for guiding a given contour robustly and precisely toward the object boundary. Specifically, we define the salient edge energy by exploiting the higher order statistics on the diffusion space, and incorporate it into a variational level set formulation with the local region-based segmentation energy for solving the problem of curve evolution. In contrast to most previous methods, the proposed salient edge energy allows the curve to find only significant local minima relevant to the object boundary even in the noisy and cluttered background. Moreover, the segmentation performance derived from our new energy is less sensitive to the size of local windows compared with other recently developed methods, owing to the ability of our energy function to suppress diverse clutters. The proposed method has been tested on various images, and experimental results show that the salient edge energy effectively drives the active contour both qualitatively and quantitatively compared to various state-of-the-art methods.

13.
IEEE Trans Image Process ; 21(3): 1272-83, 2012 Mar.
Artigo em Inglês | MEDLINE | ID: mdl-21843989

RESUMO

In this paper, a visual attention model is incorporated for efficient saliency detection, and the salient regions are employed as object seeds for our automatic object segmentation system. In contrast with existing interactive segmentation approaches that require considerable user interaction, the proposed method does not require it, i.e., the segmentation task is fulfilled in a fully automatic manner. First, we introduce a novel unified spectral-domain approach for saliency detection. Our visual attention model originates from a well-known property of the human visual system that the human visual perception is highly adaptive and sensitive to structural information in images rather than nonstructural information. Then, based on the saliency map, we propose an iterative self-adaptive segmentation framework for more accurate object segmentation. Extensive tests on a variety of cluttered natural images show that the proposed algorithm is an efficient indicator for characterizing the human perception and it can provide satisfying segmentation performance.


Assuntos
Inteligência Artificial , Atenção , Reconhecimento Automatizado de Padrão/métodos , Percepção Visual , Algoritmos , Humanos
14.
IEEE Trans Biomed Eng ; 57(10): 2600-4, 2010 Oct.
Artigo em Inglês | MEDLINE | ID: mdl-20656653

RESUMO

In this letter, we present a novel watershed-based method for segmentation of cervical and breast cell images. We formulate the segmentation of clustered nuclei as an optimization problem. A hypothesis concerning the nuclei, which involves a priori knowledge with respect to the shape of nuclei, is tested to solve the optimization problem. We first apply the distance transform to the clustered nuclei. A marker extraction scheme based on the H-minima transform is introduced to obtain the optimal segmentation result from the distance map. In order to estimate the optimal h-value, a size-invariant segmentation distortion evaluation function is defined based on the fitting residuals between the segmented region boundaries and fitted models. Ellipsoidal modeling of contours is introduced to adjust nuclei contours for more effective analysis. Experiments on a variety of real microscopic cell images show that the proposed method yields more accurate segmentation results than the state-of-the-art watershed-based methods.


Assuntos
Algoritmos , Núcleo Celular/fisiologia , Processamento de Imagem Assistida por Computador/métodos , Microscopia/métodos , Neoplasias da Mama/patologia , Carcinoma Ductal de Mama/patologia , Núcleo Celular/ultraestrutura , Feminino , Humanos , Neoplasias do Colo do Útero/patologia
15.
IEEE Trans Biomed Eng ; 57(12): 2825-32, 2010 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-20656648

RESUMO

In a fully automatic cell extraction process, one of the main issues to overcome is the problem related to extracting overlapped nuclei since such nuclei will often affect the quantitative analysis of cell images. In this paper, we present an unsupervised Bayesian classification scheme for separating overlapped nuclei. The proposed approach first involves applying the distance transform to overlapped nuclei. The topographic surface generated by distance transform is viewed as a mixture of Gaussians in the proposed algorithm. In order to learn the distribution of the topographic surface, the parametric expectation-maximization (EM) algorithm is employed. Cluster validation is performed to determine how many nuclei are overlapped. Our segmentation approach incorporates a priori knowledge about the regular shape of clumped nuclei to yield more accurate segmentation results. Experimental results show that the proposed method yields superior segmentation performance, compared to those produced by conventional schemes.


Assuntos
Algoritmos , Teorema de Bayes , Núcleo Celular/ultraestrutura , Histocitoquímica/métodos , Processamento de Imagem Assistida por Computador/métodos , Mama/citologia , Neoplasias da Mama , Carcinoma Ductal de Mama , Colo do Útero/citologia , Análise por Conglomerados , Análise Discriminante , Feminino , Humanos , Distribuição Normal , Reprodutibilidade dos Testes
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA