Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 17 de 17
Filtrar
Más filtros













Base de datos
Intervalo de año de publicación
1.
Virol Sin ; 2024 Mar 29.
Artículo en Inglés | MEDLINE | ID: mdl-38556051

RESUMEN

The Ebola virus (EBOV) is a member of the Orthoebolavirus genus, Filoviridae family, which causes severe hemorrhagic diseases in humans and non-human primates (NHPs), with a case fatality rate of up to 90%. The development of countermeasures against EBOV has been hindered by the lack of ideal animal models, as EBOV requires handling in biosafety level (BSL)-4 facilities. Therefore, accessible and convenient animal models are urgently needed to promote prophylactic and therapeutic approaches against EBOV. In this study, a recombinant vesicular stomatitis virus expressing Ebola virus glycoprotein (VSV-EBOV/GP) was constructed and applied as a surrogate virus, establishing a lethal infection in hamsters. Following infection with VSV-EBOV/GP, 3-week-old female Syrian hamsters exhibited disease signs such as weight loss, multi-organ failure, severe uveitis, high viral loads, and developed severe systemic diseases similar to those observed in human EBOV patients. All animals succumbed at 2-3 days post-infection (dpi). Histopathological changes indicated that VSV-EBOV/GP targeted liver cells, suggesting that the tissue tropism of VSV-EBOV/GP was comparable to wild-type EBOV (WT EBOV). Notably, the pathogenicity of the VSV-EBOV/GP was found to be species-specific, age-related, gender-associated, and challenge route-dependent. Subsequently, equine anti-EBOV immunoglobulins and a subunit vaccine were validated using this model. Overall, this surrogate model represents a safe, effective, and economical tool for rapid preclinical evaluation of medical countermeasures against EBOV under BSL-2 conditions, which would accelerate technological advances and breakthroughs in confronting Ebola virus disease.

2.
Artículo en Inglés | MEDLINE | ID: mdl-38335084

RESUMEN

Multiview clustering (MVC) has gained significant attention as it enables the partitioning of samples into their respective categories through unsupervised learning. However, there are a few issues as follows: 1) many existing deep clustering methods use the same latent features to achieve the conflict objectives, namely, reconstruction and view consistency. The reconstruction objective aims to preserve view-specific features for each individual view, while the view-consistency objective strives to obtain common features across all views; 2) some deep embedded clustering (DEC) approaches adopt view-wise fusion to obtain consensus feature representation. However, these approaches overlook the correlation between samples, making it challenging to derive discriminative consensus representations; and 3) many methods use contrastive learning (CL) to align the view's representations; however, they do not take into account cluster information during the construction of sample pairs, which can lead to the presence of false negative pairs. To address these issues, we propose a novel multiview representation learning network, called anchor-sharing and clusterwise CL (CwCL) network for multiview representation learning. Specifically, we separate view-specific learning and view-common learning into different network branches, which addresses the conflict between reconstruction and consistency. Second, we design an anchor-sharing feature aggregation (ASFA) module, which learns the sharing anchors from different batch data samples, establishes the bipartite relationship between anchors and samples, and further leverages it to improve the samples' representations. This module enhances the discriminative power of the common representation from different samples. Third, we design CwCL module, which incorporates the learned transition probability into CL, allowing us to focus on minimizing the similarity between representations from negative pairs with a low transition probability. It alleviates the conflict in previous sample-level contrastive alignment. Experimental results demonstrate that our method outperforms the state-of-the-art performance.

3.
IEEE Trans Image Process ; 32: 3027-3039, 2023.
Artículo en Inglés | MEDLINE | ID: mdl-37192028

RESUMEN

In recent years, various neural network architectures for computer vision have been devised, such as the visual transformer and multilayer perceptron (MLP). A transformer based on an attention mechanism can outperform a traditional convolutional neural network. Compared with the convolutional neural network and transformer, the MLP introduces less inductive bias and achieves stronger generalization. In addition, a transformer shows an exponential increase in the inference, training, and debugging times. Considering a wave function representation, we propose the WaveNet architecture that adopts a novel vision task-oriented wavelet-based MLP for feature extraction to perform salient object detection in RGB (red-green-blue)-thermal infrared images. In addition, we apply knowledge distillation to a transformer as an advanced teacher network to acquire rich semantic and geometric information and guide WaveNet learning with this information. Following the shortest-path concept, we adopt the Kullback-Leibler distance as a regularization term for the RGB features to be as similar to the thermal infrared features as possible. The discrete wavelet transform allows for the examination of frequency-domain features in a local time domain and time-domain features in a local frequency domain. We apply this representation ability to perform cross-modality feature fusion. Specifically, we introduce a progressively cascaded sine-cosine module for cross-layer feature fusion and use low-level features to obtain clear boundaries of salient objects through the MLP. Results from extensive experiments indicate that the proposed WaveNet achieves impressive performance on benchmark RGB-thermal infrared datasets. The results and code are publicly available at https://github.com/nowander/WaveNet.

4.
Artículo en Inglés | MEDLINE | ID: mdl-37022901

RESUMEN

Most recent methods for RGB (red-green-blue)-thermal salient object detection (SOD) involve several floating-point operations and have numerous parameters, resulting in slow inference, especially on common processors, and impeding their deployment on mobile devices for practical applications. To address these problems, we propose a lightweight spatial boosting network (LSNet) for efficient RGB-thermal SOD with a lightweight MobileNetV2 backbone to replace a conventional backbone (e.g., VGG, ResNet). To improve feature extraction using a lightweight backbone, we propose a boundary boosting algorithm that optimizes the predicted saliency maps and reduces information collapse in low-dimensional features. The algorithm generates boundary maps based on predicted saliency maps without incurring additional calculations or complexity. As multimodality processing is essential for high-performance SOD, we adopt attentive feature distillation and selection and propose semantic and geometric transfer learning to enhance the backbone without increasing the complexity during testing. Experimental results demonstrate that the proposed LSNet achieves state-of-the-art performance compared with 14 RGB-thermal SOD methods on three datasets while improving the numbers of floating-point operations (1.025G) and parameters (5.39M), model size (22.1 MB), and inference speed (9.95 fps for PyTorch, batch size of 1, and Intel i5-7500 processor; 93.53 fps for PyTorch, batch size of 1, and NVIDIA TITAN V graphics processor; 936.68 fps for PyTorch, batch size of 20, and graphics processor; 538.01 fps for TensorRT and batch size of 1; and 903.01 fps for TensorRT/FP16 and batch size of 1). The code and results can be found from the link of https://github.com/zyrant/LSNet.

5.
Opt Express ; 31(5): 8029-8041, 2023 Feb 27.
Artículo en Inglés | MEDLINE | ID: mdl-36859921

RESUMEN

RGB-D indoor scene parsing is a challenging task in computer vision. Conventional scene-parsing approaches based on manual feature extraction have proved inadequate in this area because indoor scenes are both unordered and complex. This study proposes a feature adaptive selection, and fusion lightweight network (FASFLNet) for RGB-D indoor scene parsing that is both efficient and accurate. The proposed FASFLNet utilizes a lightweight classification network (MobileNetV2), constituting the backbone of the feature extraction. This lightweight backbone model guarantees that FASFLNet is not only highly efficient but also provides good performance in terms of feature extraction. The additional information provided by depth images (specifically, spatial information such as the shape and scale of objects) is used in FASFLNet as supplemental information for feature-level adaptive fusion between the RGB and depth streams. Furthermore, during decoding, the features of different layers are fused from top-bottom and integrated at different layers for final pixel-level classification, resulting in an effect similar to that of pyramid supervision. Experimental results obtained on the NYU V2 and SUN RGB-D datasets indicate that the proposed FASFLNet outperforms existing state-of-the-art models and is both highly efficient and accurate.

6.
Aging Clin Exp Res ; 35(3): 581-589, 2023 Mar.
Artículo en Inglés | MEDLINE | ID: mdl-36629995

RESUMEN

BACKGROUND: The relationship between the ratio of blood urea nitrogen to creatinine (BUN/Cr) and physical frailty in elderly patients remains unclear. The study aims to investigate the association between the BUN/Cr ratio and physical frailty in the elderly Chinese population. METHODS: In this cross-sectional analysis, the clinical data of 5213 participants from 2015 were selected from the China Health and Retirement Longitudinal Study (CHARLS). The demographic variables (including age and gender) and health behavior (including smoking and drinking history), anthropometric (including systolic and diastolic blood pressure, waist circumference (WC), etc.), physical performances (i.e., grip strength, repeated chair stands, etc.), and biochemical indicators (i.e., blood urea nitrogen (BUN), creatinine(Cr), total cholesterol (TC), triglycerides (TG), etc.) were measured. The association between the BUN/Cr ratio and physical frailty was analyzed. RESULTS: After adjusting for potential confounding factors, smooth curve fitting showed a linear relationship between the BUN/Cr ratio and grip strength, a non-linear relationship between the BUN/Cr ratio, and repeated chair-rising time. The fully adjusted linear regression results showed a negative association between the BUN/Cr ratio and grip strength. In the multivariate, piecewise linear regression, when the BUN/Cr ratio was greater than 18.60, the repeated chair-rising time increased with the increase in BUN/Cr ratio (ß = 0.046, 95%CI 0.025, 0.066; p < 0.001). However, we did not observe a significant correlation when the BUN/Cr ratio was less than 18.60 (ß = -0.007, 95%CI -0.046, 0.032; p = 0.717). CONCLUSION: This study demonstrated that the BUN/Cr ratio might be associated with physical frailty in older-aged Chinese, and this association requires further investigation.


Asunto(s)
Pueblos del Este de Asia , Fragilidad , Anciano , Humanos , Persona de Mediana Edad , Nitrógeno de la Urea Sanguínea , Estudios Transversales , Creatinina , Estudios Longitudinales , Biomarcadores
7.
Appl Opt ; 61(26): 7602-7607, 2022 Sep 10.
Artículo en Inglés | MEDLINE | ID: mdl-36256359

RESUMEN

Compared with monocular images, scene discrepancies between the left- and right-view images impose additional challenges on visual quality predictions in binocular images. Herein, we propose a hierarchical feature fusion network (HFFNet) for blind binocular image quality prediction that handles scene discrepancies and uses multilevel fusion features from the left- and right-view images to reflect distortions in binocular images. Specifically, a feature extraction network based on MobileNetV2 is used to determine the feature layers from distorted binocular images; then, low-level binocular fusion features (or middle-level and high-level binocular fusion features) are obtained by fusing the left and right low-level monocular features (or middle-level and high-level monocular features) using the feature gate module; further, three feature enhancement modules are used to enrich the information of the extracted features at different levels. Finally, the total feature maps obtained from the high-, middle-, and low-level fusion features are applied to a three-input feature fusion module for feature merging. Thus, the proposed HFFNet provides better results, to the best of our knowledge, than existing methods on two benchmark datasets.

8.
IEEE Trans Image Process ; 30: 7790-7802, 2021.
Artículo en Inglés | MEDLINE | ID: mdl-34495832

RESUMEN

Semantic segmentation is a fundamental task in computer vision, and it has various applications in fields such as robotic sensing, video surveillance, and autonomous driving. A major research topic in urban road semantic segmentation is the proper integration and use of cross-modal information for fusion. Here, we attempt to leverage inherent multimodal information and acquire graded features to develop a novel multilabel-learning network for RGB-thermal urban scene semantic segmentation. Specifically, we propose a strategy for graded-feature extraction to split multilevel features into junior, intermediate, and senior levels. Then, we integrate RGB and thermal modalities with two distinct fusion modules, namely a shallow feature fusion module and deep feature fusion module for junior and senior features. Finally, we use multilabel supervision to optimize the network in terms of semantic, binary, and boundary characteristics. Experimental results confirm that the proposed architecture, the graded-feature multilabel-learning network, outperforms state-of-the-art methods for urban scene semantic segmentation, and it can be generalized to depth data.

9.
Artículo en Inglés | MEDLINE | ID: mdl-34415839

RESUMEN

Using attention mechanisms in saliency detection networks enables effective feature extraction, and using linear methods can promote proper feature fusion, as verified in numerous existing models. Current networks usually combine depth maps with red-green-blue (RGB) images for salient object detection (SOD). However, fully leveraging depth information complementary to RGB information by accurately highlighting salient objects deserves further study. We combine a gated attention mechanism and a linear fusion method to construct a dual-stream interactive recursive feature-reshaping network (IRFR-Net). The streams for RGB and depth data communicate through a backbone encoder to thoroughly extract complementary information. First, we design a context extraction module (CEM) to obtain low-level depth foreground information. Subsequently, the gated attention fusion module (GAFM) is applied to the RGB depth (RGB-D) information to obtain advantageous structural and spatial fusion features. Then, adjacent depth information is globally integrated to obtain complementary context features. We also introduce a weighted atrous spatial pyramid pooling (WASPP) module to extract the multiscale local information of depth features. Finally, global and local features are fused in a bottom-up scheme to effectively highlight salient objects. Comprehensive experiments on eight representative datasets demonstrate that the proposed IRFR-Net outperforms 11 state-of-the-art (SOTA) RGB-D approaches in various evaluation indicators.

10.
Comput Intell Neurosci ; 2021: 6610997, 2021.
Artículo en Inglés | MEDLINE | ID: mdl-34035801

RESUMEN

In recent years, the prediction of salient regions in RGB-D images has become a focus of research. Compared to its RGB counterpart, the saliency prediction of RGB-D images is more challenging. In this study, we propose a novel deep multimodal fusion autoencoder for the saliency prediction of RGB-D images. The core trainable autoencoder of the RGB-D saliency prediction model employs two raw modalities (RGB and depth/disparity information) as inputs and their corresponding eye-fixation attributes as labels. The autoencoder comprises four main networks: color channel network, disparity channel network, feature concatenated network, and feature learning network. The autoencoder can mine the complex relationship and make the utmost of the complementary characteristics between both color and disparity cues. Finally, the saliency map is predicted via a feature combination subnetwork, which combines the deep features extracted from a prior learning and convolutional feature learning subnetworks. We compare the proposed autoencoder with other saliency prediction models on two publicly available benchmark datasets. The results demonstrate that the proposed autoencoder outperforms these models by a significant margin.


Asunto(s)
Fijación Ocular
11.
Comput Intell Neurosci ; 2020: 8841681, 2020.
Artículo en Inglés | MEDLINE | ID: mdl-33293945

RESUMEN

Visual saliency prediction for RGB-D images is more challenging than that for their RGB counterparts. Additionally, very few investigations have been undertaken concerning RGB-D-saliency prediction. The proposed study presents a method based on a hierarchical multimodal adaptive fusion (HMAF) network to facilitate end-to-end prediction of RGB-D saliency. In the proposed method, hierarchical (multilevel) multimodal features are first extracted from an RGB image and depth map using a VGG-16-based two-stream network. Subsequently, the most significant hierarchical features of the said RGB image and depth map are predicted using three two-input attention modules. Furthermore, adaptive fusion of saliencies concerning the above-mentioned fused saliency features of different levels (hierarchical fusion saliency features) can be accomplished using a three-input attention module to facilitate high-accuracy RGB-D visual saliency prediction. Comparisons based on the application of the proposed HMAF-based approach against those of other state-of-the-art techniques on two challenging RGB-D datasets demonstrate that the proposed method outperforms other competing approaches consistently by a considerable margin.


Asunto(s)
Atención
12.
Opt Express ; 27(23): 34056-34066, 2019 Nov 11.
Artículo en Inglés | MEDLINE | ID: mdl-31878462

RESUMEN

Human eye-fixation prediction in 3D images is important for many 3D applications, such as fine-grained 3D video object segmentation and intelligent bulletproof curtains. While the vast majority of existing 2D-based approaches cannot be applied, the main challenge lies in the inconsistency, or even conflict, between the RGB and depth saliency maps. In this paper, we propose a three-stream architecture to accurately predict human visual attention on 3D images end-to-end. First, a two-stream feature extraction network based on advanced convolutional neural networks is trained for RGB and depth, and hierarchical information is extracted from each ResNet-18. Then, these multi-level features are fed into the channel attention mechanism to suppress the feature space inconsistency and make the network focus on a significant target. The enhanced saliency map is fused step-by-step by VGG-16 to generate the final coarse saliency map. Finally, each coarse map is refined empirically through refinement blocks, and the network's own identification errors are corrected based on the acquired knowledge, thus converting the prediction saliency map from coarse to fine. The results of comparison of our model with six other state-of-the-art approaches on the NUS dataset (CC of 0.5579, KLDiv of 1.0903, AUC of 0.8339, and NSS of 2.3373) and the NCTU dataset (CC of 0.8614, KLDiv of 0.2681, AUC of 0.9143, and NSS of 2.3795) indicate that the proposed model consistently outperforms them by a considerable margin as it fully employs the channel attention mechanism.


Asunto(s)
Atención/fisiología , Fijación Ocular/fisiología , Imagenología Tridimensional , Algoritmos , Bases de Datos como Asunto , Humanos
13.
IEEE Trans Image Process ; 27(5): 2086-2095, 2018 May.
Artículo en Inglés | MEDLINE | ID: mdl-29432092

RESUMEN

The blind quality evaluation of screen content images (SCIs) and natural scene images (NSIs) has become an important, yet very challenging issue. In this paper, we present an effective blind quality evaluation technique for SCIs and NSIs based on a dictionary of learned local and global quality features. First, a local dictionary is constructed using local normalized image patches and conventional -means clustering. With this local dictionary, the learned local quality features can be obtained using a locality-constrained linear coding with max pooling. To extract the learned global quality features, the histogram representations of binary patterns are concatenated to form a global dictionary. The collaborative representation algorithm is used to efficiently code the learned global quality features of the distorted images using this dictionary. Finally, kernel-based support vector regression is used to integrate these features into an overall quality score. Extensive experiments involving the proposed evaluation technique demonstrate that in comparison with most relevant metrics, the proposed blind metric yields significantly higher consistency in line with subjective fidelity ratings.

15.
Sci Rep ; 6: 33819, 2016 09 26.
Artículo en Inglés | MEDLINE | ID: mdl-27666087

RESUMEN

Optical coherence tomography (OCT) has been applied to inspect the internal defect of beadless Chinese ZhuJi fleshwater pearls. A novel fully automated algorithm is proposed to classify between normal and defective sub-layer in nacre layer. Our algorithm utilizes the graph segmentation approach to estimate the up and down boundaries of defect sub-layers from flattened and cropped image, and also proposes the strategy for edge and weight construction in segmentation process. The vertical gradients of boundary pixels are used to make grading decision. The algorithm is tested by typical pearl samples, and achieves 100% classification accuracy. The experiment result shows the feasibility and adaptability of the proposed approach, and proves that the OCT technique combined with proposed algorithm is a potential tool for fast and non-destructive diagnosis of internal structure of beadless pearl.

16.
Appl Opt ; 54(21): 6549-57, 2015 Jul 20.
Artículo en Inglés | MEDLINE | ID: mdl-26367842

RESUMEN

Three-dimensional (3D) technology has become immensely popular in recent years and widely adopted in various applications. Hence, perceptual quality measurement of symmetrically and asymmetrically distorted 3D images has become an important, fundamental, and challenging issue in 3D imaging research. In this paper, we propose a binocular-vision-based 3D image-quality measurement (IQM) metric. Consideration of the 3D perceptual properties of the primary visual cortex (V1) and the higher visual areas (V2) for 3D-IQM is the major technical contribution to this research. To be more specific, first, the metric simulates the receptive fields of complex cells (V1) using binocular energy response and binocular rivalry response and the higher visual areas (V2) using local binary patterns features. Then, three similarity scores of 3D perceptual properties between the reference and distorted 3D images are measured. Finally, by using support vector regression, three similarity scores are integrated into an overall 3D quality score. Experimental results for two public benchmark databases demonstrate that, in comparison with most current 2D and 3D metrics, the proposed metric achieves significantly higher consistency in alignment with subjective fidelity ratings.

17.
Opt Express ; 23(18): 23710-25, 2015 Sep 07.
Artículo en Inglés | MEDLINE | ID: mdl-26368467

RESUMEN

Perceptual quality measurement of three-dimensional (3D) visual signals has become a fundamental challenge in 3D imaging fields. This paper proposes a novel no-reference (NR) 3D visual quality measurement (VQM) metric that uses simulations of the primary visual cortex (V1) of binocular vision. As the major technical contribution of this study, perceptual properties of simple and complex cells are considered for NR 3D-VQM. More specifically, the metric simulates the receptive fields of simple cells (one class of V1 neurons) using Gaussian derivative functions, and the receptive fields of complex cells (the other class of V1 neurons) using disparity energy responses and binocular rivalry responses. Subsequently, various quality-aware features are extracted from the primary visual cortex; these will change in the presence of distortions. Finally, those features are mapped to the subjective quality score of the distorted 3D visual signal by using support vector regression (SVR). Experiments on two publicly available 3D databases confirm the effectiveness of our proposed metric, compared to the relevant full-reference (FR) and NR metrics.

SELECCIÓN DE REFERENCIAS
DETALLE DE LA BÚSQUEDA