Your browser doesn't support javascript.
loading
Mostrar: 20 | 50 | 100
Resultados 1 - 7 de 7
Filtrar
Mais filtros

Base de dados
Tipo de documento
País de afiliação
Intervalo de ano de publicação
1.
J Chem Inf Model ; 62(14): 3281-3290, 2022 Jul 25.
Artigo em Inglês | MEDLINE | ID: mdl-35574760

RESUMO

Metal-organic frameworks (MOFs) have become an active topic because of their excellent carbon capture and storage (CCS) properties. However, it is quite challenging to identify MOFs with superior performance within a massive combinatorial search space. To this end, we propose a deep-learning-based end-to-end prediction model to rapidly and accurately predict the CO2 working capacity and CO2/N2 selectivity of a given MOF under low-pressure conditions. Different from previous methods, our prediction model relies only on the data from the Crystallographic Information File (CIF) rather than handcrafted geometric descriptors and chemical descriptors. The model was developed, trained, and tested on a dataset of 342489 topologically diverse MOFs. Experimental results on the dataset show that the proposed model achieves high prediction performance, i.e., R2 = 0.916 for predicting the CO2 working capacity and R2 = 0.911 for predicting the CO2/N2 selectivity. With regard to the identification of potential high-performing MOFs, 1020 of 1027 (top 3%) high-performance MOFs were recovered while screening only 12% of the entire dataset using our provided pretrained model, reducing the computation time by nearly an order of magnitude when the model was used to prescreen material prior to computationally intensive grand canonical Monte Carlo (GCMC) simulations while still capturing 99% of the high-performance MOFs. In the ab initio training task, the method can achieve R2 = 0.85 with only 20% of the labeled data used for training and recover 995 of 1027 (top 3%) high-performance MOFs with only 12% of the entire dataset screened.

2.
Entropy (Basel) ; 25(1)2022 Dec 22.
Artigo em Inglês | MEDLINE | ID: mdl-36673161

RESUMO

Cross-modal human pose estimation has a wide range of applications. Traditional image-based pose estimation will not work well in poor light or darkness. Therefore, some sensors such as LiDAR or Radio Frequency (RF) signals are now using to estimate human pose. However, it limits the application that these methods require much high-priced professional equipment. To address these challenges, we propose a new WiFi-based pose estimation method. Based on the Channel State Information (CSI) of WiFi, a novel architecture CSI-former is proposed to innovatively realize the integration of the multi-head attention in the WiFi-based pose estimation network. To evaluate the performance of CSI-former, we establish a span-new dataset Wi-Pose. This dataset consists of 5 GHz WiFi CSI, the corresponding images, and skeleton point annotations. The experimental results on Wi-Pose demonstrate that CSI-former can significantly improve the performance in wireless pose estimation and achieve more remarkable performance over traditional image-based pose estimation. To better benefit future research on the WiFi-based pose estimation, Wi-Pose has been made publicly available.

3.
Entropy (Basel) ; 24(6)2022 May 31.
Artigo em Inglês | MEDLINE | ID: mdl-35741496

RESUMO

Recently, deep neural network-based image compressed sensing methods have achieved impressive success in reconstruction quality. However, these methods (1) have limitations in sampling pattern and (2) usually have the disadvantage of high computational complexity. To this end, a fast multi-scale generative adversarial network (FMSGAN) is implemented in this paper. Specifically, (1) an effective multi-scale sampling structure is proposed. It contains four different kernels with varying sizes so that decompose, and sample images effectively, which is capable of capturing different levels of spatial features at multiple scales. (2) An efficient lightweight multi-scale residual structure for deep image reconstruction is proposed to balance receptive field size and computational complexity. The key idea is to apply smaller convolution kernel sizes in the multi-scale residual structure to reduce the number of operations while maintaining the receptive field. Meanwhile, the channel attention structure is employed for enriching useful information. Moreover, perceptual loss is combined with MSE loss and adversarial loss as the optimization function to recover a finer image. Numerous experiments show that our FMSGAN achieves state-of-the-art image reconstruction quality with low computational complexity.

4.
Sensors (Basel) ; 20(18)2020 Sep 15.
Artigo em Inglês | MEDLINE | ID: mdl-32942579

RESUMO

In the skeleton-based human action recognition domain, the spatial-temporal graph convolution networks (ST-GCNs) have made great progress recently. However, they use only one fixed temporal convolution kernel, which is not enough to extract the temporal cues comprehensively. Moreover, simply connecting the spatial graph convolution layer (GCL) and the temporal GCL in series is not the optimal solution. To this end, we propose a novel enhanced spatial and extended temporal graph convolutional network (EE-GCN) in this paper. Three convolution kernels with different sizes are chosen to extract the discriminative temporal features from shorter to longer terms. The corresponding GCLs are then concatenated by a powerful yet efficient one-shot aggregation (OSA) + effective squeeze-excitation (eSE) structure. The OSA module aggregates the features from each layer once to the output, and the eSE module explores the interdependency between the channels of the output. Besides, we propose a new connection paradigm to enhance the spatial features, which expand the serial connection to a combination of serial and parallel connections by adding a spatial GCL in parallel with the temporal GCLs. The proposed method is evaluated on three large scale datasets, and the experimental results show that the performance of our method exceeds previous state-of-the-art methods.


Assuntos
Algoritmos , Movimento , Redes Neurais de Computação , Esqueleto/fisiologia , Humanos
5.
Artigo em Inglês | MEDLINE | ID: mdl-38416620

RESUMO

Text-based person retrieval is the process of searching a massive visual resource library for images of a particular pedestrian, based on a textual query. Existing approaches often suffer from a problem of color (CLR) over-reliance, which can result in a suboptimal person retrieval performance by distracting the model from other important visual cues such as texture and structure information. To handle this problem, we propose a novel framework to Excavate All-round Information Beyond Color for the task of text-based person retrieval, which is therefore termed EAIBC. The EAIBC architecture includes four branches, namely an RGB branch, a grayscale (GRS) branch, a high-frequency (HFQ) branch, and a CLR branch. Furthermore, we introduce a mutual learning (ML) mechanism to facilitate communication and learning among the branches, enabling them to take full advantage of all-round information in an effective and balanced manner. We evaluate the proposed method on three benchmark datasets, including CUHK-PEDES, ICFG-PEDES, and RSTPReid. The experimental results demonstrate that EAIBC significantly outperforms existing methods and achieves state-of-the-art (SOTA) performance in supervised, weakly supervised, and cross-domain settings.

6.
IEEE Trans Neural Netw Learn Syst ; 34(12): 10589-10599, 2023 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-35522636

RESUMO

Modeling the spatiotemporal relationship (STR) of traffic data is important yet challenging for existing graph networks. These methods usually capture features separately in temporal and spatial dimensions or represent the spatiotemporal data by adopting multiple local spatial-temporal graphs. The first kind of method mentioned above is difficult to capture potential temporal-spatial relationships, while the other is limited for long-term feature extraction due to its local receptive field. To handle these issues, the Synchronous Spatio-Temporal grAph Transformer (S2TAT) network is proposed for efficiently modeling the traffic data. The contributions of our method include the following: 1) the nonlocal STR can be synchronously modeled by our integrated attention mechanism and graph convolution in the proposed S2TAT block; 2) the timewise graph convolution and multihead mechanism designed can handle the heterogeneity of data; and 3) we introduce a novel attention-based strategy in the output module, being able to capture more valuable historical information to overcome the shortcoming of conventional average aggregation. Extensive experiments are conducted on PeMS datasets that demonstrate the efficacy of the S2TAT by achieving a top-one accuracy but less computational cost by comparing with the state of the art.

7.
IEEE Trans Cybern ; 51(12): 6017-6028, 2021 Dec.
Artigo em Inglês | MEDLINE | ID: mdl-32011279

RESUMO

Generating action proposals in untrimmed videos is a challenging task, since video sequences usually contain lots of irrelevant contents and the duration of an action instance is arbitrary. The quality of action proposals is key to action detection performance. The previous methods mainly rely on sliding windows or anchor boxes to cover all ground-truth actions, but this is infeasible and computationally inefficient. To this end, this article proposes a RecapNet-a novel framework for generating action proposal, by mimicking the human cognitive process of understanding video content. Specifically, this RecapNet includes a residual causal convolution module to build a short memory of the past events, based on which the joint probability actionness density ranking mechanism is designed to retrieve the action proposals. The RecapNet can handle videos with arbitrary length and more important, a video sequence will need to be processed only in one single pass in order to generate all action proposals. The experiments show that the proposed RecapNet outperforms the state of the art under all metrics on the benchmark THUMOS14 and ActivityNet-1.3 datasets. The code is available publicly at https://github.com/tianwangbuaa/RecapNet.


Assuntos
Algoritmos , Cognição , Humanos
SELEÇÃO DE REFERÊNCIAS
DETALHE DA PESQUISA