Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 7 de 7
Filter
Add more filters










Database
Language
Publication year range
1.
Article in English | MEDLINE | ID: mdl-38889037

ABSTRACT

Channel pruning is attracting increasing attention in the deep model compression community due to its capability of significantly reducing computation and memory footprints without special support from specific software and hardware. A challenge of channel pruning is designing efficient and effective criteria to select channels to prune. A widely used criterion is minimal performance degeneration, e.g., loss changes before and after pruning being the smallest. To accurately evaluate the truth performance degeneration requires retraining the survived weights to convergence, which is prohibitively slow. Hence existing pruning methods settle to use previous weights (without retraining) to evaluate the performance degeneration. However, we observe that the loss changes differ significantly with and without retraining. It motivates us to develop a technique to evaluate true loss changes without retraining, using which to select channels to prune with more reliability and confidence. We first derive a closed-form estimator of the true loss change per mask change, using influence functions without retraining. Influence function is a classic technique from robust statistics that reveals the impacts of a training sample on the model's prediction and is repurposed by us to assess impacts on true loss changes. We then show how to assess the importance of all channels simultaneously and develop a novel global channel pruning algorithm accordingly. We conduct extensive experiments to verify the effectiveness of the proposed algorithm, which significantly outperforms the competing channel pruning methods on both image classification and object detection tasks. One of the attractive properties of our algorithm is that it automatically obtains the prune percentage without the cumbersome yet commonly used sensitivity analysis by local pruning. To the best of our knowledge, we are the first that shows evaluating true loss changes for pruning without retraining is possible. This finding will open up opportunities for a series of new paradigms to emerge that differ from existing pruning methods. The code is available at https://github.com/hrcheng1066/IFSO.

2.
J Phys Chem Lett ; 14(49): 11058-11062, 2023 Dec 14.
Article in English | MEDLINE | ID: mdl-38048178

ABSTRACT

Single-atom catalysts (SACs) offer significant potential across various applications, yet our understanding of their formation mechanism remains limited. Notably, the pyrolysis of zeolitic imidazolate frameworks (ZIFs) stands as a pivotal avenue for SAC synthesis, of which the mechanism can be assessed through infrared (IR) spectroscopy. However, the prevailing analysis techniques still rely on manual interpretation. Here, we report a machine learning (ML)-driven analysis of the IR spectroscopy to unravel the pyrolysis process of Pt-doped ZIF-67 to synthesize Pt-Co3O4 SAC. Demonstrating a total Pearson correlation exceeding 0.7 with experimental data, the algorithm provides correlation coefficients for the selected structures, thereby confirming crucial structural changes with time and temperature, including the decomposition of ZIF and formation of Pt-O bonds. These findings reveal and confirm the formation mechanism of SACs. As demonstrated, the integration of ML algorithms, theoretical simulations, and experimental spectral analysis introduces an approach to deciphering experimental characterization data, implying its potential for broader adoption.

3.
Nano Lett ; 23(21): 9796-9802, 2023 Nov 08.
Article in English | MEDLINE | ID: mdl-37890870

ABSTRACT

Despite today's commercial-scale graphene production using chemical vapor deposition (CVD), the growth of high-quality single-layer graphene with controlled morphology and crystallinity remains challenging. Considerable effort is still spent on designing improved CVD catalysts for producing high-quality graphene. Conventionally, however, catalyst design has been pursued using empirical intuition or trial-and-error approaches. Here, we combine high-throughput density functional theory and machine learning to identify new prospective transition metal alloy catalysts that exhibit performance comparable to that of established graphene catalysts, such as Ni(111) and Cu(111). The alloys identified through this process generally consist of combinations of early- and late-transition metals, and a majority are alloys of Ni or Cu. Nevertheless, in many cases, these conventional catalyst metals are combined with unconventional partners, such as Zr, Hf, and Nb. The approach presented here therefore highlights an important new approach for identifying novel catalyst materials for the CVD growth of low-dimensional nanomaterials.

4.
Plant Methods ; 19(1): 96, 2023 Sep 02.
Article in English | MEDLINE | ID: mdl-37660084

ABSTRACT

BACKGROUND: Genomic prediction has become a powerful modelling tool for assessing line performance in plant and livestock breeding programmes. Among the genomic prediction modelling approaches, linear based models have proven to provide accurate predictions even when the number of genetic markers exceeds the number of data samples. However, breeding programmes are now compiling data from large numbers of lines and test environments for analyses, rendering these approaches computationally prohibitive. Machine learning (ML) now offers a solution to this problem through the construction of fully connected deep learning architectures and high parallelisation of the predictive task. However, the fully connected nature of these architectures immediately generates an over-parameterisation of the network that needs addressing for efficient and accurate predictions. RESULTS: In this research we explore the use of an ML architecture governed by variational Bayesian sparsity in its initial layers that we have called VBS-ML. The use of VBS-ML provides a mechanism for feature selection of important markers linked to the trait, immediately reducing the network over-parameterisation. Selected markers then propagate to the remaining fully connected feed-forward components of the ML network to form the final genomic prediction. We illustrated the approach with four large Australian wheat breeding data sets that range from 2665 lines to 10375 lines genotyped across a large set of markers. For all data sets, the use of the VBS-ML architecture improved genomic prediction accuracy over legacy linear based modelling approaches. CONCLUSIONS: An ML architecture governed under a variational Bayesian paradigm was shown to improve genomic prediction accuracy over legacy modelling approaches. This VBS-ML approach can be used to dramatically decrease the parameter burden on the network and provide a computationally feasible approach for improving genomic prediction conducted with large breeding population numbers and genetic markers.

5.
IEEE Trans Pattern Anal Mach Intell ; 45(10): 11898-11914, 2023 10.
Article in English | MEDLINE | ID: mdl-37247321

ABSTRACT

Compared with the progress made on human activity classification, much less success has been achieved on human interaction understanding (HIU). Apart from the latter task is much more challenging, the main causation is that recent approaches learn human interactive relations via shallow graphical representations, which are inadequate to model complicated human interactive-relations. This paper proposes a deep consistency-aware framework aiming at tackling the grouping and labelling inconsistencies in HIU. This framework consists of three components, including a backbone CNN to extract image features, a factor graph network to implicitly learn higher-order consistencies among labelling and grouping variables, and a consistency-aware reasoning module to explicitly enforcing consistencies. The last module is inspired by our key observation that the consistency-aware reasoning bias can be embedded into an energy function or a particular loss function, minimizing which delivers consistent predictions. An efficient mean-field inference algorithm is proposed, such that all modules of our network could be trained in an end-to-end fashion. Experimental results demonstrate that the two proposed consistency-learning modules complement each other, and both make considerable contributions in achieving leading performance on three benchmarks of HIU. The effectiveness of the proposed approach is further validated by experiments on detecting human-object interactions.


Subject(s)
Algorithms , Learning , Humans , Benchmarking
6.
IEEE Trans Image Process ; 32: 2857-2866, 2023.
Article in English | MEDLINE | ID: mdl-37186531

ABSTRACT

The goal of dynamic scene deblurring is to remove the motion blur presented in a given image. To recover the details from the severe blurs, conventional convolutional neural networks (CNNs) based methods typically increase the number of convolution layers, kernel-size, or different scale images to enlarge the receptive field. However, these methods neglect the non-uniform nature of blurs, and cannot extract varied local and global information. Unlike the CNNs-based methods, we propose a Transformer-based model for image deblurring, named SharpFormer, that directly learns long-range dependencies via a novel Transformer module to overcome large blur variations. Transformer is good at learning global information but is poor at capturing local information. To overcome this issue, we design a novel Locality preserving Transformer (LTransformer) block to integrate sufficient local information into global features. In addition, to effectively apply LTransformer to the medium-resolution features, a hybrid block is introduced to capture intermediate mixed features. Furthermore, we use a dynamic convolution (DyConv) block, which aggregates multiple parallel convolution kernels to handle the non-uniform blur of inputs. We leverage a powerful two-stage attentive framework composed of the above blocks to learn the global, hybrid, and local features effectively. Extensive experiments on the GoPro and REDS datasets show that the proposed SharpFormer performs favourably against the state-of-the-art methods in blurred image restoration.

7.
IEEE Trans Pattern Anal Mach Intell ; 44(12): 9011-9025, 2022 12.
Article in English | MEDLINE | ID: mdl-34705634

ABSTRACT

This paper addresses the task of set prediction using deep feed-forward neural networks. A set is a collection of elements which is invariant under permutation and the size of a set is not fixed in advance. Many real-world problems, such as image tagging and object detection, have outputs that are naturally expressed as sets of entities. This creates a challenge for traditional deep neural networks which naturally deal with structured outputs such as vectors, matrices or tensors. We present a novel approach for learning to predict sets with unknown permutation and cardinality using deep neural networks. In our formulation we define a likelihood for a set distribution represented by a) two discrete distributions defining the set cardinally and permutation variables, and b) a joint distribution over set elements with a fixed cardinality. Depending on the problem under consideration, we define different training models for set prediction using deep neural networks. We demonstrate the validity of our set formulations on relevant vision problems such as: 1) multi-label image classification where we outperform the other competing methods on the PASCAL VOC and MS COCO datasets, 2) object detection, for which our formulation outperforms popular state-of-the-art detectors, and 3) a complex CAPTCHA test, where we observe that, surprisingly, our set-based network acquired the ability of mimicking arithmetics without any rules being coded.


Subject(s)
Algorithms , Neural Networks, Computer , Machine Learning
SELECTION OF CITATIONS
SEARCH DETAIL
...