Your browser doesn't support javascript.
loading
Self-architectural knowledge distillation for spiking neural networks.
Qiu, Haonan; Ning, Munan; Song, Zeyin; Fang, Wei; Chen, Yanqi; Sun, Tao; Ma, Zhengyu; Yuan, Li; Tian, Yonghong.
Afiliação
  • Qiu H; Peking University, School of Electronic and Computer Engineering, Shenzhen Graduate School, China. Electronic address: qiuhaonan@stu.pku.edu.cn.
  • Ning M; Peking University, School of Electronic and Computer Engineering, Shenzhen Graduate School, China.
  • Song Z; Peking University, School of Electronic and Computer Engineering, Shenzhen Graduate School, China.
  • Fang W; Peking University, School of Computer Science, China; PengCheng Laboratory, China.
  • Chen Y; Peking University, School of Computer Science, China; PengCheng Laboratory, China.
  • Sun T; Peking University, School of Electronic and Computer Engineering, Shenzhen Graduate School, China.
  • Ma Z; PengCheng Laboratory, China. Electronic address: mazhy@pcl.ac.cn.
  • Yuan L; Peking University, School of Electronic and Computer Engineering, Shenzhen Graduate School, China; PengCheng Laboratory, China. Electronic address: yuanli-ece@pku.edu.cn.
  • Tian Y; Peking University, School of Electronic and Computer Engineering, Shenzhen Graduate School, China; Peking University, School of Computer Science, China; PengCheng Laboratory, China. Electronic address: yhtian@pku.edu.cn.
Neural Netw ; 178: 106475, 2024 Oct.
Article em En | MEDLINE | ID: mdl-38941738
ABSTRACT
Spiking neural networks (SNNs) have attracted attention due to their biological plausibility and the potential for low-energy applications on neuromorphic hardware. Two mainstream approaches are commonly used to obtain SNNs, i.e., ANN-to-SNN conversion methods, and Directly-trained-SNN methods. However, the former achieve excellent performance at the cost of a large number of time steps (i.e., latency), while the latter exhibit lower latency but suffers from suboptimal performance. To tackle the performance-latency trade-off, we propose Self-Architectural Knowledge Distillation (SAKD), an intuitive and effective method for SNNs leveraging Knowledge Distillation (KD). We adopt a bilevel teacher-student training strategy in SAKD, i.e., level-1 involves directly transferring same-architectural pre-trained ANN weights to SNNs, and level-2 encourages the SNNs to mimic ANN's behavior, considering both final responses and intermediate features aspects. Learning with informative supervision signals fostered by labels and ANNs, our SAKD achieves new state-of-the-art (SOTA) performance with a few time steps on widely-used classification benchmark datasets. On ImageNet-1K, with only 4 time steps, our Spiking-ResNet34 model attains a Top-1 accuracy of 70.04%, outperforming the previous same-architectural SOTA methods. Notably, our SEW-ResNet152 model reaches a Top-1 accuracy of 77.30% on ImageNet-1K, setting a new SOTA benchmark for SNNs. Furthermore, we apply our SAKD to various dense prediction downstream tasks, such as object detection and semantic segmentation, demonstrating strong generalization ability and superior performance. In conclusion, our proposed SAKD framework presents a promising approach for achieving both high performance and low latency in SNNs, potentially paving the way for future advancements in the field.
Assuntos
Palavras-chave

Texto completo: 1 Base de dados: MEDLINE Assunto principal: Redes Neurais de Computação Limite: Humans Idioma: En Ano de publicação: 2024 Tipo de documento: Article

Texto completo: 1 Base de dados: MEDLINE Assunto principal: Redes Neurais de Computação Limite: Humans Idioma: En Ano de publicação: 2024 Tipo de documento: Article