Your browser doesn't support javascript.
loading
Efficient Layer-Wise N:M Sparse CNN Accelerator with Flexible SPEC: Sparse Processing Element Clusters.
Xie, Xiaoru; Zhu, Mingyu; Lu, Siyuan; Wang, Zhongfeng.
Afiliação
  • Xie X; School of Electronic Science and Engineering, Nanjing University, Nanjing 210023, China.
  • Zhu M; School of Electronic Science and Engineering, Nanjing University, Nanjing 210023, China.
  • Lu S; School of Electronic Science and Engineering, Nanjing University, Nanjing 210023, China.
  • Wang Z; School of Electronic Science and Engineering, Nanjing University, Nanjing 210023, China.
Micromachines (Basel) ; 14(3)2023 Feb 24.
Article em En | MEDLINE | ID: mdl-36984936
ABSTRACT
Recently, the layer-wise NM fine-grained sparse neural network algorithm (i.e., every M-weights contains N non-zero values) has attracted tremendous attention, as it can effectively reduce the computational complexity with negligible accuracy loss. However, the speed-up potential of this algorithm will not be fully exploited if the right hardware support is lacking. In this work, we design an efficient accelerator for the NM sparse convolutional neural networks (CNNs) with layer-wise sparse patterns. First, we analyze the performances of different processing element (PE) structures and extensions to construct the flexible PE architecture. Second, the variable sparse convolutional dimensions and sparse ratios are involved in the hardware design. With a sparse PE cluster (SPEC) design, the hardware can efficiently accelerate CNNs with the layer-wise NM pattern. Finally, we employ the proposed SPEC into the CNN accelerator with flexible network-on-chip and specially designed dataflow. We implement hardware accelerators on Xilinx ZCU102 FPGA and Xilinx VCU118 FPGA and evaluate them with classical CNNs such as Alexnet, VGG-16, and ResNet-50. Compared with existing accelerators designed for structured and unstructured pruned networks, our design achieves the best performance in terms of power efficiency.
Palavras-chave

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Idioma: En Ano de publicação: 2023 Tipo de documento: Article

Texto completo: 1 Coleções: 01-internacional Base de dados: MEDLINE Idioma: En Ano de publicação: 2023 Tipo de documento: Article