Your browser doesn't support javascript.
loading
Research on Convolutional Neural Network Inference Acceleration and Performance Optimization for Edge Intelligence.
Liang, Yong; Tan, Junwen; Xie, Zhisong; Chen, Zetao; Lin, Daoqian; Yang, Zhenhao.
Afiliación
  • Liang Y; Key Laboratory of Advanced Manufacturing and Automation Technology (Guilin University of Technology), Education Department of Guangxi Zhuang, Autonomous Region, Guilin 541006, China.
  • Tan J; College of Mechanical and Control Engineering, Guilin University of Technology, Guilin 541006, China.
  • Xie Z; Key Laboratory of Advanced Manufacturing and Automation Technology (Guilin University of Technology), Education Department of Guangxi Zhuang, Autonomous Region, Guilin 541006, China.
  • Chen Z; College of Mechanical and Control Engineering, Guilin University of Technology, Guilin 541006, China.
  • Lin D; College of Mechanical and Control Engineering, Guilin University of Technology, Guilin 541006, China.
  • Yang Z; College of Mechanical and Control Engineering, Guilin University of Technology, Guilin 541006, China.
Sensors (Basel) ; 24(1)2023 Dec 31.
Article en En | MEDLINE | ID: mdl-38203102
ABSTRACT
In recent years, edge intelligence (EI) has emerged, combining edge computing with AI, and specifically deep learning, to run AI algorithms directly on edge devices. In practical applications, EI faces challenges related to computational power, power consumption, size, and cost, with the primary challenge being the trade-off between computational power and power consumption. This has rendered traditional computing platforms unsustainable, making heterogeneous parallel computing platforms a crucial pathway for implementing EI. In our research, we leveraged the Xilinx Zynq 7000 heterogeneous computing platform, employed high-level synthesis (HLS) for design, and implemented two different accelerators for LeNet-5 using loop unrolling and pipelining optimization techniques. The experimental results show that when running at a clock speed of 100 MHz, the PIPELINE accelerator, compared to the UNROLL accelerator, experiences an 8.09% increase in power consumption but speeds up by 14.972 times, making the PIPELINE accelerator superior in performance. Compared to the CPU, the PIPELINE accelerator reduces power consumption by 91.37% and speeds up by 70.387 times, while compared to the GPU, it reduces power consumption by 93.35%. This study provides two different optimization schemes for edge intelligence applications through design and experimentation and demonstrates the impact of different quantization methods on FPGA resource consumption. These experimental results can provide a reference for practical applications, thereby providing a reference hardware acceleration scheme for edge intelligence applications.
Palabras clave

Texto completo: 1 Colección: 01-internacional Banco de datos: MEDLINE Idioma: En Revista: Sensors (Basel) Año: 2023 Tipo del documento: Article País de afiliación: China

Texto completo: 1 Colección: 01-internacional Banco de datos: MEDLINE Idioma: En Revista: Sensors (Basel) Año: 2023 Tipo del documento: Article País de afiliación: China