The search functionality is under construction.

IEICE TRANSACTIONS on Information

Efficient Hardware Accelerator for Compressed Sparse Deep Neural Network

Hao XIAO, Kaikai ZHAO, Guangzhu LIU

  • Full Text Views

    0

  • Cite this

Summary :

This work presents a DNN accelerator architecture specifically designed for performing efficient inference on compressed and sparse DNN models. Leveraging the data sparsity, a runtime processing scheme is proposed to deal with the encoded weights and activations directly in the compressed domain without decompressing. Furthermore, a new data flow is proposed to facilitate the reusage of input activations across the fully-connected (FC) layers. The proposed design is implemented and verified using the Xilinx Virtex-7 FPGA. Experimental results show it achieves 1.99×, 1.95× faster and 20.38×, 3.04× more energy efficient than CPU and mGPU platforms, respectively, running AlexNet.

Publication
IEICE TRANSACTIONS on Information Vol.E104-D No.5 pp.772-775
Publication Date
2021/05/01
Publicized
2021/02/19
Online ISSN
1745-1361
DOI
10.1587/transinf.2020EDL8153
Type of Manuscript
LETTER
Category
Computer System

Authors

Hao XIAO
  HeFei University of Technology
Kaikai ZHAO
  HeFei University of Technology
Guangzhu LIU
  HeFei University of Technology

Keyword