본문 바로가기
로그인

RESEARCH

Semiconductor System Lab

Through this homepage, we would like to share our sweats, pains,
excitements and experiences with you.

Neuromorphic 

SNPU: Ultralow-power Spike Domain Deep-Neural-Network Processor

본문

Overview

In this paper, we propose SNPU, a spike-domain DNN accelerator, to achieve ultralow power consumption with neuron-level event-driven inference. The spike domain processing can replace many MACs with small number of accumulations. The proposed SNPU has 3 key features; spike train decomposition to reduce the accumulations, time shrinking multi-level encoding to replace the multiple accumulations with single shift-and-accumulation which adopts bit scalability to enable different always-on applications, and neuron link to support various time-windows for optimal energy consumption by minimizing time-window in layer-by-layer and to increase the PE utilization. For face recognition with the LFW dataset, the proposed spike domain processing can reduce the energy consumption by 43.9% with less than 1% accuracy loss. If there is no face in the input image, the energy consumption can be reduced further by 87.6%. The SNPU which is fabricated in 28-nm CMOS technology. It occupies 6.3 mm2 die area with 120 KB on-chip SRAM, and can operate from 2 MHz to 200 MHz clock frequencies with 0.63 V to 1.1 V supply voltages. And it achieves energy efficiency of 150.1 TOPS/W during face recognition. Original DNN before the spike domain conversion is composed of 5 convolutional layers and 1 fully connected layer with 326 KB parameters. The shift-and-accumulations of each layer in SNPU is only ~25% of MAC operations in the original CNN. The SNPU shows 96% accuracy of LFW dataset, and its latency is only 0.681 ms. Also, the maximum throughput is 147K fps, and the peak performance is 15.9 TOPS. In conclusion, a 63.2 μW ultralow-power spike domain DNN processor with spike train decomposition, time shrinking multi-level encoding and neuron link is proposed.

Implementation results


Performance comparison



Architecture

 
Features

  - spike train decomposition to reduce the accumulations

  - time shrinking multi-level encoding to replace the multiple accumulations with single shift-and-accumulation which adopts bit scalability to enable different always-on applications

  - neuron link to support various time-windows for optimal energy consumption by minimizing time-window in layer-by-layer and to increase the PE utilization.




Related Papers

  - ASSCC 2022



Address#1233, School of Electrical Engineering, KAIST, 291 Daehak-ro (373-1 Guseong-dong), Yuseong-gu, Daejeon 34141, Republic of Korea
Tel +82-42-350-8068 Fax +82-42-350-3410E-mail [email protected]·© SSL. All Rights Reserved.·Design by NSTAR