A 1.625 TOPS/W SOC for Deep CNN Training and Inference in 28nm CMOS
Journal
ESSCIRC 2021 - IEEE 47th European Solid State Circuits Conference, Proceedings
Pages
107-110
Date Issued
2021
Author(s)
Abstract
This work presents a FloatSD8-based system on chip (SOC) for the inference as well as the training of a convolutional neural networks (CNNs). A novel number format (FloatSD8) is employed to reduce the computational complexity of the convolution circuit. By co-designing data representation and circuit, we demonstrate that the AISOC can achieve high convolution performance and optimal energy efficiency without sacrificing the quality of training. At its normal operating condition (200MHz), the AISOC prototype is capable of 0.69 TFLOPS peak performance and 1.625 TOPS/W in 28nm CMOS. ? 2021 IEEE.
Subjects
AI accelerator
low-precision neural network
machine learning
SOC
CMOS integrated circuits
Energy efficiency
Machine learning
Neural networks
Programmable logic controllers
System-on-chip
28nm CMOS
Co-designing
Convolutional neural network
Data representations
Low-precision neural network
Lower precision
Network inference
Neural networks trainings
Neural-networks
Convolution
SDGs
Type
conference paper