Repository logo
 
No Thumbnail Available
Publication

Lite-CNN: a high-performance architecture to execute CNNs in low density FPGAs

Use this identifier to reference this record.
Name:Description:Size:Format: 
Lite-CNN_MVestias.pdf173.6 KBAdobe PDF Download

Advisor(s)

Abstract(s)

Due to the computational complexity of Convolutional Neural Networks (CNNs), high performance platforms are generally considered for their execution. However, CNNs are very useful in embedded systems and its execution right next to the source of data has many advantages, like avoiding the need for data communication. In this paper, we propose an architecture for CNN inference (Lite-CNN) that can achieve high performance in low density FPGAs. Lite-CNN adopts a fixed-point representation for both neurons and weights, which was already shown to be sufficient for most CNNs. Also, with a simple and known dot product reorganization, the number of multiplications is reduced to half. We show implementation results for 8 bit fixed-point in a ZYNQ7020 and extrapolate for other larger FPGAs. Lite-CNN achieves 410 GOPs in a ZYNQ7020.

Description

Keywords

Embedded computing Deep learning Convolutional neural network Field-programmable gate array

Citation

VÉSTIAS, Mário; [et al] – Lite-CNN: a high-performance architecture to execute CNNs in low density FPGAs. In 28th International Conference on Field Programmable Logic & Applications. Dublin, Ireland: 2018. Pp. 399-402

Research Projects

Research ProjectShow more

Organizational Units

Journal Issue