Name: | Description: | Size: | Format: | |
---|---|---|---|---|
173.6 KB | Adobe PDF |
Advisor(s)
Abstract(s)
Due to the computational complexity of Convolutional Neural Networks (CNNs), high performance platforms are generally considered for their execution. However, CNNs are very useful in embedded systems and its execution right next to the source of data has many advantages, like avoiding the need for data communication. In this paper, we propose an architecture for CNN inference (Lite-CNN) that can achieve high performance in low density FPGAs. Lite-CNN adopts a fixed-point representation for both neurons and weights, which was already shown to be sufficient for most CNNs. Also, with a simple and known dot product reorganization, the number of multiplications is reduced to half. We show implementation results for 8 bit fixed-point in a ZYNQ7020 and extrapolate for other larger FPGAs. Lite-CNN achieves 410 GOPs in a ZYNQ7020.
Description
Keywords
Embedded computing Deep learning Convolutional neural network Field-programmable gate array
Citation
VÉSTIAS, Mário; [et al] – Lite-CNN: a high-performance architecture to execute CNNs in low density FPGAs. In 28th International Conference on Field Programmable Logic & Applications. Dublin, Ireland: 2018. Pp. 399-402