SciELO - Scientific Electronic Library Online

 
vol.30 issue1A Data Fusion System for Simulating Critical Scenarios and Decision-MakingIdentifying Actors in a Disaster Using Twitter: Sinabung 2018 Case Study author indexsubject indexarticles search
Home Pagealphabetic serial listing  

Services on Demand

Journal

Article

Indicators

Related links

  • On index processCited by Google
  • Have no similar articlesSimilars in SciELO
  • On index processSimilars in Google

Share


Ciencia e Ingeniería Neogranadina

Print version ISSN 0124-8170On-line version ISSN 1909-7735

Abstract

GONZALEZ, Edwin; VILLAMIZAR LUNA, Walter D.  and  FAJARDO ARIZA, Carlos Augusto. A Hardware Accelerator for The Inference of a Convolutional Neural Network. Cienc. Ing. Neogranad. [online]. 2020, vol.30, n.1, pp.107-116.  Epub Aug 16, 2020. ISSN 0124-8170.  https://doi.org/10.18359/rcin.4194.

Convolutional Neural Networks (CNNs) are becoming increasingly popular in deep learning applications, e.g. image classification, speech recognition, medicine, to name a few. However, CNN inference is computationally intensive and demands a large number of memory resources. This work proposes a CNN inference hardware accelerator, which was implemented in a co-processing scheme. The aim is to reduce hardware resources and achieve the best possible throughput. The design is implemented in the Digilent Arty Z7-20 development board, which is based on the Xilinx Zynq-7000 System on Chip (SoC). Our implementation achieved a of accuracy for the MNIST database using only a 12-bits fixed-point format. Results show that the co-processing scheme operating at a conservative speed of 100 MHz can identify around 441 images per second, which is about 17% times faster than a 650 MHz - software implementation. It is difficult to compare our results against other Field-Programmable Gate Array (FPGA)-based implementations because they are not exactly like ours. However, some comparisons, regarding logical resources used and accuracy, suggest that our work could be better than previous ones. Besides, the proposed scheme is compared with a hardware implementation in terms of power consumption and throughput.

Keywords : CNN; FPGA; hardware accelerator; MNIST; Zynq.

        · abstract in Spanish     · text in English     · English ( pdf )