A Hardware Accelerator for the Inference of a Convolutional Neural network

Convolutional Neural Networks (CNNs) are becoming increasingly popular in deep learning applications, e.g. image classification, speech recognition, medicine, to name a few. However, the CNN inference is computationally intensive and demanding a large among of memory resources. In this work is prop...

Full description

Saved in:
Bibliographic Details
Main Authors: Edwin González, Walter D. Villamizar Luna, Carlos Augusto Fajardo Ariza
Format: Article
Language:English
Published: Editorial Neogranadina 2019-11-01
Series:Ciencia e Ingeniería Neogranadina
Subjects:
Online Access:https://revistasunimilitareduco.biteca.online/index.php/rcin/article/view/4194
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832539815463092224
author Edwin González
Walter D. Villamizar Luna
Carlos Augusto Fajardo Ariza
author_facet Edwin González
Walter D. Villamizar Luna
Carlos Augusto Fajardo Ariza
author_sort Edwin González
collection DOAJ
description Convolutional Neural Networks (CNNs) are becoming increasingly popular in deep learning applications, e.g. image classification, speech recognition, medicine, to name a few. However, the CNN inference is computationally intensive and demanding a large among of memory resources. In this work is proposed a CNN inference hardware accelerator, which was implemented in a co-processing scheme. The aim is to reduce the hardware resources and achieve the better possible throughput. The design was implemented in the Digilent Arty Z7-20 development board, which is based on System on Chip (SoC) Zynq-7000 of Xilinx. Our implementation achieved a  of accuracy for the MNIST database using only 12-bits fixed-point format. The results show that the co-processing scheme operating at a conservative speed of 100 MHz can identify around 441 images per second, which is about 17% times faster than a 650 MHz - software implementation. It is difficult to compare our results against other implementations based on Field-Programmable Gate Array (FPGA), because the others implementations are not exactly like ours. However, some comparisons, regarding the logical resources used and accuracy, suggest that our work could be better than previous works.
format Article
id doaj-art-4605b0f9555a4e429e448659cdbedade
institution Kabale University
issn 0124-8170
1909-7735
language English
publishDate 2019-11-01
publisher Editorial Neogranadina
record_format Article
series Ciencia e Ingeniería Neogranadina
spelling doaj-art-4605b0f9555a4e429e448659cdbedade2025-02-05T08:57:44ZengEditorial NeogranadinaCiencia e Ingeniería Neogranadina0124-81701909-77352019-11-01301A Hardware Accelerator for the Inference of a Convolutional Neural networkEdwin González 0https://orcid.org/0000-0003-2217-9817Walter D. Villamizar Luna 1https://orcid.org/0000-0003-4341-8020Carlos Augusto Fajardo Ariza2https://orcid.org/0000-0002-8995-4585Universidad Industrial de SantanderUniversidad Industrial de SantanderUniversidad Industrial de Santander Convolutional Neural Networks (CNNs) are becoming increasingly popular in deep learning applications, e.g. image classification, speech recognition, medicine, to name a few. However, the CNN inference is computationally intensive and demanding a large among of memory resources. In this work is proposed a CNN inference hardware accelerator, which was implemented in a co-processing scheme. The aim is to reduce the hardware resources and achieve the better possible throughput. The design was implemented in the Digilent Arty Z7-20 development board, which is based on System on Chip (SoC) Zynq-7000 of Xilinx. Our implementation achieved a  of accuracy for the MNIST database using only 12-bits fixed-point format. The results show that the co-processing scheme operating at a conservative speed of 100 MHz can identify around 441 images per second, which is about 17% times faster than a 650 MHz - software implementation. It is difficult to compare our results against other implementations based on Field-Programmable Gate Array (FPGA), because the others implementations are not exactly like ours. However, some comparisons, regarding the logical resources used and accuracy, suggest that our work could be better than previous works. https://revistasunimilitareduco.biteca.online/index.php/rcin/article/view/4194CNNFPGAHardware acceleratorMNISTZynq
spellingShingle Edwin González
Walter D. Villamizar Luna
Carlos Augusto Fajardo Ariza
A Hardware Accelerator for the Inference of a Convolutional Neural network
Ciencia e Ingeniería Neogranadina
CNN
FPGA
Hardware accelerator
MNIST
Zynq
title A Hardware Accelerator for the Inference of a Convolutional Neural network
title_full A Hardware Accelerator for the Inference of a Convolutional Neural network
title_fullStr A Hardware Accelerator for the Inference of a Convolutional Neural network
title_full_unstemmed A Hardware Accelerator for the Inference of a Convolutional Neural network
title_short A Hardware Accelerator for the Inference of a Convolutional Neural network
title_sort hardware accelerator for the inference of a convolutional neural network
topic CNN
FPGA
Hardware accelerator
MNIST
Zynq
url https://revistasunimilitareduco.biteca.online/index.php/rcin/article/view/4194
work_keys_str_mv AT edwingonzalez ahardwareacceleratorfortheinferenceofaconvolutionalneuralnetwork
AT walterdvillamizarluna ahardwareacceleratorfortheinferenceofaconvolutionalneuralnetwork
AT carlosaugustofajardoariza ahardwareacceleratorfortheinferenceofaconvolutionalneuralnetwork
AT edwingonzalez hardwareacceleratorfortheinferenceofaconvolutionalneuralnetwork
AT walterdvillamizarluna hardwareacceleratorfortheinferenceofaconvolutionalneuralnetwork
AT carlosaugustofajardoariza hardwareacceleratorfortheinferenceofaconvolutionalneuralnetwork