A Hardware Accelerator for the Inference of a Convolutional Neural network
Convolutional Neural Networks (CNNs) are becoming increasingly popular in deep learning applications, e.g. image classification, speech recognition, medicine, to name a few. However, the CNN inference is computationally intensive and demanding a large among of memory resources. In this work is prop...
Saved in:
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Editorial Neogranadina
2019-11-01
|
Series: | Ciencia e Ingeniería Neogranadina |
Subjects: | |
Online Access: | https://revistasunimilitareduco.biteca.online/index.php/rcin/article/view/4194 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1832539815463092224 |
---|---|
author | Edwin González Walter D. Villamizar Luna Carlos Augusto Fajardo Ariza |
author_facet | Edwin González Walter D. Villamizar Luna Carlos Augusto Fajardo Ariza |
author_sort | Edwin González |
collection | DOAJ |
description |
Convolutional Neural Networks (CNNs) are becoming increasingly popular in deep learning applications, e.g. image classification, speech recognition, medicine, to name a few. However, the CNN inference is computationally intensive and demanding a large among of memory resources. In this work is proposed a CNN inference hardware accelerator, which was implemented in a co-processing scheme. The aim is to reduce the hardware resources and achieve the better possible throughput. The design was implemented in the Digilent Arty Z7-20 development board, which is based on System on Chip (SoC) Zynq-7000 of Xilinx. Our implementation achieved a of accuracy for the MNIST database using only 12-bits fixed-point format. The results show that the co-processing scheme operating at a conservative speed of 100 MHz can identify around 441 images per second, which is about 17% times faster than a 650 MHz - software implementation. It is difficult to compare our results against other implementations based on Field-Programmable Gate Array (FPGA), because the others implementations are not exactly like ours. However, some comparisons, regarding the logical resources used and accuracy, suggest that our work could be better than previous works.
|
format | Article |
id | doaj-art-4605b0f9555a4e429e448659cdbedade |
institution | Kabale University |
issn | 0124-8170 1909-7735 |
language | English |
publishDate | 2019-11-01 |
publisher | Editorial Neogranadina |
record_format | Article |
series | Ciencia e Ingeniería Neogranadina |
spelling | doaj-art-4605b0f9555a4e429e448659cdbedade2025-02-05T08:57:44ZengEditorial NeogranadinaCiencia e Ingeniería Neogranadina0124-81701909-77352019-11-01301A Hardware Accelerator for the Inference of a Convolutional Neural networkEdwin González 0https://orcid.org/0000-0003-2217-9817Walter D. Villamizar Luna 1https://orcid.org/0000-0003-4341-8020Carlos Augusto Fajardo Ariza2https://orcid.org/0000-0002-8995-4585Universidad Industrial de SantanderUniversidad Industrial de SantanderUniversidad Industrial de Santander Convolutional Neural Networks (CNNs) are becoming increasingly popular in deep learning applications, e.g. image classification, speech recognition, medicine, to name a few. However, the CNN inference is computationally intensive and demanding a large among of memory resources. In this work is proposed a CNN inference hardware accelerator, which was implemented in a co-processing scheme. The aim is to reduce the hardware resources and achieve the better possible throughput. The design was implemented in the Digilent Arty Z7-20 development board, which is based on System on Chip (SoC) Zynq-7000 of Xilinx. Our implementation achieved a of accuracy for the MNIST database using only 12-bits fixed-point format. The results show that the co-processing scheme operating at a conservative speed of 100 MHz can identify around 441 images per second, which is about 17% times faster than a 650 MHz - software implementation. It is difficult to compare our results against other implementations based on Field-Programmable Gate Array (FPGA), because the others implementations are not exactly like ours. However, some comparisons, regarding the logical resources used and accuracy, suggest that our work could be better than previous works. https://revistasunimilitareduco.biteca.online/index.php/rcin/article/view/4194CNNFPGAHardware acceleratorMNISTZynq |
spellingShingle | Edwin González Walter D. Villamizar Luna Carlos Augusto Fajardo Ariza A Hardware Accelerator for the Inference of a Convolutional Neural network Ciencia e Ingeniería Neogranadina CNN FPGA Hardware accelerator MNIST Zynq |
title | A Hardware Accelerator for the Inference of a Convolutional Neural network |
title_full | A Hardware Accelerator for the Inference of a Convolutional Neural network |
title_fullStr | A Hardware Accelerator for the Inference of a Convolutional Neural network |
title_full_unstemmed | A Hardware Accelerator for the Inference of a Convolutional Neural network |
title_short | A Hardware Accelerator for the Inference of a Convolutional Neural network |
title_sort | hardware accelerator for the inference of a convolutional neural network |
topic | CNN FPGA Hardware accelerator MNIST Zynq |
url | https://revistasunimilitareduco.biteca.online/index.php/rcin/article/view/4194 |
work_keys_str_mv | AT edwingonzalez ahardwareacceleratorfortheinferenceofaconvolutionalneuralnetwork AT walterdvillamizarluna ahardwareacceleratorfortheinferenceofaconvolutionalneuralnetwork AT carlosaugustofajardoariza ahardwareacceleratorfortheinferenceofaconvolutionalneuralnetwork AT edwingonzalez hardwareacceleratorfortheinferenceofaconvolutionalneuralnetwork AT walterdvillamizarluna hardwareacceleratorfortheinferenceofaconvolutionalneuralnetwork AT carlosaugustofajardoariza hardwareacceleratorfortheinferenceofaconvolutionalneuralnetwork |