A low functional redundancy-based network slimming method for accelerating deep neural networks
Deep neural networks (DNNs) have been widely criticized for their large parameters and computation demands, hindering deployment to edge and embedded devices. In order to reduce the floating point operations (FLOPs) running DNNs and accelerate the inference speed, we start from the model pruning, an...
Saved in:
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
Elsevier
2025-04-01
|
Series: | Alexandria Engineering Journal |
Subjects: | |
Online Access: | http://www.sciencedirect.com/science/article/pii/S1110016824017162 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1823864378258096128 |
---|---|
author | Zheng Fang Bo Yin |
author_facet | Zheng Fang Bo Yin |
author_sort | Zheng Fang |
collection | DOAJ |
description | Deep neural networks (DNNs) have been widely criticized for their large parameters and computation demands, hindering deployment to edge and embedded devices. In order to reduce the floating point operations (FLOPs) running DNNs and accelerate the inference speed, we start from the model pruning, and realize this goal by removing useless network parameters. In this research, we propose a low functional redundancy-based network slimming method (LFRNS) that can find and remove functional redundant filters by feature clustering algorithm. However, the redundancy of some key features is beneficial to the model, and removing these features will limit the potential of the model to some extent. Build on this view, we propose feature contribution ranking unit (FCR unit) which can automatically learn the feature maps' contribution to the key information with training iterations. FCR unit can assist LFRNS restore some important elements in the pruning set to break the performance bottleneck of the slimming model. Our method mainly removes feature maps with similar functions instead of only pruning the unimportant parts, thus effectively ensuring the integrity of features’ functions and avoiding network degradation. We conduct experiments on image classification task based on CIFAR-10 and CIFAR-100 datasets. Our framework achieves over 2.0 × parameters and FLOPs reductions, while maintaining < 1 % loss in accuracy, and even improve accuracy of large-volume models. We also introduce our method to the vision transformer model (ViT) and achieve performance comparable to state-of-the-art methods with nearly 1.5 × less computation. |
format | Article |
id | doaj-art-149ec8add3ec4e29a981d2d014cb9256 |
institution | Kabale University |
issn | 1110-0168 |
language | English |
publishDate | 2025-04-01 |
publisher | Elsevier |
record_format | Article |
series | Alexandria Engineering Journal |
spelling | doaj-art-149ec8add3ec4e29a981d2d014cb92562025-02-09T04:59:45ZengElsevierAlexandria Engineering Journal1110-01682025-04-01119437450A low functional redundancy-based network slimming method for accelerating deep neural networksZheng Fang0Bo Yin1College of Information Science and Engineering, Ocean University of China, Qingdao, ChinaCollege of Information Science and Engineering, Ocean University of China, Qingdao, China; Pilot National Laboratory for Marine Science and Technology, Qingdao, China; Corresponding author.Deep neural networks (DNNs) have been widely criticized for their large parameters and computation demands, hindering deployment to edge and embedded devices. In order to reduce the floating point operations (FLOPs) running DNNs and accelerate the inference speed, we start from the model pruning, and realize this goal by removing useless network parameters. In this research, we propose a low functional redundancy-based network slimming method (LFRNS) that can find and remove functional redundant filters by feature clustering algorithm. However, the redundancy of some key features is beneficial to the model, and removing these features will limit the potential of the model to some extent. Build on this view, we propose feature contribution ranking unit (FCR unit) which can automatically learn the feature maps' contribution to the key information with training iterations. FCR unit can assist LFRNS restore some important elements in the pruning set to break the performance bottleneck of the slimming model. Our method mainly removes feature maps with similar functions instead of only pruning the unimportant parts, thus effectively ensuring the integrity of features’ functions and avoiding network degradation. We conduct experiments on image classification task based on CIFAR-10 and CIFAR-100 datasets. Our framework achieves over 2.0 × parameters and FLOPs reductions, while maintaining < 1 % loss in accuracy, and even improve accuracy of large-volume models. We also introduce our method to the vision transformer model (ViT) and achieve performance comparable to state-of-the-art methods with nearly 1.5 × less computation.http://www.sciencedirect.com/science/article/pii/S1110016824017162Deep neural networksNetwork pruningFunctional redundancyContribution Ranking |
spellingShingle | Zheng Fang Bo Yin A low functional redundancy-based network slimming method for accelerating deep neural networks Alexandria Engineering Journal Deep neural networks Network pruning Functional redundancy Contribution Ranking |
title | A low functional redundancy-based network slimming method for accelerating deep neural networks |
title_full | A low functional redundancy-based network slimming method for accelerating deep neural networks |
title_fullStr | A low functional redundancy-based network slimming method for accelerating deep neural networks |
title_full_unstemmed | A low functional redundancy-based network slimming method for accelerating deep neural networks |
title_short | A low functional redundancy-based network slimming method for accelerating deep neural networks |
title_sort | low functional redundancy based network slimming method for accelerating deep neural networks |
topic | Deep neural networks Network pruning Functional redundancy Contribution Ranking |
url | http://www.sciencedirect.com/science/article/pii/S1110016824017162 |
work_keys_str_mv | AT zhengfang alowfunctionalredundancybasednetworkslimmingmethodforacceleratingdeepneuralnetworks AT boyin alowfunctionalredundancybasednetworkslimmingmethodforacceleratingdeepneuralnetworks AT zhengfang lowfunctionalredundancybasednetworkslimmingmethodforacceleratingdeepneuralnetworks AT boyin lowfunctionalredundancybasednetworkslimmingmethodforacceleratingdeepneuralnetworks |