Optimizing FCN for devices with limited resources using quantization and sparsity enhancement

Abstract This study addresses the optimization of fully convolutional networks (FCNs) for deployment on resource-limited devices in real-time scenarios. While prior research has extensively applied quantization techniques to architectures like VGG-16, there is limited exploration of comprehensive la...

Full description

Saved in:
Bibliographic Details
Main Authors: Muhammad Faizan-Khan, Nisar Ali, Raja Hashim Ali, Areej Alasiry, Mehrez Marzougui, Shabbab Ali Algamdi, Yunyoung Nam
Format: Article
Language:English
Published: Nature Portfolio 2025-08-01
Series:Scientific Reports
Subjects:
Online Access:https://doi.org/10.1038/s41598-025-06848-3
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Abstract This study addresses the optimization of fully convolutional networks (FCNs) for deployment on resource-limited devices in real-time scenarios. While prior research has extensively applied quantization techniques to architectures like VGG-16, there is limited exploration of comprehensive layer-wise quantization specifically within the FCN-8 architecture. To fill this gap, we propose an innovative approach utilizing full-layer quantization with an $$L_2$$ error minimization algorithm, accompanied by sensitivity analysis to optimize fixed-point representation of network weights. Our results demonstrate that this method significantly enhances sparsity, achieving up to 40%, while preserving performance, yielding an impressive 89.3% pixel accuracy under extreme quantization conditions. The findings highlight the efficacy of full-layer quantization and retraining in simultaneously reducing network complexity and maintaining accuracy in both image classification and semantic segmentation tasks.
ISSN:2045-2322