Cooperative inference analysis based on DNN convolutional kernel partitioning

With the popularity of intelligent chip in the application of edge terminal devices, a large number of AI applications will be deployed on the edge of networks closer to data sources in the future.The method based on DNN partition can realize deep learning model training and deployment on resource-c...

Full description

Saved in:
Bibliographic Details
Main Authors: Jialin ZHI, Yinglei TENG, Xinyang ZHANG, Tao NIU, Mei SONG
Format: Article
Language:zho
Published: China InfoCom Media Group 2022-12-01
Series:物联网学报
Subjects:
Online Access:http://www.wlwxb.com.cn/thesisDetails#10.11959/j.issn.2096-3750.2022.00308
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1850212442470088704
author Jialin ZHI
Yinglei TENG
Xinyang ZHANG
Tao NIU
Mei SONG
author_facet Jialin ZHI
Yinglei TENG
Xinyang ZHANG
Tao NIU
Mei SONG
author_sort Jialin ZHI
collection DOAJ
description With the popularity of intelligent chip in the application of edge terminal devices, a large number of AI applications will be deployed on the edge of networks closer to data sources in the future.The method based on DNN partition can realize deep learning model training and deployment on resource-constrained terminal devices, and solve the bottleneck problem of edge AI computing ability.Thekernel based partition method (KPM) was proposed as a new scheme on the basis of traditional workload based partition method (WPM).The quantitative analysis of inference performance was carried out from three aspects of computation FLOPS, memory consumption and communication cost respectively, and the qualitative analysis of the above two schemes was carried out from the perspective of flexibility, robustness and privacy of inference process.Finally, a software and hardware experimental platform was built, and AlexNet and VGG11 networks were implemented using PyTorch to further verify the performance advantages of the proposed scheme in terms of delay and energy consumption.It was concluded that, compared with the WPM scheme, the KPM scheme had better DNN reasoning acceleration effect in large-scale computing scenarios.And it has lower memory usage and energy consumption.
format Article
id doaj-art-2dedb9ddc6944b1db822b731fb75d494
institution OA Journals
issn 2096-3750
language zho
publishDate 2022-12-01
publisher China InfoCom Media Group
record_format Article
series 物联网学报
spelling doaj-art-2dedb9ddc6944b1db822b731fb75d4942025-08-20T02:09:21ZzhoChina InfoCom Media Group物联网学报2096-37502022-12-016728159580497Cooperative inference analysis based on DNN convolutional kernel partitioningJialin ZHIYinglei TENGXinyang ZHANGTao NIUMei SONGWith the popularity of intelligent chip in the application of edge terminal devices, a large number of AI applications will be deployed on the edge of networks closer to data sources in the future.The method based on DNN partition can realize deep learning model training and deployment on resource-constrained terminal devices, and solve the bottleneck problem of edge AI computing ability.Thekernel based partition method (KPM) was proposed as a new scheme on the basis of traditional workload based partition method (WPM).The quantitative analysis of inference performance was carried out from three aspects of computation FLOPS, memory consumption and communication cost respectively, and the qualitative analysis of the above two schemes was carried out from the perspective of flexibility, robustness and privacy of inference process.Finally, a software and hardware experimental platform was built, and AlexNet and VGG11 networks were implemented using PyTorch to further verify the performance advantages of the proposed scheme in terms of delay and energy consumption.It was concluded that, compared with the WPM scheme, the KPM scheme had better DNN reasoning acceleration effect in large-scale computing scenarios.And it has lower memory usage and energy consumption.http://www.wlwxb.com.cn/thesisDetails#10.11959/j.issn.2096-3750.2022.00308edge intelligence;deep neural network partition;cooperative computation;parallel partition
spellingShingle Jialin ZHI
Yinglei TENG
Xinyang ZHANG
Tao NIU
Mei SONG
Cooperative inference analysis based on DNN convolutional kernel partitioning
物联网学报
edge intelligence;deep neural network partition;cooperative computation;parallel partition
title Cooperative inference analysis based on DNN convolutional kernel partitioning
title_full Cooperative inference analysis based on DNN convolutional kernel partitioning
title_fullStr Cooperative inference analysis based on DNN convolutional kernel partitioning
title_full_unstemmed Cooperative inference analysis based on DNN convolutional kernel partitioning
title_short Cooperative inference analysis based on DNN convolutional kernel partitioning
title_sort cooperative inference analysis based on dnn convolutional kernel partitioning
topic edge intelligence;deep neural network partition;cooperative computation;parallel partition
url http://www.wlwxb.com.cn/thesisDetails#10.11959/j.issn.2096-3750.2022.00308
work_keys_str_mv AT jialinzhi cooperativeinferenceanalysisbasedondnnconvolutionalkernelpartitioning
AT yingleiteng cooperativeinferenceanalysisbasedondnnconvolutionalkernelpartitioning
AT xinyangzhang cooperativeinferenceanalysisbasedondnnconvolutionalkernelpartitioning
AT taoniu cooperativeinferenceanalysisbasedondnnconvolutionalkernelpartitioning
AT meisong cooperativeinferenceanalysisbasedondnnconvolutionalkernelpartitioning