Edge Server Placement and Task Allocation for Maximum Delay Reduction

When edge computing is deployed for delay-sensitive applications such as autonomous driving systems and online gaming, it is important to reduce the maximum delay because real-time performance for all users must be ensured from Quality-of-Service (QoS) perspective. The primary delays in edge computi...

Full description

Saved in:
Bibliographic Details
Main Authors: Koki Shibata, Sumiko Miyata
Format: Article
Language:English
Published: IEEE 2025-01-01
Series:IEEE Open Journal of the Communications Society
Subjects:
Online Access:https://ieeexplore.ieee.org/document/11099545/
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:When edge computing is deployed for delay-sensitive applications such as autonomous driving systems and online gaming, it is important to reduce the maximum delay because real-time performance for all users must be ensured from Quality-of-Service (QoS) perspective. The primary delays in edge computing include network delay during data transmission and waiting time at the edge server. Since the waiting time at edge servers depends on server utilization, an increase in utilization bias leads to an increase in maximum delay. If a user is extremely far from the edge server, the network delay for that user will also increase. Conventional edge computing methods focus on reducing the average propagation delay of user-processing requests (tasks). However, these methods increase the utilization variance of each edge server, thus increasing the maximum delay. In this paper, we propose a method for determining both edge server placement and task allocation to reduce the maximum delay. Our method uses a genetic algorithm to optimize server utilization and the distance between users and servers. The maximum delay has been successfully reduced compared with that using conventional methods by simultaneously optimizing the server utilization and distance between users and servers.
ISSN:2644-125X