Edge Server Placement and Task Allocation for Maximum Delay Reduction
When edge computing is deployed for delay-sensitive applications such as autonomous driving systems and online gaming, it is important to reduce the maximum delay because real-time performance for all users must be ensured from Quality-of-Service (QoS) perspective. The primary delays in edge computi...
Saved in:
| Main Authors: | , |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
IEEE
2025-01-01
|
| Series: | IEEE Open Journal of the Communications Society |
| Subjects: | |
| Online Access: | https://ieeexplore.ieee.org/document/11099545/ |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| _version_ | 1849340374942744576 |
|---|---|
| author | Koki Shibata Sumiko Miyata |
| author_facet | Koki Shibata Sumiko Miyata |
| author_sort | Koki Shibata |
| collection | DOAJ |
| description | When edge computing is deployed for delay-sensitive applications such as autonomous driving systems and online gaming, it is important to reduce the maximum delay because real-time performance for all users must be ensured from Quality-of-Service (QoS) perspective. The primary delays in edge computing include network delay during data transmission and waiting time at the edge server. Since the waiting time at edge servers depends on server utilization, an increase in utilization bias leads to an increase in maximum delay. If a user is extremely far from the edge server, the network delay for that user will also increase. Conventional edge computing methods focus on reducing the average propagation delay of user-processing requests (tasks). However, these methods increase the utilization variance of each edge server, thus increasing the maximum delay. In this paper, we propose a method for determining both edge server placement and task allocation to reduce the maximum delay. Our method uses a genetic algorithm to optimize server utilization and the distance between users and servers. The maximum delay has been successfully reduced compared with that using conventional methods by simultaneously optimizing the server utilization and distance between users and servers. |
| format | Article |
| id | doaj-art-76f8b19b73df4c7cb42a5c983ded7549 |
| institution | Kabale University |
| issn | 2644-125X |
| language | English |
| publishDate | 2025-01-01 |
| publisher | IEEE |
| record_format | Article |
| series | IEEE Open Journal of the Communications Society |
| spelling | doaj-art-76f8b19b73df4c7cb42a5c983ded75492025-08-20T03:43:55ZengIEEEIEEE Open Journal of the Communications Society2644-125X2025-01-0166207621710.1109/OJCOMS.2025.359364111099545Edge Server Placement and Task Allocation for Maximum Delay ReductionKoki Shibata0https://orcid.org/0009-0001-0418-5132Sumiko Miyata1https://orcid.org/0000-0001-8023-7435Electrical Engineering and Computer Science Department, Shibaura Institute of Technology, Tokyo, JapanSchool of Engineering, Institute of Science Tokyo, Tokyo, JapanWhen edge computing is deployed for delay-sensitive applications such as autonomous driving systems and online gaming, it is important to reduce the maximum delay because real-time performance for all users must be ensured from Quality-of-Service (QoS) perspective. The primary delays in edge computing include network delay during data transmission and waiting time at the edge server. Since the waiting time at edge servers depends on server utilization, an increase in utilization bias leads to an increase in maximum delay. If a user is extremely far from the edge server, the network delay for that user will also increase. Conventional edge computing methods focus on reducing the average propagation delay of user-processing requests (tasks). However, these methods increase the utilization variance of each edge server, thus increasing the maximum delay. In this paper, we propose a method for determining both edge server placement and task allocation to reduce the maximum delay. Our method uses a genetic algorithm to optimize server utilization and the distance between users and servers. The maximum delay has been successfully reduced compared with that using conventional methods by simultaneously optimizing the server utilization and distance between users and servers.https://ieeexplore.ieee.org/document/11099545/Maximum delayedge server placementtask allocationserver utilizationeccentricity centralityqueueing theory |
| spellingShingle | Koki Shibata Sumiko Miyata Edge Server Placement and Task Allocation for Maximum Delay Reduction IEEE Open Journal of the Communications Society Maximum delay edge server placement task allocation server utilization eccentricity centrality queueing theory |
| title | Edge Server Placement and Task Allocation for Maximum Delay Reduction |
| title_full | Edge Server Placement and Task Allocation for Maximum Delay Reduction |
| title_fullStr | Edge Server Placement and Task Allocation for Maximum Delay Reduction |
| title_full_unstemmed | Edge Server Placement and Task Allocation for Maximum Delay Reduction |
| title_short | Edge Server Placement and Task Allocation for Maximum Delay Reduction |
| title_sort | edge server placement and task allocation for maximum delay reduction |
| topic | Maximum delay edge server placement task allocation server utilization eccentricity centrality queueing theory |
| url | https://ieeexplore.ieee.org/document/11099545/ |
| work_keys_str_mv | AT kokishibata edgeserverplacementandtaskallocationformaximumdelayreduction AT sumikomiyata edgeserverplacementandtaskallocationformaximumdelayreduction |