TY - JOUR
T1 - Reinforcement learning-based dynamic load balancing in edge computing networks
AU - Esmaeili, Mohammad Esmaeil
AU - Khonsari, Ahmad
AU - Sohrabi, Vahid
AU - Dadlani, Aresh
N1 - Publisher Copyright:
© 2024 Elsevier B.V.
PY - 2024/6/1
Y1 - 2024/6/1
N2 - Edge computing (EC) has emerged as a paradigm aimed at reducing data transmission latency by bringing computing resources closer to users. However, the limited scale and constrained processing power of EC pose challenges in matching the resource availability of larger cloud networks. Load balancing (LB) algorithms play a crucial role in distributing workload among edge servers and minimizing user latency. This paper presents a novel set of distributed LB algorithms that leverage machine learning techniques to overcome the three limitations of our previous LB algorithm, EVBLB: (i) its reliance on static time intervals for execution, (ii) the need for comprehensive information about all server resources and queued requests for neighbor selection, and (iii) the use of a central coordinator to dispatch incoming user requests over edge servers. To offer increased control, custom configuration, and scalability for LB on edge servers, we propose three efficient algorithms: Q-learning (QL), multi-armed bandit (MAB), and gradient bandit (GB) algorithms. The QL algorithm predicts the subsequent execution time of the EVBLB algorithm by incorporating rewards obtained from previous executions, thereby improving performance across various metrics. The MAB and GB algorithms prioritize near-optimal neighbor node servers while considering dynamic changes in request rate, request size, and edge server resources. Through simulations, we evaluate and compare the algorithms in terms of network throughput, average user response time, and a novel LB metric for workload distribution across edge servers.
AB - Edge computing (EC) has emerged as a paradigm aimed at reducing data transmission latency by bringing computing resources closer to users. However, the limited scale and constrained processing power of EC pose challenges in matching the resource availability of larger cloud networks. Load balancing (LB) algorithms play a crucial role in distributing workload among edge servers and minimizing user latency. This paper presents a novel set of distributed LB algorithms that leverage machine learning techniques to overcome the three limitations of our previous LB algorithm, EVBLB: (i) its reliance on static time intervals for execution, (ii) the need for comprehensive information about all server resources and queued requests for neighbor selection, and (iii) the use of a central coordinator to dispatch incoming user requests over edge servers. To offer increased control, custom configuration, and scalability for LB on edge servers, we propose three efficient algorithms: Q-learning (QL), multi-armed bandit (MAB), and gradient bandit (GB) algorithms. The QL algorithm predicts the subsequent execution time of the EVBLB algorithm by incorporating rewards obtained from previous executions, thereby improving performance across various metrics. The MAB and GB algorithms prioritize near-optimal neighbor node servers while considering dynamic changes in request rate, request size, and edge server resources. Through simulations, we evaluate and compare the algorithms in terms of network throughput, average user response time, and a novel LB metric for workload distribution across edge servers.
KW - Edge computing
KW - Gradient bandit algorithm
KW - Load balancing
KW - Multi-armed bandit
KW - Q-learning
KW - Reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85192222866&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85192222866&partnerID=8YFLogxK
U2 - 10.1016/j.comcom.2024.04.009
DO - 10.1016/j.comcom.2024.04.009
M3 - Article
AN - SCOPUS:85192222866
SN - 0140-3664
VL - 222
SP - 188
EP - 197
JO - Computer Communications
JF - Computer Communications
ER -