Loading…

Deep reinforcement learning-based resource scheduling for energy optimization and load balancing in SDN-driven edge computing

Traditional techniques for edge computing resource scheduling may result in large amounts of wasted server resources and energy consumption; thus, exploring new approaches to achieve higher resource and energy efficiency is a new challenge. Deep reinforcement learning (DRL) offers a promising soluti...

Full description

Saved in:
Bibliographic Details
Published in:Computer communications 2024-10, Vol.226-227, p.107925, Article 107925
Main Authors: Zhou, Xu, Yang, Jing, Li, Yijun, Li, Shaobo, Su, Zhidong
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:Traditional techniques for edge computing resource scheduling may result in large amounts of wasted server resources and energy consumption; thus, exploring new approaches to achieve higher resource and energy efficiency is a new challenge. Deep reinforcement learning (DRL) offers a promising solution by balancing resource utilization, latency, and energy optimization. However, current methods often focus solely on energy optimization for offloading and computing tasks, neglecting the impact of server numbers and resource operation status on energy efficiency and load balancing. On the other hand, prioritizing latency optimization may result in resource imbalance and increased energy waste. To address these challenges, we propose a novel energy optimization method coupled with a load balancing strategy. Our approach aims to minimize overall energy consumption and achieve server load balancing under latency constraints. This is achieved by controlling the number of active servers and individual server load states through a two stage DRL-based energy and resource optimization algorithm. Experimental results demonstrate that our scheme can save an average of 19.84% energy compared to mainstream reinforcement learning methods and 49.60% and 45.33% compared to Round Robin (RR) and random scheduling, respectively. Additionally, our method is optimized for reward value, load balancing, runtime, and anti-interference capability.
ISSN:0140-3664
DOI:10.1016/j.comcom.2024.107925