LAUSR.org creates dashboard-style pages of related content for over 1.5 million academic articles. Sign Up to like articles & get recommendations!

DDDQN‐TS: A task scheduling and load balancing method based on optimized deep reinforcement learning in heterogeneous computing environment

Photo by theblowup from unsplash

Task scheduling and load balancing problem of heterogeneous computing environment (HCE) is getting more and more attention these days and has become a research hotspot in this field. The task… Click to show full abstract

Task scheduling and load balancing problem of heterogeneous computing environment (HCE) is getting more and more attention these days and has become a research hotspot in this field. The task scheduling and load balancing problem of heterogeneous environment, which refers to assigning a set of tasks to a specific set of machines with different hardware and different computing performance with the goal of minimizing task processing time and keeping load balance among machines, has been proved to be an NP‐complete problem. The development of artificial intelligence provides new ideas to solve this problem. In this paper, we propose a novel task scheduling and load balancing method based on optimized deep reinforcement learning in HCE. First, we formulate task scheduling problem as a Markov decision process and then adopt a dueling double deep Q‐learning network to search the optimal task allocation solution. Then we use two well‐known large‐scale cluster data sets Google Cloud Jobs data set and Alibaba Cluster Trace data set to validate our approach. The experimental results show that compared with other existing solutions, our proposed method can achieve much shorter task response time and better load balancing effect.

Keywords: problem; scheduling load; task; load balancing; task scheduling

Journal Title: International Journal of Intelligent Systems
Year Published: 2022

Link to full text (if available)


Share on Social Media:                               Sign Up to like & get
recommendations!

Related content

More Information              News              Social Media              Video              Recommended



                Click one of the above tabs to view related content.