Development of the Reinforcement Learning-based Adaptive Scheduling Algorithm for Panel Block Shop

2021 ◽  
Vol 26 (2) ◽  
pp. 81-92
Author(s):  
Young In Cho ◽  
So Hyun Nam ◽  
Jong Hun Woo
2019 ◽  
Vol 11 (7) ◽  
pp. 1826 ◽  
Author(s):  
Yuxia Cheng ◽  
Zhiwei Wu ◽  
Kui Liu ◽  
Qing Wu ◽  
Yu Wang

Task scheduling is critical for improving system performance in the distributed heterogeneous computing environment. The Directed Acyclic Graph (DAG) tasks scheduling problem is NP-complete and it is hard to find an optimal schedule. Due to its key importance, the DAG tasks scheduling problem has been extensively studied in the literature. However, many previously proposed traditional heuristic algorithms are usually based on greedy methods and also lack the consideration of scheduling tasks between trusted and untrusted entities, which makes the problem more complicated, but there still exists a large optimization space to be explored. In this paper, we propose a trust-aware adaptive DAG tasks scheduling algorithm using the reinforcement learning and Monte Carlo Tree Search (MCTS) methods. The scheduling problem is defined using the reinforcement learning model. Efficient scheduling state space, action space and reward function are designed to train the policy gradient-based REINFORCE agent. The MCTS method is proposed to determine actual scheduling policies when DAG tasks are simultaneously executed in trusted and untrusted entities. Leveraging the algorithm’s capability of exploring long term reward, the proposed algorithm could achieve good scheduling policies while guaranteeing trusted tasks scheduled within trusted entities. Experimental results showed the effectiveness of the proposed algorithm compared with the classic HEFT/CPOP algorithms.


2019 ◽  
Vol 92 ◽  
pp. 43-51 ◽  
Author(s):  
Chao Qiu ◽  
Shaohua Cui ◽  
Haipeng Yao ◽  
Fangmin Xu ◽  
F. Richard Yu ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document