A Deep Q-Learning Network for Dynamic Constraint-Satisfied Service Composition

2020 ◽  
Vol 17 (4) ◽  
pp. 55-75
Author(s):  
Xuezhi Yu ◽  
Chunyang Ye ◽  
Bingzhuo Li ◽  
Hui Zhou ◽  
Mengxing Huang

Traditional service composition methods usually address the constraint-satisfied service composition (CSSC) problem with static web services. Such solutions however are inapplicable to the dynamic scenarios where the services or their QoS values may change over time. Some recent studies are proposed to use reinforcement learning, especially, integrate the idea of Q-learning, to solve the dynamic CSSC problem. However, such Q-learning algorithm relies on Q-table to search for optimal candidate services. When the problem of CSSC becomes complex, the number of states in Q-table is very large and the cost of the Q-learning model will become extremely high. In this paper, the authors propose a novel solution to address this issue. By training a DQN network to replace the Q-table, this solution can effectively model the uncertainty of services with fine-grained QoS attributes and choose suitable candidate services to compose on the fly in the dynamic scenarios. Experimental results on both artificial and real datasets demonstrate the effectiveness of the method.

2009 ◽  
Vol 28 (12) ◽  
pp. 3268-3270
Author(s):  
Chao WANG ◽  
Jing GUO ◽  
Zhen-qiang BAO

Aerospace ◽  
2021 ◽  
Vol 8 (4) ◽  
pp. 113
Author(s):  
Pedro Andrade ◽  
Catarina Silva ◽  
Bernardete Ribeiro ◽  
Bruno F. Santos

This paper presents a Reinforcement Learning (RL) approach to optimize the long-term scheduling of maintenance for an aircraft fleet. The problem considers fleet status, maintenance capacity, and other maintenance constraints to schedule hangar checks for a specified time horizon. The checks are scheduled within an interval, and the goal is to, schedule them as close as possible to their due date. In doing so, the number of checks is reduced, and the fleet availability increases. A Deep Q-learning algorithm is used to optimize the scheduling policy. The model is validated in a real scenario using maintenance data from 45 aircraft. The maintenance plan that is generated with our approach is compared with a previous study, which presented a Dynamic Programming (DP) based approach and airline estimations for the same period. The results show a reduction in the number of checks scheduled, which indicates the potential of RL in solving this problem. The adaptability of RL is also tested by introducing small disturbances in the initial conditions. After training the model with these simulated scenarios, the results show the robustness of the RL approach and its ability to generate efficient maintenance plans in only a few seconds.


Entropy ◽  
2021 ◽  
Vol 23 (6) ◽  
pp. 737
Author(s):  
Fengjie Sun ◽  
Xianchang Wang ◽  
Rui Zhang

An Unmanned Aerial Vehicle (UAV) can greatly reduce manpower in the agricultural plant protection such as watering, sowing, and pesticide spraying. It is essential to develop a Decision-making Support System (DSS) for UAVs to help them choose the correct action in states according to the policy. In an unknown environment, the method of formulating rules for UAVs to help them choose actions is not applicable, and it is a feasible solution to obtain the optimal policy through reinforcement learning. However, experiments show that the existing reinforcement learning algorithms cannot get the optimal policy for a UAV in the agricultural plant protection environment. In this work we propose an improved Q-learning algorithm based on similar state matching, and we prove theoretically that there has a greater probability for UAV choosing the optimal action according to the policy learned by the algorithm we proposed than the classic Q-learning algorithm in the agricultural plant protection environment. This proposed algorithm is implemented and tested on datasets that are evenly distributed based on real UAV parameters and real farm information. The performance evaluation of the algorithm is discussed in detail. Experimental results show that the algorithm we proposed can efficiently learn the optimal policy for UAVs in the agricultural plant protection environment.


2021 ◽  
Vol 58 (3) ◽  
pp. 102540
Author(s):  
Xiaoyu Duan ◽  
Shi Ying ◽  
Wanli Yuan ◽  
Hailong Cheng ◽  
Xiang Yin

Sign in / Sign up

Export Citation Format

Share Document