A Q-Leaning-Based On-Line Planning Approach to Autonomous Architecture Discovery for Self-managed Software

Author(s):  
Dongsun Kim ◽  
Sooyong Park
2018 ◽  
Vol 8 (5) ◽  
pp. 3321-3328 ◽  
Author(s):  
Ι. Marouani ◽  
A. Boudjemline ◽  
T. Guesmi ◽  
H. H. Abdallah

This paper presents an improved artificial bee colony (ABC) technique for solving the dynamic economic emission dispatch (DEED) problem. Ramp rate limits, valve-point loading effects and prohibited operating zones (POZs) have been considered. The proposed technique integrates the grenade explosion method and Cauchy operator in the original ABC algorithm, to avoid random search mechanism. However, the DEED is a multi-objective optimization problem with two conflicting criteria which need to be minimized simultaneously. Thus, it is recommended to provide the best solution for the decision-makers. Shannon’s entropy-based method is used for the first time within the context of the on-line planning of generator outputs to extract the best compromise solution among the Pareto set. The robustness of the proposed technique is verified on six-unit and ten-unit system tests. Results proved that the proposed algorithm gives better optimum solutions in comparison with more than ten metaheuristic techniques.


2017 ◽  
Vol 26 (05) ◽  
pp. 1760018 ◽  
Author(s):  
Aurélie Beynier

Multiagent patrolling is the problem faced by a set of agents that have to visit a set of sites to prevent or detect some threats or illegal actions. Although it is commonly assumed that patrollers share a common objective, the issue of cooperation between the patrollers has received little attention. Over the last years, the focus has been put on patrolling strategies to prevent a one-shot attack from an adversary. This adversary is usually assumed to be fully rational and to have full observability of the system. Most approaches are then based on game theory and consists in computing a best response strategy. Nonetheless, when patrolling frontiers, detecting illegal fishing or poaching; patrollers face multiple adversaries with limited observability and rationality. Moreover, adversaries can perform multiple illegal actions over time and space and may change their strategies as time passes. In this paper, we propose a multiagent planning approach that enables effective cooperation between a team of patrollers in uncertain environments. Patrolling agents are assumed to have partial observability of the system. Our approach allows the patrollers to learn a generic and stochastic model of the adversaries based on the history of observations. A wide variety of adversaries can thus be considered with strategies ranging from random behaviors to fully rational and informed behaviors. We show that the multiagent planning problem can be formalized by a non-stationary DEC- POMDP. In order to deal with the non-stationary, we introduce the notion of context. We then describe an evolutionary algorithm to compute patrolling strategies on-line, and we propose methods to improve the patrollers’ performance.


1995 ◽  
Vol 13 (8) ◽  
pp. 1160-1168 ◽  
Author(s):  
Akihiko Matsushita ◽  
Takeshi Tsuchiya

1994 ◽  
Vol 30 (8) ◽  
pp. 917-925 ◽  
Author(s):  
Akihiko MATSUSHITA ◽  
Takeshi TSUCHIYA
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document