Automating look-ahead schedule generation for construction using linked-data based constraint checking and reinforcement learning

2022 ◽  
Vol 134 ◽  
pp. 104069
Author(s):  
Ranjith K. Soman ◽  
Miguel Molina-Solana
2020 ◽  
Vol 120 ◽  
pp. 103369
Author(s):  
Ranjith K. Soman ◽  
Miguel Molina-Solana ◽  
Jennifer K. Whyte

2020 ◽  
Vol 34 (04) ◽  
pp. 4577-4584
Author(s):  
Xian Yeow Lee ◽  
Sambit Ghadai ◽  
Kai Liang Tan ◽  
Chinmay Hegde ◽  
Soumik Sarkar

Robustness of Deep Reinforcement Learning (DRL) algorithms towards adversarial attacks in real world applications such as those deployed in cyber-physical systems (CPS) are of increasing concern. Numerous studies have investigated the mechanisms of attacks on the RL agent's state space. Nonetheless, attacks on the RL agent's action space (corresponding to actuators in engineering systems) are equally perverse, but such attacks are relatively less studied in the ML literature. In this work, we first frame the problem as an optimization problem of minimizing the cumulative reward of an RL agent with decoupled constraints as the budget of attack. We propose the white-box Myopic Action Space (MAS) attack algorithm that distributes the attacks across the action space dimensions. Next, we reformulate the optimization problem above with the same objective function, but with a temporally coupled constraint on the attack budget to take into account the approximated dynamics of the agent. This leads to the white-box Look-ahead Action Space (LAS) attack algorithm that distributes the attacks across the action and temporal dimensions. Our results showed that using the same amount of resources, the LAS attack deteriorates the agent's performance significantly more than the MAS attack. This reveals the possibility that with limited resource, an adversary can utilize the agent's dynamics to malevolently craft attacks that causes the agent to fail. Additionally, we leverage these attack strategies as a possible tool to gain insights on the potential vulnerabilities of DRL agents.


Author(s):  
Ling Pan ◽  
Qingpeng Cai ◽  
Zhixuan Fang ◽  
Pingzhong Tang ◽  
Longbo Huang

Bike sharing provides an environment-friendly way for traveling and is booming all over the world. Yet, due to the high similarity of user travel patterns, the bike imbalance problem constantly occurs, especially for dockless bike sharing systems, causing significant impact on service quality and company revenue. Thus, it has become a critical task for bike sharing operators to resolve such imbalance efficiently. In this paper, we propose a novel deep reinforcement learning framework for incentivizing users to rebalance such systems. We model the problem as a Markov decision process and take both spatial and temporal features into consideration. We develop a novel deep reinforcement learning algorithm called Hierarchical Reinforcement Pricing (HRP), which builds upon the Deep Deterministic Policy Gradient algorithm. Different from existing methods that often ignore spatial information and rely heavily on accurate prediction, HRP captures both spatial and temporal dependencies using a divide-and-conquer structure with an embedded localized module. We conduct extensive experiments to evaluate HRP, based on a dataset from Mobike, a major Chinese dockless bike sharing company. Results show that HRP performs close to the 24-timeslot look-ahead optimization, and outperforms state-of-the-art methods in both service level and bike distribution. It also transfers well when applied to unseen areas.


IEEE Access ◽  
2018 ◽  
Vol 6 ◽  
pp. 5716-5726 ◽  
Author(s):  
Chong Wei ◽  
Yinhu Wang ◽  
Xuedong Yan ◽  
Chunfu Shao

2010 ◽  
Vol 36 (2) ◽  
pp. 289-296 ◽  
Author(s):  
Hao TANG ◽  
Hai-Feng WAN ◽  
Jiang-Hong HAN ◽  
Lei ZHOU

Sign in / Sign up

Export Citation Format

Share Document