scholarly journals Regularized Cost-Model Oblivious Database Tuning with Reinforcement Learning

Author(s):  
Debabrota Basu ◽  
Qian Lin ◽  
Weidong Chen ◽  
Hoang Tam Vo ◽  
Zihong Yuan ◽  
...  
Author(s):  
Debabrota Basu ◽  
Qian Lin ◽  
Weidong Chen ◽  
Hoang Tam Vo ◽  
Zihong Yuan ◽  
...  

2017 ◽  
Vol 36 (10) ◽  
pp. 1073-1087 ◽  
Author(s):  
Markus Wulfmeier ◽  
Dushyant Rao ◽  
Dominic Zeng Wang ◽  
Peter Ondruska ◽  
Ingmar Posner

We present an approach for learning spatial traversability maps for driving in complex, urban environments based on an extensive dataset demonstrating the driving behaviour of human experts. The direct end-to-end mapping from raw input data to cost bypasses the effort of manually designing parts of the pipeline, exploits a large number of data samples, and can be framed additionally to refine handcrafted cost maps produced based on manual hand-engineered features. To achieve this, we introduce a maximum-entropy-based, non-linear inverse reinforcement learning (IRL) framework which exploits the capacity of fully convolutional neural networks (FCNs) to represent the cost model underlying driving behaviours. The application of a high-capacity, deep, parametric approach successfully scales to more complex environments and driving behaviours, while at deployment being run-time independent of training dataset size. After benchmarking against state-of-the-art IRL approaches, we focus on demonstrating scalability and performance on an ambitious dataset collected over the course of 1 year including more than 25,000 demonstration trajectories extracted from over 120 km of urban driving. We evaluate the resulting cost representations by showing the advantages over a carefully, manually designed cost map and furthermore demonstrate its robustness towards systematic errors by learning accurate representations even in the presence of calibration perturbations. Importantly, we demonstrate that a manually designed cost map can be refined to more accurately handle corner cases that are scarcely seen in the environment, such as stairs, slopes and underpasses, by further incorporating human priors into the training framework.


Energies ◽  
2021 ◽  
Vol 14 (22) ◽  
pp. 7749
Author(s):  
Wenying Li ◽  
Ming Tang ◽  
Xinzhen Zhang ◽  
Danhui Gao ◽  
Jian Wang

Battery energy storage systems (BESSs) are able to facilitate economical operation of the grid through demand response (DR), and are regarded as the most significant DR resource. Among them, distributed BESS integrating home photovoltaics (PV) have developed rapidly, and account for nearly 40% of newly installed capacity. However, the use scenarios and use efficiency of distributed BESS are far from sufficient to be able to utilize the potential loads and overcome uncertainties caused by disorderly operation. In this paper, the low-voltage transformer-powered area (LVTPA) is firstly defined, and then a DR grid edge controller was implemented based on deep reinforcement learning to maximize the total DR benefits and promote three-phase balance in the LVTPA. The proposed DR problem is formulated as a Markov decision process (MDP). In addition, the deep deterministic policy gradient (DDPG) algorithm is applied to train the controller in order to learn the optimal DR strategy. Additionally, a life cycle cost model of the BESS is established and implemented in the DR scheme to measure the income. The numerical results, compared to deep Q learning and model-based methods, demonstrate the effectiveness and validity of the proposed method.


2021 ◽  
pp. 101-110
Author(s):  
Yanfeng Chai ◽  
Jiake Ge ◽  
Yunpeng Chai ◽  
Xin Wang ◽  
BoXuan Zhao

1994 ◽  
Vol 11 (1) ◽  
pp. 47-56
Author(s):  
Virginia C. Day ◽  
Zachary F. Lansdowne ◽  
Richard A Moynihan ◽  
John A. Vitkevich

Decision ◽  
2016 ◽  
Vol 3 (2) ◽  
pp. 115-131 ◽  
Author(s):  
Helen Steingroever ◽  
Ruud Wetzels ◽  
Eric-Jan Wagenmakers

Sign in / Sign up

Export Citation Format

Share Document