Research on decision-making of lane-changing of automated vehicles in highway confluence area based on deep reinforcement learning

Author(s):  
Shuang Tang ◽  
Hong Shu ◽  
Yu Tang
Author(s):  
Qiong Wu ◽  
Shuo Cheng ◽  
Liang Li ◽  
Fan Yang ◽  
Li Jun Meng ◽  
...  

Intelligent decision control is one key issue of automated vehicles. Complex dynamic traffic flow and multi-requirement of passengers including vehicle safety, comfort, vehicle efficiency bring about tremendous challenges to vehicle decision making. Overtaking maneuver is a demanding task due to its large potential of traffic collision. Therefore, this paper proposes a fuzzy-inference-based reinforcement learning (FIRL) approach of autonomous overtaking decision making. Firstly, the problem of overtaking is formulated as a multi-objective Markov decision process (MDP) considering vehicle safety, driving comfort, and vehicle efficiency. Secondly, a temporal difference learning based on dynamic fuzzy (DF-TDL) is presented to learn optimized policies for autonomous overtaking decision making. Fuzzy inference is introduced to deal with continuous states and boost learning process. The RL algorithm decides whether to overtake or not based on the learned policies. Then, the automated vehicle executes local path planning and tracking. Furthermore, a simulation platform based on simulation of urban mobility (SUMO) is established to generate the random training data, that is, various traffic flows for algorithm iterative learning and validate the proposed method, extensive test results demonstrate the effectiveness of the overtaking decision-making method.


2021 ◽  
Vol 11 (1) ◽  
Author(s):  
Batel Yifrah ◽  
Ayelet Ramaty ◽  
Genela Morris ◽  
Avi Mendelsohn

AbstractDecision making can be shaped both by trial-and-error experiences and by memory of unique contextual information. Moreover, these types of information can be acquired either by means of active experience or by observing others behave in similar situations. The interactions between reinforcement learning parameters that inform decision updating and memory formation of declarative information in experienced and observational learning settings are, however, unknown. In the current study, participants took part in a probabilistic decision-making task involving situations that either yielded similar outcomes to those of an observed player or opposed them. By fitting alternative reinforcement learning models to each subject, we discerned participants who learned similarly from experience and observation from those who assigned different weights to learning signals from these two sources. Participants who assigned different weights to their own experience versus those of others displayed enhanced memory performance as well as subjective memory strength for episodes involving significant reward prospects. Conversely, memory performance of participants who did not prioritize their own experience over others did not seem to be influenced by reinforcement learning parameters. These findings demonstrate that interactions between implicit and explicit learning systems depend on the means by which individuals weigh relevant information conveyed via experience and observation.


2021 ◽  
Vol 13 (4) ◽  
pp. 1948
Author(s):  
Qiaoning Zhang ◽  
Xi Jessie Yang ◽  
Lionel P. Robert

Automated vehicles (AV) have the potential to benefit our society. Providing explanations is one approach to facilitating AV trust by decreasing uncertainty about automated decision-making. However, it is not clear whether explanations are equally beneficial for drivers across age groups in terms of trust and anxiety. To examine this, we conducted a mixed-design experiment with 40 participants divided into three age groups (i.e., younger, middle-age, and older). Participants were presented with: (1) no explanation, or (2) explanation given before or (3) after the AV took action, or (4) explanation along with a request for permission to take action. Results highlight both commonalities and differences between age groups. These results have important implications in designing AV explanations and promoting trust.


Author(s):  
Ming-Sheng Ying ◽  
Yuan Feng ◽  
Sheng-Gang Ying

AbstractMarkov decision process (MDP) offers a general framework for modelling sequential decision making where outcomes are random. In particular, it serves as a mathematical framework for reinforcement learning. This paper introduces an extension of MDP, namely quantum MDP (qMDP), that can serve as a mathematical model of decision making about quantum systems. We develop dynamic programming algorithms for policy evaluation and finding optimal policies for qMDPs in the case of finite-horizon. The results obtained in this paper provide some useful mathematical tools for reinforcement learning techniques applied to the quantum world.


Sign in / Sign up

Export Citation Format

Share Document