scholarly journals Behavior and Path Planning Algorithm of Autonomous Vehicle A1 in Structured Environments

2013 ◽  
Vol 46 (10) ◽  
pp. 36-41 ◽  
Author(s):  
Junsoo Kim ◽  
Kichun Jo ◽  
Dongchul Kim ◽  
Keonyup Chu ◽  
Myoungho Sunwoo
Author(s):  
Amr Mohamed ◽  
Moustafa El-Gindy ◽  
Jing Ren ◽  
Haoxiang Lang

This paper presents an optimal collision-free path planning algorithm of an autonomous multi-wheeled combat vehicle using optimal control theory and artificial potential field function (APF). The optimal path of the autonomous vehicle between a given starting and goal points is generated by an optimal path planning algorithm. The cost function of the path planning is solved together with vehicle dynamics equations to satisfy the vehicle dynamics constraints and the boundary conditions. For this purpose, a simplified four-axle bicycle model of the actual vehicle considering the vehicle body lateral and yaw dynamics while neglecting roll dynamics is used. The obstacle avoidance technique is mathematically modeled based on the proposed sigmoid function as the artificial potential field method. This potential function is assigned to each obstacle as a repulsive potential field. The inclusion of these potential fields results in a new APF which controls the steering angle of the autonomous vehicle to reach the goal point. A full nonlinear multi-wheeled combat vehicle model in TruckSim software is used for validation. This is done by importing the generated optimal path data from the introduced optimal path planning MATLAB algorithm and comparing lateral acceleration, yaw rate and curvature at different speeds (9 km/h, 28 km/h) for both simplified and TruckSim vehicle model. The simulation results show that the obtained optimal path for the autonomous multi-wheeled combat vehicle satisfies all vehicle dynamics constraints and successfully validated with TruckSim vehicle model.


Author(s):  
Kiwon Yeom ◽  

—The applications of mobile robots are more and more diverse and extensive. The motion planning of the mobile robots should be considered in aspect of effectiveness of the navigation, and collision-free motion is essential for mobile robots. In addition, dynamic path planning of unknown environment has always been a challenge for mobile robots. Aiming at navigation problems, this paper proposes a Deep Reinforcement Learning (DRL) based path planning algorithm which can navigate nonholonomic car-like mobile robots in an unknown dynamic environment. The output of the learned network are the robot’s translational and angular velocities for the next time step. The method combines path planning on a 2D grid with reinforcement learning and does not need any supervision. The experiments illustrate that our trained policy can be applied to solve complex navigation tasks. Furthermore, we compare the performance of our learned controller to the popular approaches. Keywords— Deep reinforcement learning, path planning, , artificial neural network, mobile robot, autonomous vehicle


Author(s):  
Nurul Saliha Amani Ibrahim ◽  
Faiz Asraf Saparudin

The path planning problem has been a crucial topic to be solved in autonomous vehicles. Path planning consists operations to find the route that passes through all of the points of interest in a given area. Several algorithms have been proposed and outlined in the various literature for the path planning of autonomous vehicle especially for unmanned aerial vehicles (UAV). The algorithms are not guaranteed to give full performance in each path planning cases but each one of them has their own specification which makes them suitable in sophisticated situation. This review paper evaluates several possible different path planning approaches of UAVs in terms optimal path, probabilistic completeness and computation time along with their application in specific problems.


2020 ◽  
Vol 83 (1) ◽  
pp. 133-143
Author(s):  
Sanjoy Kumar Debnath ◽  
Rosli Omar ◽  
Nor Badariyah Abdul Latip ◽  
Susama Bagchi ◽  
Elia Nadira Sabudin ◽  
...  

This paper analyses an experimental path planning performance between the Iterative Equilateral Space Oriented Visibility Graph (IESOVG) and conventional Visibility Graph (VG) algorithms in terms of computation time and path length for an autonomous vehicle. IESOVG is a path planning algorithm that was proposed to overcome the limitations of VG which is slow in obstacle-rich environment. The performance assessment was done in several identical scenarios through simulation. The results showed that the proposed IESOVG algorithm was much faster in comparison to VG. In terms of path length, IESOVG was found to have almost similar performance with VG.  It was also found that IESOVG was complete as it could find a collision-free path in all scenarios.


2021 ◽  
Vol 9 (3) ◽  
pp. 252
Author(s):  
Yushan Sun ◽  
Xiaokun Luo ◽  
Xiangrui Ran ◽  
Guocheng Zhang

This research aims to solve the safe navigation problem of autonomous underwater vehicles (AUVs) in deep ocean, which is a complex and changeable environment with various mountains. When an AUV reaches the deep sea navigation, it encounters many underwater canyons, and the hard valley walls threaten its safety seriously. To solve the problem on the safe driving of AUV in underwater canyons and address the potential of AUV autonomous obstacle avoidance in uncertain environments, an improved AUV path planning algorithm based on the deep deterministic policy gradient (DDPG) algorithm is proposed in this work. This method refers to an end-to-end path planning algorithm that optimizes the strategy directly. It takes sensor information as input and driving speed and yaw angle as outputs. The path planning algorithm can reach the predetermined target point while avoiding large-scale static obstacles, such as valley walls in the simulated underwater canyon environment, as well as sudden small-scale dynamic obstacles, such as marine life and other vehicles. In addition, this research aims at the multi-objective structure of the obstacle avoidance of path planning, modularized reward function design, and combined artificial potential field method to set continuous rewards. This research also proposes a new algorithm called deep SumTree-deterministic policy gradient algorithm (SumTree-DDPG), which improves the random storage and extraction strategy of DDPG algorithm experience samples. According to the importance of the experience samples, the samples are classified and stored in combination with the SumTree structure, high-quality samples are extracted continuously, and SumTree-DDPG algorithm finally improves the speed of the convergence model. Finally, this research uses Python language to write an underwater canyon simulation environment and builds a deep reinforcement learning simulation platform on a high-performance computer to conduct simulation learning training for AUV. Data simulation verified that the proposed path planning method can guide the under-actuated underwater robot to navigate to the target without colliding with any obstacles. In comparison with the DDPG algorithm, the stability, training’s total reward, and robustness of the improved Sumtree-DDPG algorithm planner in this study are better.


Sign in / Sign up

Export Citation Format

Share Document