A Convolutional Neural Network for End to End Structural Prediction and Lane Detection for Autonomous Vehicle

Author(s):  
Md. Rashidul Islam ◽  
Taushif Ahmed Siddique ◽  
Md. Imtiazul Hoque Sakib ◽  
Shafayat Hossain
Webology ◽  
2020 ◽  
Vol 17 (2) ◽  
pp. 694-705
Author(s):  
T. Kirthiga Devi ◽  
Akshat Srivatsava ◽  
Kritesh Kumar Mudgal ◽  
Ranjnish Raj Jayanti ◽  
T. Karthick

The objective of this project is to automate the process of driving a car. The result of this project will surely reduce the number of hazards happening everyday. Our world is in progress and self driving car is on its way to reach consumer‟s door-step but the big question still lies that will people accept such a car which is fully automated and driverless. The idea is to create an autonomous Vehicle that uses only some sensors (collision detectors, temperature detectors etc.) and camera module to travel between destinations with minimal/no human intervention. The car will be using a trained Convolutional Neural Network (CNN) which would control the parameters that are required for smoothly driving a car. They are directly connected to the main steering mechanism and the output of the deep learning model will control the steering angle of the vehicle. Many algorithms like Lane Detection, Object Detection are used in tandem to provide the necessary functionalities in the car.


2021 ◽  
Vol 13 (2) ◽  
pp. 274
Author(s):  
Guobiao Yao ◽  
Alper Yilmaz ◽  
Li Zhang ◽  
Fei Meng ◽  
Haibin Ai ◽  
...  

The available stereo matching algorithms produce large number of false positive matches or only produce a few true-positives across oblique stereo images with large baseline. This undesired result happens due to the complex perspective deformation and radiometric distortion across the images. To address this problem, we propose a novel affine invariant feature matching algorithm with subpixel accuracy based on an end-to-end convolutional neural network (CNN). In our method, we adopt and modify a Hessian affine network, which we refer to as IHesAffNet, to obtain affine invariant Hessian regions using deep learning framework. To improve the correlation between corresponding features, we introduce an empirical weighted loss function (EWLF) based on the negative samples using K nearest neighbors, and then generate deep learning-based descriptors with high discrimination that is realized with our multiple hard network structure (MTHardNets). Following this step, the conjugate features are produced by using the Euclidean distance ratio as the matching metric, and the accuracy of matches are optimized through the deep learning transform based least square matching (DLT-LSM). Finally, experiments on Large baseline oblique stereo images acquired by ground close-range and unmanned aerial vehicle (UAV) verify the effectiveness of the proposed approach, and comprehensive comparisons demonstrate that our matching algorithm outperforms the state-of-art methods in terms of accuracy, distribution and correct ratio. The main contributions of this article are: (i) our proposed MTHardNets can generate high quality descriptors; and (ii) the IHesAffNet can produce substantial affine invariant corresponding features with reliable transform parameters.


2021 ◽  
Vol 10 (1) ◽  
pp. 31-36
Author(s):  
Junghwan Lee ◽  
Heesang Eom ◽  
Yuli Sun Hariyani ◽  
Cheonjung Kim ◽  
Yongkyoung Yoo ◽  
...  

2021 ◽  
Vol 7 (4) ◽  
pp. 117
Author(s):  
Linling Fang ◽  
Yingle Fan

<p>A biomimetic vision computing model based on multi-level feature channel optimization coding is proposed and applied to image contour detection, combining the end-to-end detection method of full convolutional neural network and the traditional contour detection method based on biological vision mechanism. Considering the effectiveness of the Gabor filter in perceiving the scale and direction of the image target, the Gabor filter is introduced to simulate the multi-level feature response on the visual path. The optimal scale and direction of the Gabor filter are obtained based on the similarity index, and they are used as the frequency separation parameter of the NSCT transform. The contour sub-image obtained by the NSCT transform is combined with the original image for feature enhancement and fusion to realize the primary contour response. The low-dimensional and low-redundancy primary contour response is used as the input sample of the network model to relieve network pressure and reduce computational complexity. A fully improved convolutional neural network model is constructed for multi-scale training, through feature encoder to feature decoder, to achieve end-to-end pixel prediction, and obtain a complete and continuous detection image of the subject contour. Using the BSDS500 atlas as the experimental sample, the average accuracy index is 0.85, which runs on the device CPU at a detection rate of 20+ FPS to achieve a good balance between training efficiency and detection effect.</p>


Sign in / Sign up

Export Citation Format

Share Document