Estimate sequential poses for wireless endoscopic capsule based on encoder-decoder convolutional neural network toward autonomous navigation

Author(s):  
Navid Panchi ◽  
Mengya Xu ◽  
Mobarakol Islam ◽  
Archana Gahiwad ◽  
Hongliang Ren
Electronics ◽  
2021 ◽  
Vol 10 (22) ◽  
pp. 2764
Author(s):  
Syed-Ali Hassan ◽  
Tariq Rahim ◽  
Soo-Young Shin

Recent advancements in the field of machine learning (ML) provide opportunity to conduct research on autonomous devices for a variety of applications. Intelligent decision-making is a critical task for self-driving systems. An attempt is made in this study to use a deep learning (DL) approach for the early detection of road cracks, potholes, and the yellow lane. The accuracy is not sufficient after training with the default model. To enhance accuracy, a convolutional neural network (CNN) model with 13 convolutional layers, a softmax layer as an output layer, and two fully connected layers (FCN) are constructed. In order to achieve the deeper propagation and to prevent saturation in the training phase, mish activation is employed in the first 12 layers with a rectified linear unit (ReLU) activation function. The upgraded CNN model performs better than the default CNN model in terms of accuracy. For the varied situation, a revised and enriched dataset for road cracks, potholes, and the yellow lane is created. The yellow lane is detected and tracked in order to move the unmanned aerial vehicle (UAV) autonomously by following yellow lane. After identifying a yellow lane, the UAV performs autonomous navigation while concurrently detecting road cracks and potholes using the robot operating system within the UAV. The performance model is benchmarked using performance measures, such as accuracy, sensitivity, F1-score, F2-score, and dice-coefficient, which demonstrate that the suggested technique produces better outcomes.


2020 ◽  
Vol 39 (4) ◽  
pp. 5475-5486
Author(s):  
Walead Kaled Sleaman ◽  
Sırma Yavuz

Robot can help human in their everyday life and routine. These are not an indoor robot which was designed to perform desired task, but they can adapt to our environment by themselves and to learn from their own experiences. In this research we focus on high degree of autonomy, which is a must for social robots. For training purpose autonomous exploration and unknown environments is used along with proper algorithm so that robot can adapt to unknown environments. For testing purpose, simulation is carried with sensor fusion method, so that real world noise can be reduced and accuracy can be increased. This dissertation focuses on the intelligent robot control in autonomous navigation tasks and investigates the robot learning in following aspects. This method is based on human instinct of imitation. In this standard real time data set is provided to the robot for training purpose, it gets train from these data and generalize over all unseen potential situations and environments. Convolutional Neural Network is used to determine the probability and based on that robot can act. After acceptable number of demonstrations, robot can predict output with high accuracy and hence can acquire the independent navigation skills. State-of-the-art reinforcement learning techniques is used to train the robot via interaction with the robots. Convolutional Neural Network is also incorporated for fast generalization. Robot is train based on all past state-action pairs collected during interaction. This training model can predict output which helps robot for autonomous navigation.


2022 ◽  
pp. 1-18
Author(s):  
Binghua Shi ◽  
Yixin Su ◽  
Cheng Lian ◽  
Chang Xiong ◽  
Yang Long ◽  
...  

Abstract Recognition of obstacle type based on visual sensors is important for navigation by unmanned surface vehicles (USV), including path planning, obstacle avoidance, and reactive control. Conventional detection techniques may fail to distinguish obstacles that are similar in visual appearance in a cluttered environment. This work proposes a novel obstacle type recognition approach that combines a dilated operator with the deep-level features map of ResNet50 for autonomous navigation. First, visual images are collected and annotated from various different scenarios for USV test navigation. Second, the deep learning model, based on a dilated convolutional neural network, is set and trained. Dilated convolution allows the whole network to learn deep features with increased receptive field and further improves the performance of obstacle type recognition. Third, a series of evaluation parameters are utilised to evaluate the obtained model, such as the mean average precision (mAP), missing rate and detection speed. Finally, some experiments are designed to verify the accuracy of the proposed approach using visual images in a cluttered environment. Experimental results demonstrate that the dilated convolutional neural network obtains better recognition performance than the other methods, with an mAP of 88%.


2018 ◽  
Vol 275 ◽  
pp. 1861-1870 ◽  
Author(s):  
Mehmet Turan ◽  
Yasin Almalioglu ◽  
Helder Araujo ◽  
Ender Konukoglu ◽  
Metin Sitti

2020 ◽  
Author(s):  
S Kashin ◽  
D Zavyalov ◽  
A Rusakov ◽  
V Khryashchev ◽  
A Lebedev

2020 ◽  
Vol 2020 (10) ◽  
pp. 181-1-181-7
Author(s):  
Takahiro Kudo ◽  
Takanori Fujisawa ◽  
Takuro Yamaguchi ◽  
Masaaki Ikehara

Image deconvolution has been an important issue recently. It has two kinds of approaches: non-blind and blind. Non-blind deconvolution is a classic problem of image deblurring, which assumes that the PSF is known and does not change universally in space. Recently, Convolutional Neural Network (CNN) has been used for non-blind deconvolution. Though CNNs can deal with complex changes for unknown images, some CNN-based conventional methods can only handle small PSFs and does not consider the use of large PSFs in the real world. In this paper we propose a non-blind deconvolution framework based on a CNN that can remove large scale ringing in a deblurred image. Our method has three key points. The first is that our network architecture is able to preserve both large and small features in the image. The second is that the training dataset is created to preserve the details. The third is that we extend the images to minimize the effects of large ringing on the image borders. In our experiments, we used three kinds of large PSFs and were able to observe high-precision results from our method both quantitatively and qualitatively.


Sign in / Sign up

Export Citation Format

Share Document