vision based localization
Recently Published Documents


TOTAL DOCUMENTS

136
(FIVE YEARS 34)

H-INDEX

16
(FIVE YEARS 2)

Author(s):  
Xiliang Yin ◽  
Lin Ma ◽  
Ping Sun ◽  
Xuezhi Tan

AbstractRecently, deep learning and vision-based technologies have shown their great significance for the prospective development of smart Internet of Vehicle (IoV). When the smart vehicle enters the indoor parking of a shopping mall, the vision-based localization technology can provide reliable parking service. As known, the vision-based technique relies on a visual map without a change in the position of the reference object. Although, some researchers have proposed a few automatic visual fingerprinting (AVF) methods, which are aiming at reducing the cost of building the visual map database. However, the AVF method still costs too much under such a situation, since it is impossible to determine the specific location of the displaced object. Given the smart IoV and the development of deep learning approach, we propose an algorithm for solving the problem based on crowdsourcing and deep learning in this paper. Firstly, we propose a Region-based Fully Convolutional Network (R-FCN) based method with the feedback of crowdsourced images to locate the specific displaced object in the visual map database. Secondly, we propose a method based on quadratic programming (QP) for solving the translation vector of the displaced objects, which finally solves the problem of updating the visual map database. The simulation results show that our method can provide a higher detection sensitivity and correction accuracy as well as the relocation results. It means that our proposed algorithm outperforms the compared one, which is verified by both synthetic and real data simulation.


2021 ◽  
Author(s):  
N. Q. Pham ◽  
K. A. Mekonnen ◽  
A. Mefleh ◽  
A. M. J. Koonen ◽  
E. Tangdiongga

2021 ◽  
Author(s):  
Xiliang Yin ◽  
Lin Ma ◽  
Ping Sun

Abstract Recently, the deep learning and vision-based technologies has shown their great significance for the prospective development of smart Internet of Vehicle (IoV). When the smart vehicle enters the indoor parking of a shopping mall, the vision-based localization technology can provide reliable parking service. As known, the vision-based technique relies on a visual map without a change in the position of the reference object. Although, some researchers have proposed a few automatic visual fingerprinting (AVF) methods, which are aiming at reducing the cost of building the visual map database. However, the AVF method still costs too much under such situation, since it is impossible to determine the specific location of the displaced object. In view of the smart IoV and the development of deep learning approach, we propose a crowdsourcing and deep learning based algorithm for solving the problem in this paper. Firstly, we propose a Region-based Fully Convolutional Network (R-FCN) based method with the feedback of crowdsourced images to locate the specific displaced object in the visual map database. Secondly, we propose a method based on quadratic programming (QP) for solving the translation vector of the displaced objects, which finally solves the problem of updating the visual map database. The simulation results show that our method can provide a higher detection sensitivity and correction accuracy as well as the relocation results. It means that our proposed algorithm outperforms the compared one, which is verified by both synthetic and real data simulation.


CONVERTER ◽  
2021 ◽  
pp. 397-406
Author(s):  
Shuwen Pan, Yuanyuan Li, Pengying Du, Yan Liu

This paper designed an intelligent service robot system in highway based on multi-sensor fusion. The mobile robot attempts to fuse the lidar information and monocular vision information to estimate the pose of itself and obtain an environmental map. It adapts a new SLAM method which combines lidar and vision information. Lidar is used to obtain the 2D occupancy grid map and the monocular vision SLAM algorithm uses the Extended Kalman Filter (EKF) to magnify the pose estimation. The 3-DOF pose provided by lidar is obtained through Cartographer algorithm and the monocular vision SLAM who offers the 6-DOF pose is realized with ORB-SLAM. The experimental results show that the system is effective in application as an intelligent service robot of highway.


Author(s):  
A. Elashry ◽  
B. Sluis ◽  
C. Toth

Abstract. Feature Matching between images is an essential task for many computer vision and photogrammetry applications, such as Structure from Motion (SFM), Surface Extraction, Visual Simultaneous Localization and Mapping (VSLAM), and vision-based localization and navigation. Among the matched point pairs, there are typically false positive matches. Therefore, outlier detection and rejection are important steps in any vision application. RANSAC has been a well-established approach for outlier detection. The outlier ratio and the number of required correspondences used in RANSAC determine the number of iterations needed, which ultimately, determines the computation time. We propose a simple algorithm (GR_RANSAC) based on the two-dimensional spatial relationships between points in the image domain. The assumption is that the distances and bearing angles between the 2D feature points should be similar in images with small disparity, such as the case for video image sequences. In the proposed approach, the distances and angles are measured from a reference point in the first image and its correspondence in the other image, and the points with any significant differences are considered as outliers. This process can pre-filter the matched points, and thus increase the inliers’ ratio. As a result, GR_RANSAC can converge to the correct hypothesis in fewer trial runs than ordinary RANSAC.


2021 ◽  
Vol 7 (2) ◽  
pp. 20
Author(s):  
Carlos Lassance ◽  
Yasir Latif ◽  
Ravi Garg ◽  
Vincent Gripon ◽  
Ian Reid

Vision-based localization is the problem of inferring the pose of the camera given a single image. One commonly used approach relies on image retrieval where the query input is compared against a database of localized support examples and its pose is inferred with the help of the retrieved items. This assumes that images taken from the same places consist of the same landmarks and thus would have similar feature representations. These representations can learn to be robust to different variations in capture conditions like time of the day or weather. In this work, we introduce a framework which aims at enhancing the performance of such retrieval-based localization methods. It consists in taking into account additional information available, such as GPS coordinates or temporal proximity in the acquisition of the images. More precisely, our method consists in constructing a graph based on this additional information that is later used to improve reliability of the retrieval process by filtering the feature representations of support and/or query images. We show that the proposed method is able to significantly improve the localization accuracy on two large scale datasets, as well as the mean average precision in classical image retrieval scenarios.


Author(s):  
Lin Bai ◽  
Yang Yang ◽  
Mingzhe Chen ◽  
Chunyan Feng ◽  
Caili Guo ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document