Using visual features to build topological maps of indoor environments

Author(s):  
P.E. Rybski ◽  
F. Zacharias ◽  
J.-F. Lett
2017 ◽  
Vol 2017 ◽  
pp. 1-18 ◽  
Author(s):  
Guanyuan Feng ◽  
Lin Ma ◽  
Xuezhi Tan

RGB-D sensors capture RGB images and depth images simultaneously, which makes it possible to acquire the depth information at pixel level. This paper focuses on the use of RGB-D sensors to construct a visual map which is an extended dense 3D map containing essential elements for image-based localization, such as poses of the database camera, visual features, and 3D structures of the building. Taking advantage of matched visual features and corresponding depth values, a novel local optimization algorithm is proposed to achieve point cloud registration and database camera pose estimation. Next, graph-based optimization is used to obtain the global consistency of the map. On the basis of the visual map, the image-based localization method is investigated, making use of the epipolar constraint. The performance of the visual map construction and the image-based localization are evaluated on typical indoor scenes. The simulation results show that the average position errors of the database camera and the query camera can be limited to within 0.2 meters and 0.9 meters, respectively.


2001 ◽  
Author(s):  
Donald A. Varakin ◽  
Sheena Rogers ◽  
Jeffrey T. Andre ◽  
Susan L. Davis

Author(s):  
Piotr Rajchowski ◽  
Jaroslaw Sadowski ◽  
Olga Blaszkiewicz ◽  
Krzysztof K. Cwalina ◽  
Alicja Olejniczak

2019 ◽  
Author(s):  
Sushrut Thorat

A mediolateral gradation in neural responses for images spanning animals to artificial objects is observed in the ventral temporal cortex (VTC). Which information streams drive this organisation is an ongoing debate. Recently, in Proklova et al. (2016), the visual shape and category (“animacy”) dimensions in a set of stimuli were dissociated using a behavioural measure of visual feature information. fMRI responses revealed a neural cluster (extra-visual animacy cluster - xVAC) which encoded category information unexplained by visual feature information, suggesting extra-visual contributions to the organisation in the ventral visual stream. We reassess these findings using Convolutional Neural Networks (CNNs) as models for the ventral visual stream. The visual features developed in the CNN layers can categorise the shape-matched stimuli from Proklova et al. (2016) in contrast to the behavioural measures used in the study. The category organisations in xVAC and VTC are explained to a large degree by the CNN visual feature differences, casting doubt over the suggestion that visual feature differences cannot account for the animacy organisation. To inform the debate further, we designed a set of stimuli with animal images to dissociate the animacy organisation driven by the CNN visual features from the degree of familiarity and agency (thoughtfulness and feelings). Preliminary results from a new fMRI experiment designed to understand the contribution of these non-visual features are presented.


2018 ◽  
Vol 482 (2) ◽  
pp. 224-227
Author(s):  
E. Mikhaylova ◽  
◽  
N. Gerasimenko ◽  
P. Prokudin ◽  
◽  
...  

Sign in / Sign up

Export Citation Format

Share Document