scholarly journals Introducing Depth Information Into Generative Target Tracking

2021 ◽  
Vol 15 ◽  
Author(s):  
Dongyue Sun ◽  
Xian Wang ◽  
Yonghong Lin ◽  
Tianlong Yang ◽  
Shixu Wu

Common visual features used in target tracking, including colour and grayscale, are prone to failure in a confusingly similar-looking background. As the technology of three-dimensional visual information acquisition has gradually gained ground in recent years, the conditions for the wide use of depth information in target tracking has been made available. This study focuses on discussing the possible ways to introduce depth information into the generative target tracking methods based on a kernel density estimation as well as the performance of different methods of introduction, thereby providing a reference for the use of depth information in actual target tracking systems. First, an analysis of the mean-shift technical framework, a typical algorithm used for generative target tracking, is described, and four methods of introducing the depth information are proposed, i.e., the thresholding of the data source, thresholding of the density distribution of the dataset applied, weighting of the data source, and weighting of the density distribution of the dataset. Details of an experimental study conducted to evaluate the validity, characteristics, and advantages of each method are then described. The experimental results showed that the four methods can improve the validity of the basic method to a certain extent and meet the requirements of real-time target tracking in a confusingly similar background. The method of weighting the density distribution of the dataset, into which depth information is introduced, is the prime choice in engineering practise because it delivers an excellent comprehensive performance and the highest level of accuracy, whereas methods such as the thresholding of both the data sources and the density distribution of the dataset are less time-consuming. The performance in comparison with that of a state-of-the-art tracker further verifies the practicality of the proposed approach. Finally, the research results also provide a reference for improvements in other target tracking methods in which depth information can be introduced.

2020 ◽  
Vol 6 (2) ◽  
pp. eaay6036 ◽  
Author(s):  
R. C. Feord ◽  
M. E. Sumner ◽  
S. Pusdekar ◽  
L. Kalra ◽  
P. T. Gonzalez-Bellido ◽  
...  

The camera-type eyes of vertebrates and cephalopods exhibit remarkable convergence, but it is currently unknown whether the mechanisms for visual information processing in these brains, which exhibit wildly disparate architecture, are also shared. To investigate stereopsis in a cephalopod species, we affixed “anaglyph” glasses to cuttlefish and used a three-dimensional perception paradigm. We show that (i) cuttlefish have also evolved stereopsis (i.e., the ability to extract depth information from the disparity between left and right visual fields); (ii) when stereopsis information is intact, the time and distance covered before striking at a target are shorter; (iii) stereopsis in cuttlefish works differently to vertebrates, as cuttlefish can extract stereopsis cues from anticorrelated stimuli. These findings demonstrate that although there is convergent evolution in depth computation, cuttlefish stereopsis is likely afforded by a different algorithm than in humans, and not just a different implementation.


2007 ◽  
Vol 04 (02) ◽  
pp. 107-115
Author(s):  
XIAO-JUN TIAN ◽  
YUE-CHAO WANG ◽  
NING XI ◽  
ZAI-LI DONG ◽  
STEVE TUNG

Real-time force and visual information during MWCNT manipulation is required for online controlling MWCNT assembly based on atomic force microscope (AFM). Here real-time three-dimensional (3D) interactive forces between probe and sample are obtained according to PSD signals based on the proposed force model, and MWCNT manipulation process can be online displayed on the visual interface according to probe's position and applied force based on the proposed MWCNT motion model. With real-time force and visual information acquisition and feedback, the operator can control online MWCNT's manipulation process by adjusting the probe's 3D motion and applied forces. MWCNT push and assembly experiments verify the effectiveness of the method, which will be used in assembling MWCNT based nano device.


Sensors ◽  
2020 ◽  
Vol 20 (9) ◽  
pp. 2567
Author(s):  
Dong-hoon Kwak ◽  
Seung-ho Lee

Modern image processing techniques use three-dimensional (3D) images, which contain spatial information such as depth and scale, in addition to visual information. These images are indispensable in virtual reality, augmented reality (AR), and autonomous driving applications. We propose a novel method to estimate monocular depth using a cycle generative adversarial network (GAN) and segmentation. In this paper, we propose a method for estimating depth information by combining segmentation. It uses three processes: segmentation and depth estimation, adversarial loss calculations, and cycle consistency loss calculations. The cycle consistency loss calculation process evaluates the similarity of two images when they are restored to their original forms after being estimated separately from two adversarial losses. To evaluate the objective reliability of the proposed method, we compared our proposed method with other monocular depth estimation (MDE) methods using the NYU Depth Dataset V2. Our results show that the benchmark value for our proposed method is better than other methods. Therefore, we demonstrated that our proposed method is more efficient in determining depth estimation.


2013 ◽  
Vol 25 (3) ◽  
pp. 352-364 ◽  
Author(s):  
Tom Theys ◽  
Pierpaolo Pani ◽  
Johannes van Loon ◽  
Jan Goffin ◽  
Peter Janssen

Depth information is necessary for adjusting the hand to the three-dimensional (3-D) shape of an object to grasp it. The transformation of visual information into appropriate distal motor commands is critically dependent on the anterior intraparietal area (AIP) and the ventral premotor cortex (area F5), particularly the F5p sector. Recent studies have demonstrated that both AIP and the F5a sector of the ventral premotor cortex contain neurons that respond selectively to disparity-defined 3-D shape. To investigate the neural coding of 3-D shape and the behavioral role of 3-D shape-selective neurons in these two areas, we recorded single-cell activity in AIP and F5a during passive fixation of curved surfaces and during grasping of real-world objects. Similar to those in AIP, F5a neurons were either first- or second-order disparity selective, frequently showed selectivity for discrete approximations of smoothly curved surfaces that contained disparity discontinuities, and exhibited mostly monotonic tuning for the degree of disparity variation. Furthermore, in both areas, 3-D shape-selective neurons were colocalized with neurons that were active during grasping of real-world objects. Thus, area AIP and F5a contain highly similar representations of 3-D shape, which is consistent with the proposed transfer of object information from AIP to the motor system through the ventral premotor cortex.


Author(s):  
R. A. Crowther

The reconstruction of a three-dimensional image of a specimen from a set of electron micrographs reduces, under certain assumptions about the imaging process in the microscope, to the mathematical problem of reconstructing a density distribution from a set of its plane projections.In the absence of noise we can formulate a purely geometrical criterion, which, for a general object, fixes the resolution attainable from a given finite number of views in terms of the size of the object. For simplicity we take the ideal case of projections collected by a series of m equally spaced tilts about a single axis.


Choonpa Igaku ◽  
2018 ◽  
Vol 45 (2) ◽  
pp. 149-157 ◽  
Author(s):  
Ryu NAKADATE ◽  
Makoto HASHIZUME

2009 ◽  
Vol 29 (6) ◽  
pp. 1680-1682
Author(s):  
Chang-tao CHEN ◽  
Qin ZHU ◽  
Sheng-yi ZHOU ◽  
Jia-ming ZHANG

Sign in / Sign up

Export Citation Format

Share Document