scholarly journals A Theory of Size-Invariance in Human Object Recognition

2010 ◽  
Vol 10 (7) ◽  
pp. 1006-1006
Author(s):  
L. Zhao
1989 ◽  
Vol 12 (3) ◽  
pp. 381-397 ◽  
Author(s):  
Gary W. Strong ◽  
Bruce A. Whitehead

AbstractPurely parallel neural networks can model object recognition in brief displays – the same conditions under which illusory conjunctions (the incorrect combination of features into perceived objects in a stimulus array) have been demonstrated empirically (Treisman 1986; Treisman & Gelade 1980). Correcting errors of illusory conjunction is the “tag-assignment” problem for a purely parallel processor: the problem of assigning a spatial tag to nonspatial features, feature combinations, and objects. This problem must be solved to model human object recognition over a longer time scale. Our model simulates both the parallel processes that may underlie illusory conjunctions and the serial processes that may solve the tag-assignment problem in normal perception. One component of the model extracts pooled features and another provides attentional tags that correct illusory conjunctions. Our approach addresses two questions: (i) How can objects be identified from simultaneously attended features in a parallel, distributed representation? (ii) How can the spatial selectional requirements of such an attentional process be met by a separation of pathways for spatial and nonspatial processing? Our analysis of these questions yields a neurally plausible simulation of tag assignment based on synchronizing feature processing activity in a spatial focus of attention.


Perception ◽  
1997 ◽  
Vol 26 (1_suppl) ◽  
pp. 33-33
Author(s):  
G M Wallis ◽  
H H Bülthoff

The view-based approach to object recognition supposes that objects are stored as a series of associated views. Although representation of these views as combinations of 2-D features allows generalisation to similar views, it remains unclear how very different views might be associated together to allow recognition from any viewpoint. One cue present in the real world other than spatial similarity, is that we usually experience different objects in temporally constrained, coherent order, and not as randomly ordered snapshots. In a series of recent neural-network simulations, Wallis and Baddeley (1997 Neural Computation9 883 – 894) describe how the association of views on the basis of temporal as well as spatial correlations is both theoretically advantageous and biologically plausible. We describe an experiment aimed at testing their hypothesis in human object-recognition learning. We investigated recognition performance of faces previously presented in sequences. These sequences consisted of five views of five different people's faces, presented in orderly sequence from left to right profile in 45° steps. According to the temporal-association hypothesis, the visual system should associate the images together and represent them as different views of the same person's face, although in truth they are images of different people's faces. In a same/different task, subjects were asked to say whether two faces seen from different viewpoints were views of the same person or not. In accordance with theory, discrimination errors increased for those faces seen earlier in the same sequence as compared with those faces which were not ( p<0.05).


2000 ◽  
Vol 40 (5) ◽  
pp. 473-484 ◽  
Author(s):  
Christopher S. Furmanski ◽  
Stephen A. Engel

2014 ◽  
Vol 17 (3) ◽  
pp. 455-462 ◽  
Author(s):  
Radoslaw Martin Cichy ◽  
Dimitrios Pantazis ◽  
Aude Oliva

Perception ◽  
1996 ◽  
Vol 25 (1_suppl) ◽  
pp. 148-148
Author(s):  
B J Stankiewicz ◽  
J E Hummel

Researchers in the field of visual perception have dedicated a great deal of effort to understanding how humans recognise known objects from novel viewpoints (often referred to as shape constancy). This research has produced a variety of theories—some that emphasise the use of invariant representations, others that emphasise alignment processes used in conjunction with viewpoint-specific representations. Although researchers disagree on the specifics of the representations and processes used during human object recognition, most agree that achieving shape constancy is computationally expensive—that is, it requires work. If it is assumed that attention provides the necessary resources for these computations, these theories suggest that recognition with attention should be qualitatively different from recognition without attention. Specifically, recognition with attention should be more invariant with viewpoint than recognition without attention. We recently reported a series of experiments, in which we used a response-time priming paradigm in which attention and viewpoint were manipulated, that showed attention is necessary for generating a representation of shape that is invariant with left-right reflection. We are now reporting new experiments showing that shape representation activated without attention is not completely view-specific. These experiments demonstrate that the automatic shape representation is invariant with the size and location of an image in the visual field. The results are reported in the context of a recent model proposed by Hummel and Stankiewicz ( Attention and Performance16 in press), as well as in the context of other models of human object recognition that make explicit predictions about the role of attention in generating a viewpoint-invariant representation of object shape.


2007 ◽  
Author(s):  
Trevor Wolff ◽  
Jeremiah D. Still ◽  
Derrick J. Parkhurst ◽  
Veronica J. Dark

Sign in / Sign up

Export Citation Format

Share Document