Visual Recognition Based on Temporal Cortex Cells: Viewer-Centred Processing of Pattern Configuration

1998 ◽  
Vol 53 (7-8) ◽  
pp. 518-541 ◽  
Author(s):  
David I. Perrett ◽  
Mike W. Oram

Abstract A model of recognition is described based on cell properties in the ventral cortical stream of visual processing in the primate brain. At a critical intermediate stage in this system, ‘Elaborate’ feature sensitive cells respond selectively to visual features in a way that depends on size (± 1 octave), orientation (± 4 5 °) but does not depend on position within central vision (± 5 °). These features are simple conjunctions of 2-D elements (e.g. a horizontal dark area above a dark smoothly convex area). They can arise either as elements of an object’s surface pattern or as a 3-D component bounded by an object’s external contour. By requiring a combination of several such features without regard to their position within the central region of the visual image, ‘Pattern’ sensitive cells at higher levels can exhibit selectivity for complex configurations that typify objects seen under particular viewing conditions. Given that input features to such Pattern sensitive cells are specified in approximate size and orientation, initial cellular ‘representations’ of the visual appearance of object type (or object example) are also selective for orientation and size. At this level, sensitivity to object view (± 6 0 °) arises because visual features disappear as objects are rotated in perspective. Processing is thus viewer-centred and the neurones only respond to objects seen from particular viewing conditions or ‘object instances’. Combined sensitivity to multiple features (conjunctions of elements) independent of their position, establishes selectivity for the configurations of ob­ject parts (from one view) because rearranged configurations of the same parts yield images lacking some of the 2-D visual features present in the normal configuration. Different neural populations appear to be selectively tuned to particular components of the same biological object (e.g. face, eyes, hands, legs), perhaps because the independent articulation of these components gives rise to correlated activity in different sets of input visual features. Generalisation over viewing conditions for a given object can be established by hierarchically pooling outputs of view-condition specific cells with pooling operations dependent on the continuity in experience across viewing conditions. Different object parts are seen together and different views are seen in succession when the observer walks around the object. The view specific coding that characterises the selectivity of cells in the temporal lobe can be seen as a natural consequence of selective experience of objects from particular vantage points. View specific coding for the face and body also has great utility in understand­ing complex social signals, a property that may not be feasible with object-centred processing.

2020 ◽  
Author(s):  
Bahareh Jozranjbar ◽  
Arni Kristjansson ◽  
Heida Maria Sigurdardottir

While dyslexia is typically described as a phonological deficit, recent evidence suggests that ventral stream regions, important for visual categorization and object recognition, are hypoactive in dyslexic readers who might accordingly show visual recognition deficits. By manipulating featural and configural information of faces and houses, we investigated whether dyslexic readers are disadvantaged at recognizing certain object classes or utilizing particular visual processing mechanisms. Dyslexic readers found it harder to recognize objects (houses), suggesting that visual problems in dyslexia are not completely domain-specific. Mean accuracy for faces was equivalent in the two groups, compatible with domain-specificity in face processing. While face recognition abilities correlated with reading ability, lower house accuracy was nonetheless related to reading difficulties even when accuracy for faces was kept constant, suggesting a specific relationship between visual word recognition and the recognition of non-face objects. Representational similarity analyses (RSA) revealed that featural and configural processes were clearly separable in typical readers, while dyslexic readers appeared to rely on a single process. This occurred for both faces and houses and was not restricted to particular visual categories. We speculate that reading deficits in some dyslexic readers reflect their reliance on a single process for object recognition.


2019 ◽  
Author(s):  
Sushrut Thorat

A mediolateral gradation in neural responses for images spanning animals to artificial objects is observed in the ventral temporal cortex (VTC). Which information streams drive this organisation is an ongoing debate. Recently, in Proklova et al. (2016), the visual shape and category (“animacy”) dimensions in a set of stimuli were dissociated using a behavioural measure of visual feature information. fMRI responses revealed a neural cluster (extra-visual animacy cluster - xVAC) which encoded category information unexplained by visual feature information, suggesting extra-visual contributions to the organisation in the ventral visual stream. We reassess these findings using Convolutional Neural Networks (CNNs) as models for the ventral visual stream. The visual features developed in the CNN layers can categorise the shape-matched stimuli from Proklova et al. (2016) in contrast to the behavioural measures used in the study. The category organisations in xVAC and VTC are explained to a large degree by the CNN visual feature differences, casting doubt over the suggestion that visual feature differences cannot account for the animacy organisation. To inform the debate further, we designed a set of stimuli with animal images to dissociate the animacy organisation driven by the CNN visual features from the degree of familiarity and agency (thoughtfulness and feelings). Preliminary results from a new fMRI experiment designed to understand the contribution of these non-visual features are presented.


2009 ◽  
Vol 102 (6) ◽  
pp. 3469-3480 ◽  
Author(s):  
H. M. Van Ettinger-Veenstra ◽  
W. Huijbers ◽  
T. P. Gutteling ◽  
M. Vink ◽  
J. L. Kenemans ◽  
...  

It is well known that parts of a visual scene are prioritized for visual processing, depending on the current situation. How the CNS moves this focus of attention across the visual image is largely unknown, although there is substantial evidence that preparation of an action is a key factor. Our results support the view that direct corticocortical feedback connections from frontal oculomotor areas to the visual cortex are responsible for the coupling between eye movements and shifts of visuospatial attention. Functional magnetic resonance imaging (fMRI)–guided transcranial magnetic stimulation (TMS) was applied to the frontal eye fields (FEFs) and intraparietal sulcus (IPS). A single pulse was delivered 60, 30, or 0 ms before a discrimination target was presented at, or next to, the target of a saccade in preparation. Results showed that the known enhancement of discrimination performance specific to locations to which eye movements are being prepared was enhanced by early TMS on the FEF contralateral to eye movement direction, whereas TMS on the IPS resulted in a general performance increase. The current findings indicate that the FEF affects selective visual processing within the visual cortex itself through direct feedback projections.


2013 ◽  
Vol 25 (4) ◽  
pp. 547-557 ◽  
Author(s):  
Maital Neta ◽  
William M. Kelley ◽  
Paul J. Whalen

Extant research has examined the process of decision making under uncertainty, specifically in situations of ambiguity. However, much of this work has been conducted in the context of semantic and low-level visual processing. An open question is whether ambiguity in social signals (e.g., emotional facial expressions) is processed similarly or whether a unique set of processors come on-line to resolve ambiguity in a social context. Our work has examined ambiguity using surprised facial expressions, as they have predicted both positive and negative outcomes in the past. Specifically, whereas some people tended to interpret surprise as negatively valenced, others tended toward a more positive interpretation. Here, we examined neural responses to social ambiguity using faces (surprise) and nonface emotional scenes (International Affective Picture System). Moreover, we examined whether these effects are specific to ambiguity resolution (i.e., judgments about the ambiguity) or whether similar effects would be demonstrated for incidental judgments (e.g., nonvalence judgments about ambiguously valenced stimuli). We found that a distinct task control (i.e., cingulo-opercular) network was more active when resolving ambiguity. We also found that activity in the ventral amygdala was greater to faces and scenes that were rated explicitly along the dimension of valence, consistent with findings that the ventral amygdala tracks valence. Taken together, there is a complex neural architecture that supports decision making in the presence of ambiguity: (a) a core set of cortical structures engaged for explicit ambiguity processing across stimulus boundaries and (b) other dedicated circuits for biologically relevant learning situations involving faces.


eLife ◽  
2017 ◽  
Vol 6 ◽  
Author(s):  
Ivan Larderet ◽  
Pauline MJ Fritsch ◽  
Nanae Gendre ◽  
G Larisa Neagu-Maier ◽  
Richard D Fetter ◽  
...  

Visual systems transduce, process and transmit light-dependent environmental cues. Computation of visual features depends on photoreceptor neuron types (PR) present, organization of the eye and wiring of the underlying neural circuit. Here, we describe the circuit architecture of the visual system of Drosophila larvae by mapping the synaptic wiring diagram and neurotransmitters. By contacting different targets, the two larval PR-subtypes create two converging pathways potentially underlying the computation of ambient light intensity and temporal light changes already within this first visual processing center. Locally processed visual information then signals via dedicated projection interneurons to higher brain areas including the lateral horn and mushroom body. The stratified structure of the larval optic neuropil (LON) suggests common organizational principles with the adult fly and vertebrate visual systems. The complete synaptic wiring diagram of the LON paves the way to understanding how circuits with reduced numerical complexity control wide ranges of behaviors.


2015 ◽  
Vol 282 (1799) ◽  
pp. 20142384 ◽  
Author(s):  
Aurore Avarguès-Weber ◽  
Adrian G. Dyer ◽  
Noha Ferrah ◽  
Martin Giurfa

Traditional models of insect vision have assumed that insects are only capable of low-level analysis of local cues and are incapable of global, holistic perception. However, recent studies on honeybee ( Apis mellifera ) vision have refuted this view by showing that this insect also processes complex visual information by using spatial configurations or relational rules. In the light of these findings, we asked whether bees prioritize global configurations or local cues by setting these two levels of image analysis in competition. We trained individual free-flying honeybees to discriminate hierarchical visual stimuli within a Y-maze and tested bees with novel stimuli in which local and/or global cues were manipulated. We demonstrate that even when local information is accessible, bees prefer global information, thus relying mainly on the object's spatial configuration rather than on elemental, local information. This preference can be reversed if bees are pre-trained to discriminate isolated local cues. In this case, bees prefer the hierarchical stimuli with the local elements previously primed even if they build an incorrect global configuration. Pre-training with local cues induces a generic attentional bias towards any local elements as local information is prioritized in the test, even if the local cues used in the test are different from the pre-trained ones. Our results thus underline the plasticity of visual processing in insects and provide new insights for the comparative analysis of visual recognition in humans and animals.


1991 ◽  
Vol 3 (1) ◽  
pp. 9-24 ◽  
Author(s):  
M. H. Harries ◽  
D. I. Perrett

Physiological recordings along the length of the upper bank of the superior temporal sulcus (STS) revealed cells each of which was selectively responsive to a particular view of the head and body. Such cells were grouped in large patches 3-4 mm across. The patches were separated by regions of cortex containing cells responsive to other stimuli. The distribution of cells projecting from temporal cortex to the posterior regions of the inferior parietal lobe was studied with retrogradely transported fluorescent dyes. A strong temporoparietal projection was found originating from the upper bank of the STS. Cells projecting to the parietal cortex occurred in large patches or bands. The size and periodicity of modules defined through anatomical connections matched the functional subdivisions of the STS cortex involved in face processing evident in physiological recordings. It is speculated that the temporoparietal projections could provide a route through which temporal lobe analysis of facial signals about the direction of others' attention can be passed to parietal systems concerned with spatial awareness.


2018 ◽  
Vol 48 (10) ◽  
pp. 1608-1615 ◽  
Author(s):  
G. Donohoe ◽  
J. Holland ◽  
D. Mothersill ◽  
S. McCarthy-Jones ◽  
D. Cosgrove ◽  
...  

AbstractBackgroundThe longstanding association between the major histocompatibility complex (MHC) locus and schizophrenia (SZ) risk has recently been accounted for, partially, by structural variation at the complement component 4 (C4) gene. This structural variation generates varying levels ofC4RNA expression, and genetic information from the MHC region can now be used to predictC4RNA expression in the brain. Increased predictedC4ARNA expression is associated with the risk of SZ, andC4is reported to influence synaptic pruning in animal models.MethodsBased on our previous studies associating MHC SZ risk variants with poorer memory performance, we tested whether increased predictedC4ARNA expression was associated with reduced memory function in a large (n= 1238) dataset of psychosis cases and healthy participants, and with altered task-dependent cortical activation in a subset of these samples.ResultsWe observed that increased predictedC4ARNA expression predicted poorer performance on measures of memory recall (p= 0.016, corrected). Furthermore, in healthy participants, we found that increased predictedC4ARNA expression was associated with a pattern of reduced cortical activity in middle temporal cortex during a measure of visual processing (p< 0.05, corrected).ConclusionsThese data suggest that the effects ofC4on cognition were observable at both a cortical and behavioural level, and may represent one mechanism by which illness risk is mediated. As such, deficits in learning and memory may represent a therapeutic target for new molecular developments aimed at alteringC4’s developmental role.


Psych ◽  
2020 ◽  
Vol 2 (4) ◽  
pp. 287-295
Author(s):  
Satoshi Shibuya

Various devices have been developed to enable humans to control remote objects using active hand movements. However, it is still unclear how the visual characteristics of a synchronously moving object influences hand movements. This study investigates the effects of visual appearance and orientation of a hand-controlled object on hand movements using a novel visuomotor task. The motion of a visual image on a monitor reflected the participants’ right hand movements in the forwards-backwards direction, but not in the lateral direction (i.e., the lateral position of the image was fixed). Participants performed continuous goal-directed back and forth movements of the visual image for one minute. The image’s appearance (hand and arrow) and orientation (forward (FW), leftward (LW), and rightward (RW)) were manipulated. Unconscious lateral deviations (i.e., drift movements) of the participant’s hand during the task were evaluated. Regardless of appearance, the leftward and rightward image induced leftward and rightward drift movements, compared to the forward image. However, the modulation sizes were similar using arrow images, but not using hand images. Specifically, anatomically plausible hand images elicited greater drift movements than anatomically implausible images. This suggests that both orientation and appearance of a hand-controlled object influences hand movements according to stimulus-response compatibility and body-representation changes.


2012 ◽  
Vol 108 (11) ◽  
pp. 3073-3086 ◽  
Author(s):  
Arjun K. Bansal ◽  
Jedediah M. Singer ◽  
William S. Anderson ◽  
Alexandra Golby ◽  
Joseph R. Madsen ◽  
...  

The cerebral cortex needs to maintain information for long time periods while at the same time being capable of learning and adapting to changes. The degree of stability of physiological signals in the human brain in response to external stimuli over temporal scales spanning hours to days remains unclear. Here, we quantitatively assessed the stability across sessions of visually selective intracranial field potentials (IFPs) elicited by brief flashes of visual stimuli presented to 27 subjects. The interval between sessions ranged from hours to multiple days. We considered electrodes that showed robust visual selectivity to different shapes; these electrodes were typically located in the inferior occipital gyrus, the inferior temporal cortex, and the fusiform gyrus. We found that IFP responses showed a strong degree of stability across sessions. This stability was evident in averaged responses as well as single-trial decoding analyses, at the image exemplar level as well as at the category level, across different parts of visual cortex, and for three different visual recognition tasks. These results establish a quantitative evaluation of the degree of stationarity of visually selective IFP responses within and across sessions and provide a baseline for studies of cortical plasticity and for the development of brain-machine interfaces.


Sign in / Sign up

Export Citation Format

Share Document