scholarly journals Dissociable signatures of visual salience and behavioral relevance across attentional priority maps in human cortex

2017 ◽  
Author(s):  
Thomas C. Sprague ◽  
Sirawaj Itthipuripat ◽  
Vy A. Vo ◽  
John T. Serences

AbstractComputational models posit that visual attention is guided by activity within spatial maps that index the image-computable salience and the behavioral relevance of objects in the scene. However, the simultaneous influence of these factors on putative neural ‘attentional priority maps’ in human cortex is not well understood. We tested the hypothesis that visual salience and behavioral relevance independently impact the activation profile across retinotopically-organized cortical regions by quantifying attentional priority maps measured in human brains using functional MRI while participants attended one of two differentially-salient stimuli. We find that the topography of activation in priority maps, as reflected in the modulation of region-level patterns of population activity, independently indexed the physical salience and behavioral relevance of each scene element. Moreover, salience strongly impacted activation patterns in early visual areas, whereas later visual areas were dominated by relevance. This suggests that prioritizing spatial locations relies on distributed neural codes containing graded representations of salience and relevance across the visual hierarchy.Significance StatementOften, it is necessary to orient towards bright, unique, or sudden events in the environment – that is, salient stimuli. However, we can focus processing resources on less salient visual information if it is relevant to the task at hand. We tested a theory which supposes that we represent different scene elements according to both their salience and their relevance in a series of ‘priority maps’ by measuring fMRI activation patterns across the human brain and reconstructing spatial maps of the visual scene under different task conditions. We found that different regions indexed either the salience or the relevance of scene items, but not their interaction, suggesting an evolving representation of salience and relevance across different visual areas.

2018 ◽  
Vol 119 (6) ◽  
pp. 2153-2165 ◽  
Author(s):  
Thomas C. Sprague ◽  
Sirawaj Itthipuripat ◽  
Vy A. Vo ◽  
John T. Serences

Computational models posit that visual attention is guided by activity within spatial maps that index the image-computable salience and the behavioral relevance of objects in the scene. These spatial maps are theorized to be instantiated as activation patterns across a series of retinotopic visual regions in occipital, parietal, and frontal cortex. Whereas previous research has identified sensitivity to either the behavioral relevance or the image-computable salience of different scene elements, the simultaneous influence of these factors on neural “attentional priority maps” in human cortex is not well understood. We tested the hypothesis that visual salience and behavioral relevance independently impact the activation profile across retinotopically organized cortical regions by quantifying attentional priority maps measured in human brains using functional MRI while participants attended one of two differentially salient stimuli. We found that the topography of activation in priority maps, as reflected in the modulation of region-level patterns of population activity, independently indexed the physical salience and behavioral relevance of each scene element. Moreover, salience strongly impacted activation patterns in early visual areas, whereas later visual areas were dominated by relevance. This suggests that prioritizing spatial locations relies on distributed neural codes containing graded representations of salience and relevance across the visual hierarchy.NEW & NOTEWORTHY We tested a theory which supposes that neural systems represent scene elements according to both their salience and their relevance in a series of “priority maps” by measuring functional MRI activation patterns across human brains and reconstructing spatial maps of the visual scene. We found that different regions indexed either the salience or the relevance of scene items, but not their interaction, suggesting an evolving representation of salience and relevance across different visual areas.


2016 ◽  
Vol 23 (2) ◽  
pp. 220-227 ◽  
Author(s):  
Tal Benoliel ◽  
Noa Raz ◽  
Tamir Ben-Hur ◽  
Netta Levin

Background: We have recently suggested that delayed visual evoked potential (VEP) latencies in the fellow eye (FE) of optic neuritis patients reflect a cortical adaptive process, to compensate for the delayed arrival of visual information via the affected eye (AE). Objective: To define the cortical mechanism that underlies this adaptive process. Methods: Cortical activations to moving stimuli and connectivity patterns within the visual network were tested using functional magnetic resonance imaging (MRI) in 11 recovered optic neuritis patients and in 11 matched controls. Results: Reduced cortical activation in early but not in higher visual areas was seen in both eyes, compared to controls. VEP latencies in the AEs inversely correlated with activation in motion-related visual cortices. Inter-eye differences in VEP latencies inversely correlated with cortical activation following FE stimulation, throughout the visual hierarchy. Functional correlation between visual regions was more pronounced in the FE compared with the AE. Conclusion: The different correlation patterns between VEP latencies and cortical activation in the AE and FE support different pathophysiology of VEP prolongation in each eye. Similar cortical activation patterns in both eyes and the fact that stronger links between early and higher visual areas were found following FE stimulation suggest a cortical modulatory process in the FE.


1996 ◽  
Vol 75 (4) ◽  
pp. 1673-1686 ◽  
Author(s):  
H. D. Critchley ◽  
E. T. Rolls

1. The primate orbitofrontal cortex is the site of convergence of information from primary taste and primary olfactory cortical regions. In addition, it receives projections from temporal lobe visual areas concerned with the representation of objects such as foods. Previous work has shown that the responses of gustatory neurons in the secondary taste area within the orbitofrontal cortex are modulated by hunger and satiety, in that they stop responding to the taste of a food on which an animal has been fed to behavioral satiation, yet may continue to respond to the taste of other foods. 2. This study demonstrates a similar modulation of the responses of olfactory and visual orbitofrontal cortex neurons after feeding to satiety. Seven of nine olfactory neurons that were responsive to the odors of foods, such as blackcurrant juice, were found to decrease their responses to the odor of the satiating food in a selective and statistically significant manner. 3. It also was found for eight of nine neurons that had selective responses to the sight of food, that they demonstrated a sensory-specific reduction in their visual responses to foods after satiation. 4. The responses of orbitofrontal cortex neurons selective for foods in more than one modality also were analyzed before and after feeding to satiation. Satiety often affected the responses of these multimodal neurons across all modalities, but a sensory-specific effect was not always demonstrable for both modalities. 5. These findings show that the olfactory and visual representations of food, as well as the taste representation of food, in the primate orbitofrontal cortex are modulated by hunger. Usually a component related to sensory-specific satiety can be demonstrated. The findings link at least part of the processing of olfactory and visual information in this brain region to the control of feeding-related behavior.


2008 ◽  
Vol 99 (3) ◽  
pp. 1380-1393 ◽  
Author(s):  
Joakim Vinberg ◽  
Kalanit Grill-Spector

The lateral occipital complex (LOC) responds preferentially to objects compared with random stimuli or textures independent of the visual cue. However, it is unknown whether the LOC (or other cortical regions) are involved in the processing of edges or global surfaces without shape information. Here, we examined processing of 1) global shape, 2) disconnected edges without a global shape, and 3) global surfaces without edges versus random stimuli across motion and stereo cues. The LOC responded more strongly to global shapes than to edges, surfaces, or random stimuli, for both motion and stereo cues. However, its responses to local edges or global surfaces were not different from random stimuli. This suggests that the LOC processes shapes, not edges or surfaces. LOC also responded more strongly to objects than to holes with the same shape, suggesting sensitivity to border ownership. V7 responded more strongly to edges than to surfaces or random stimuli for both motion and stereo cues, whereas V3a and V4 preferred motion edges. Finally, a region in the caudal intraparietal sulcus (cIPS) responded more strongly to both stereo versus motion and to stereo surfaces versus random stereo (but not to motion surfaces vs. random motion). Thus we found evidence for cue-specific responses to surfaces in the cIPS, both cue-specific and cue-independent responses to edges in intermediate visual areas, and shape-selective responses across multiple cues in the LOC. Overall, these data suggest that integration of visual information across multiple cues is mainly achieved at the level of shape and underscore LOC's role in shape computations.


Sensors ◽  
2021 ◽  
Vol 21 (9) ◽  
pp. 3099
Author(s):  
V. Javier Traver ◽  
Judith Zorío ◽  
Luis A. Leiva

Temporal salience considers how visual attention varies over time. Although visual salience has been widely studied from a spatial perspective, its temporal dimension has been mostly ignored, despite arguably being of utmost importance to understand the temporal evolution of attention on dynamic contents. To address this gap, we proposed Glimpse, a novel measure to compute temporal salience based on the observer-spatio-temporal consistency of raw gaze data. The measure is conceptually simple, training free, and provides a semantically meaningful quantification of visual attention over time. As an extension, we explored scoring algorithms to estimate temporal salience from spatial salience maps predicted with existing computational models. However, these approaches generally fall short when compared with our proposed gaze-based measure. Glimpse could serve as the basis for several downstream tasks such as segmentation or summarization of videos. Glimpse’s software and data are publicly available.


1999 ◽  
Vol 11 (3) ◽  
pp. 300-311 ◽  
Author(s):  
Edmund T. Rolls ◽  
Martin J. Tovée ◽  
Stefano Panzeri

Backward masking can potentially provide evidence of the time needed for visual processing, a fundamental constraint that must be incorporated into computational models of vision. Although backward masking has been extensively used psychophysically, there is little direct evidence for the effects of visual masking on neuronal responses. To investigate the effects of a backward masking paradigm on the responses of neurons in the temporal visual cortex, we have shown that the response of the neurons is interrupted by the mask. Under conditions when humans can just identify the stimulus, with stimulus onset asynchronies (SOA) of 20 msec, neurons in macaques respond to their best stimulus for approximately 30 msec. We now quantify the information that is available from the responses of single neurons under backward masking conditions when two to six faces were shown. We show that the information available is greatly decreased as the mask is brought closer to the stimulus. The decrease is more marked than the decrease in firing rate because it is the selective part of the firing that is especially attenuated by the mask, not the spontaneous firing, and also because the neuronal response is more variable at short SOAs. However, even at the shortest SOA of 20 msec, the information available is on average 0.1 bits. This compares to 0.3 bits with only the 16-msec target stimulus shown and a typical value for such neurons of 0.4 to 0.5 bits with a 500-msec stimulus. The results thus show that considerable information is available from neuronal responses even under backward masking conditions that allow the neurons to have their main response in 30 msec. This provides evidence for how rapid the processing of visual information is in a cortical area and provides a fundamental constraint for understanding how cortical information processing operates.


2010 ◽  
Vol 104 (4) ◽  
pp. 2075-2081 ◽  
Author(s):  
Lars Strother ◽  
Adrian Aldcroft ◽  
Cheryl Lavell ◽  
Tutis Vilis

Functional MRI (fMRI) studies of the human object recognition system commonly identify object-selective cortical regions by comparing blood oxygen level–dependent (BOLD) responses to objects versus those to scrambled objects. Object selectivity distinguishes human lateral occipital cortex (LO) from earlier visual areas. Recent studies suggest that, in addition to being object selective, LO is retinotopically organized; LO represents both object and location information. Although LO responses to objects have been shown to depend on location, it is not known whether responses to scrambled objects vary similarly. This is important because it would suggest that the degree of object selectivity in LO does not vary with retinal stimulus position. We used a conventional functional localizer to identify human visual area LO by comparing BOLD responses to objects versus scrambled objects presented to either the upper (UVF) or lower (LVF) visual field. In agreement with recent findings, we found evidence of position-dependent responses to objects. However, we observed the same degree of position dependence for scrambled objects and thus object selectivity did not differ for UVF and LVF stimuli. We conclude that, in terms of BOLD response, LO discriminates objects from non-objects equally well in either visual field location, despite stronger responses to objects in the LVF.


eLife ◽  
2018 ◽  
Vol 7 ◽  
Author(s):  
Sean R O'Bryan ◽  
Darrell A Worthy ◽  
Evan J Livesey ◽  
Tyler Davis

Extensive evidence suggests that people use base rate information inconsistently in decision making. A classic example is the inverse base rate effect (IBRE), whereby participants classify ambiguous stimuli sharing features of both common and rare categories as members of the rare category. Computational models of the IBRE have either posited that it arises from associative similarity-based mechanisms or dissimilarity-based processes that may depend upon higher-level inference. Here we develop a hybrid model, which posits that similarity- and dissimilarity-based evidence both contribute to the IBRE, and test it using functional magnetic resonance imaging data collected from human subjects completing an IBRE task. Consistent with our model, multivoxel pattern analysis reveals that activation patterns on ambiguous test trials contain information consistent with dissimilarity-based processing. Further, trial-by-trial activation in left rostrolateral prefrontal cortex tracks model-based predictions for dissimilarity-based processing, consistent with theories positing a role for higher-level symbolic processing in the IBRE.


Author(s):  
Jessica A.F. Thompson ◽  
Yoshua Bengio ◽  
Elia Formisano ◽  
Marc Schönwiesner

AbstractThe correspondence between the activity of artificial neurons in convolutional neural networks (CNNs) trained to recognize objects in images and neural activity collected throughout the primate visual system has been well documented. Shallower layers of CNNs are typically more similar to early visual areas and deeper layers tend to be more similar to later visual areas, providing evidence for a shared representational hierarchy. This phenomenon has not been thoroughly studied in the auditory domain. Here, we compared the representations of CNNs trained to recognize speech (triphone recognition) to 7-Tesla fMRI activity collected throughout the human auditory pathway, including subcortical and cortical regions, while participants listened to speech. We found no evidence for a shared representational hierarchy of acoustic speech features. Instead, all auditory regions of interest were most similar to a single layer of the CNNs: the first fully-connected layer. This layer sits at the boundary between the relatively task-general intermediate layers and the highly task-specific final layers. This suggests that alternative architectural designs and/or training objectives may be needed to achieve fine-grained layer-wise correspondence with the human auditory pathway.HighlightsTrained CNNs more similar to auditory fMRI activity than untrainedNo evidence of a shared representational hierarchy for acoustic featuresAll ROIs were most similar to the first fully-connected layerCNN performance on speech recognition task positively associated with fmri similarity


2012 ◽  
Vol 25 (0) ◽  
pp. 58
Author(s):  
Katrina Quinn ◽  
Francia Acosta-Saltos ◽  
Jan W. de Fockert ◽  
Charles Spence ◽  
Andrew J. Bremner

Information about where our hands are arises from different sensory modalities; chiefly proprioception and vision. These inputs differ in variability from situation to situation (or task to task). According to the idea of ‘optimal integration’, the information provided by different sources is combined in proportion to their relative reliabilities, thus maximizing the reliability of the combined estimate. It is uncertain whether optimal multisensory integration of multisensory contributions to limb position requires executive resources. If so, then it should be possible to observe effects of secondary task performance and/or working memory load (WML) on the relative weighting of the senses under conditions of crossmodal sensory conflict. Alternatively, an integrated signal may be affected by upstream influences of WML or a secondary task on the reliabilities of the individual sensory inputs. We examine these possibilities in two experiments which examine effects of WML on reaching tasks in which bisensory visual-proprioceptive (Exp. 1), and unisensory proprioceptive (Exp. 2) cues to hand position are provided. WML increased visual capture under conditions of visual-proprioceptive conflict, regardless of the direction of visual-proprioceptive conflict, and the degree of load imposed. This indicates that task-switching (rather than WML load) leads to an increased reliance on visual information regardless of its task-specific reliability (Exp. 1). This could not be explained due to an increase in the variability of proprioception under secondary working memory task conditions (Exp. 2). We conclude that executive resources are involved in the relative weighting of visual and proprioceptive cues to hand position.


Sign in / Sign up

Export Citation Format

Share Document