Neural Basis of Semantically Dependent and Independent Cross-Modal Boosts on the Attentional Blink

2020 ◽  
Author(s):  
Song Zhao ◽  
Chengzhi Feng ◽  
Xinyin Huang ◽  
Yijun Wang ◽  
Wenfeng Feng

Abstract The present study recorded event-related potentials (ERPs) in a visual object-recognition task under the attentional blink paradigm to explore the temporal dynamics of the cross-modal boost on attentional blink and whether this auditory benefit would be modulated by semantic congruency between T2 and the simultaneous sound. Behaviorally, the present study showed that not only a semantically congruent but also a semantically incongruent sound improved T2 discrimination during the attentional blink interval, whereas the enhancement was larger for the congruent sound. The ERP results revealed that the behavioral improvements induced by both the semantically congruent and incongruent sounds were closely associated with an early cross-modal interaction on the occipital N195 (192–228 ms). In contrast, the lower T2 accuracy for the incongruent than congruent condition was accompanied by a larger late occurring cento-parietal N440 (424–448 ms). These findings suggest that the cross-modal boost on attentional blink is hierarchical: the task-irrelevant but simultaneous sound, irrespective of its semantic relevance, firstly enables T2 to escape the attentional blink via cross-modally strengthening the early stage of visual object-recognition processing, whereas the semantic conflict of the sound begins to interfere with visual awareness only at a later stage when the representation of visual object is extracted.

Perception ◽  
1996 ◽  
Vol 25 (1_suppl) ◽  
pp. 183-183
Author(s):  
R P Würtz ◽  
C von der Malsburg

Conventional neural networks try to solve the problem of object recognition in a single step by building a stimulus — response system that codes its result as cell activities. We take a different approach assuming that recognition is an active process with temporal dynamics and results in an ordered state. We present a structure of neuronal layers, interconnected by dynamic links (von der Malsburg, 1985 Berichte der Bunsengesellschaft für Physikalische Chemie89 703 – 710) that solves the correspondence problem between two images and thus constitutes an important building block for a model of recognition. Images as well as stored models are represented as Gabor pyramids. This allows the dynamics to proceed from coarse to fine scale and reduces the sequential processing time inherent in the concept. Invariance under background changes is also made possible. On the lowest frequency level, a single blob of activity moves across the image and model layer, respectively. Dynamic links between these layers are initialised to the (highly ambiguous) feature similarities. Links grow or decline according to a combination of feature similarity and correlated activation. This enforces correct neighbourhood relationships in addition to feature similarity. On the higher levels the established correspondences are refined by several blobs in parallel. We present an improved version of the dynamical system proposed by Würtz [1995 Multilayer Dynamic Link Networks for Establishing Image Point Correspondences and Visual Object Recognition (Thun, Frankfurt a.M.: Harri Deutsch)] and show, with examples of human faces, that it evolves from an unordered link distribution to any ordered state where only corresponding point pairs are connected by strong links. Correspondences between sample points are population-coded by a set of neighbouring links.


2011 ◽  
Vol 23 (8) ◽  
pp. 1887-1899 ◽  
Author(s):  
Alex Clarke ◽  
Kirsten I. Taylor ◽  
Lorraine K. Tyler

Research on the spatio-temporal dynamics of visual object recognition suggests a recurrent, interactive model whereby an initial feedforward sweep through the ventral stream to prefrontal cortex is followed by recurrent interactions. However, critical questions remain regarding the factors that mediate the degree of recurrent interactions necessary for meaningful object recognition. The novel prediction we test here is that recurrent interactivity is driven by increasing semantic integration demands as defined by the complexity of semantic information required by the task and driven by the stimuli. To test this prediction, we recorded magnetoencephalography data while participants named living and nonliving objects during two naming tasks. We found that the spatio-temporal dynamics of neural activity were modulated by the level of semantic integration required. Specifically, source reconstructed time courses and phase synchronization measures showed increased recurrent interactions as a function of semantic integration demands. These findings demonstrate that the cortical dynamics of object processing are modulated by the complexity of semantic information required from the visual input.


2015 ◽  
Author(s):  
Radoslaw Cichy ◽  
Dimitrios Pantazis ◽  
Aude Oliva

Every human cognitive function, such as visual object recognition, is realized in a complex spatio-temporal activity pattern in the brain. Current brain imaging techniques in isolation cannot resolve the brain's spatio-temporal dynamics because they provide either high spatial or temporal resolution but not both. To overcome this limitation, we developed a new integration approach that uses representational similarities to combine measurements from different imaging modalities - magnetoencephalography (MEG) and functional MRI (fMRI) - to yield a spatially and temporally integrated characterization of neuronal activation. Applying this approach to two independent MEG-fMRI data sets, we observed that neural activity first emerged in the occipital pole at 50-80ms, before spreading rapidly and progressively in the anterior direction along the ventral and dorsal visual streams. These results provide a novel and comprehensive, spatio-temporally resolved view of the rapid neural dynamics during the first few hundred milliseconds of object vision. They further demonstrate the feasibility of spatially unbiased representational similarity based fusion of MEG and fMRI, promising new insights into how the brain computes complex cognitive functions.


Author(s):  
Ion Juvina ◽  
Priya Ganapathy ◽  
Matt Sherwood ◽  
Mohd Saif Usmani ◽  
Gautam Kunapuli ◽  
...  

2015 ◽  
Vol 585 ◽  
pp. 149-154 ◽  
Author(s):  
Seppo P. Ahlfors ◽  
Stephanie R. Jones ◽  
Jyrki Ahveninen ◽  
Matti S. Hämäläinen ◽  
John W. Belliveau ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document