scholarly journals A link between individual differences in multisensory speech perception and eye movements

2015 ◽  
Vol 77 (4) ◽  
pp. 1333-1341 ◽  
Author(s):  
Demet Gurler ◽  
Nathan Doyle ◽  
Edgar Walker ◽  
John Magnotti ◽  
Michael Beauchamp
Cortex ◽  
2016 ◽  
Vol 85 ◽  
pp. 182-193 ◽  
Author(s):  
Rosanna K. Olsen ◽  
Vinoja Sebanayagam ◽  
Yunjo Lee ◽  
Morris Moscovitch ◽  
Cheryl L. Grady ◽  
...  

2021 ◽  
Vol 11 (7) ◽  
pp. 915
Author(s):  
Marianna Stella ◽  
Paul E. Engelhardt

In this study, we examined eye movements and comprehension in sentences containing a relative clause. To date, few studies have focused on syntactic processing in dyslexia and so one goal of the study is to contribute to this gap in the experimental literature. A second goal is to contribute to theoretical psycholinguistic debate concerning the cause and the location of the processing difficulty associated with object-relative clauses. We compared dyslexic readers (n = 50) to a group of non-dyslexic controls (n = 50). We also assessed two key individual differences variables (working memory and verbal intelligence), which have been theorised to impact reading times and comprehension of subject- and object-relative clauses. The results showed that dyslexics and controls had similar comprehension accuracy. However, reading times showed participants with dyslexia spent significantly longer reading the sentences compared to controls (i.e., a main effect of dyslexia). In general, sentence type did not interact with dyslexia status. With respect to individual differences and the theoretical debate, we found that processing difficulty between the subject and object relatives was no longer significant when individual differences in working memory were controlled. Thus, our findings support theories, which assume that working memory demands are responsible for the processing difficulty incurred by (1) individuals with dyslexia and (2) object-relative clauses as compared to subject relative clauses.


2011 ◽  
Vol 55 (6) ◽  
pp. 563-571 ◽  
Author(s):  
M. Elsabbagh ◽  
H. Cohen ◽  
M. Cohen ◽  
S. Rosen ◽  
A. Karmiloff-Smith

2020 ◽  
Vol 24 ◽  
pp. 233121652093054 ◽  
Author(s):  
Tali Rotman ◽  
Limor Lavie ◽  
Karen Banai

Challenging listening situations (e.g., when speech is rapid or noisy) result in substantial individual differences in speech perception. We propose that rapid auditory perceptual learning is one of the factors contributing to those individual differences. To explore this proposal, we assessed rapid perceptual learning of time-compressed speech in young adults with normal hearing and in older adults with age-related hearing loss. We also assessed the contribution of this learning as well as that of hearing and cognition (vocabulary, working memory, and selective attention) to the recognition of natural-fast speech (NFS; both groups) and speech in noise (younger adults). In young adults, rapid learning and vocabulary were significant predictors of NFS and speech in noise recognition. In older adults, hearing thresholds, vocabulary, and rapid learning were significant predictors of NFS recognition. In both groups, models that included learning fitted the speech data better than models that did not include learning. Therefore, under adverse conditions, rapid learning may be one of the skills listeners could employ to support speech recognition.


2018 ◽  
Vol 120 (4) ◽  
pp. 1602-1615 ◽  
Author(s):  
Anouk J. de Brouwer ◽  
Mohammed Albaghdadi ◽  
J. Randall Flanagan ◽  
Jason P. Gallivan

Successful motor performance relies on our ability to adapt to changes in the environment by learning novel mappings between motor commands and sensory outcomes. Such adaptation is thought to involve two distinct mechanisms: an implicit, error-based component linked to slow learning and an explicit, strategic component linked to fast learning and savings (i.e., faster relearning). Because behavior, at any given moment, is the resultant combination of these two processes, it has remained a challenge to parcellate their relative contributions to performance. The explicit component to visuomotor rotation (VMR) learning has recently been measured by having participants verbally report their aiming strategy used to counteract the rotation. However, this procedure has been shown to magnify the explicit component. Here we tested whether task-specific eye movements, a natural component of reach planning, but poorly studied in motor learning tasks, can provide a direct readout of the state of the explicit component during VMR learning. We show, by placing targets on a visible ring and including a delay between target presentation and reach onset, that individual differences in gaze patterns during sensorimotor learning are linked to participants’ rates of learning and their expression of savings. Specifically, we find that participants who, during reach planning, naturally fixate an aimpoint rotated away from the target location, show faster initial adaptation and readaptation 24 h later. Our results demonstrate that gaze behavior cannot only uniquely identify individuals who implement cognitive strategies during learning but also how their implementation is linked to differences in learning. NEW & NOTEWORTHY Although it is increasingly well appreciated that sensorimotor learning is driven by two separate components, an error-based process and a strategic process, it has remained a challenge to identify their relative contributions to performance. Here we demonstrate that task-specific eye movements provide a direct read-out of explicit strategies during sensorimotor learning in the presence of visual landmarks. We further show that individual differences in gaze behavior are linked to learning rate and savings.


2019 ◽  
Vol 40 (2) ◽  
pp. 255-287 ◽  
Author(s):  
Anselm R. Strohmaier ◽  
Matthias C. Lehner ◽  
Jana T. Beitlich ◽  
Kristina M. Reiss

Sign in / Sign up

Export Citation Format

Share Document