Do "Autistic" Traits Predict Audiovisual Integration in Speech Perception?

2012 ◽  
Author(s):  
Joseph D. W. Stephens ◽  
Julian L. Scrivens ◽  
Amy A. Overman
2005 ◽  
Vol 167 (1) ◽  
pp. 66-75 ◽  
Author(s):  
Maurizio Gentilucci ◽  
Luigi Cattaneo

2018 ◽  
Vol 31 (6) ◽  
pp. 523-536 ◽  
Author(s):  
Ayako Yaguchi ◽  
Souta Hidaka

Autism spectrum disorder (ASD) is a neurodevelopmental disorder characterized by deficits in social communication and interaction, and restricted interests and behavior patterns. These characteristics are considered as a continuous distribution in the general population. People with ASD show atypical temporal processing in multisensory integration. Regarding the flash–beep illusion, which refers to how a single flash can be illusorily perceived as multiple flashes when multiple auditory beeps are concurrently presented, some studies reported that people with ASD have a wider temporal binding window and greater integration than typically developed people; others found the opposite or inconsistent tendencies. Here, we investigated the relationships between the manner of the flash–beep illusion and the various dimensions of ASD traits by estimating the degree of typically developed participants’ ASD traits including five subscales using the Autism-Spectrum Quotient. We found that stronger ASD traits of communication and social skill were associated with a wider and narrower temporal binding window respectively. These results suggest that specific ASD traits are differently involved in the particular temporal binding processes of audiovisual integration.


1995 ◽  
Vol 48 (2) ◽  
pp. 320-333 ◽  
Author(s):  
Eugen Diesch

If a place-of-articulation contrast is created between the auditory and the visual component syllables of videotaped speech, frequently the syllable that listeners report they have heard differs phonetically from the auditory component. These “McGurk effects”, as they have come to be called, show that speech perception may involve some kind of intermodal process. There are two classes of these phenomena: fusions and combinations. Perception of the syllable /da/ when auditory /ba/ and visual /ga/ are presented provides a clear example of the former, and perception of the string /bga/ after presentation of auditory /ga/ and visual /ba/ an unambiguous instance of the latter. Besides perceptual fusions and combinations, hearing visually presented component syllables also shows an influence of vision on audition. It is argued that these “visual” responses arise from basically the same underlying processes that yield fusions and combinations, respectively. In the present study, the visual component of audiovisually incongruous CV-syllables was presented in the left and the right visual hemifield, respectively. Audiovisual fusion responses showed a left hemifield advantage, and audiovisual combination responses a right hemifield advantage. This finding suggests that the process of audiovisual integration differs between audiovisual fusions and combinations and, furthermore, that the two cerebral hemispheres contribute differentially to the two classes of response.


Cognition ◽  
2008 ◽  
Vol 109 (1) ◽  
pp. 157-162 ◽  
Author(s):  
Mary E. Stewart ◽  
Mitsuhiko Ota

2012 ◽  
Vol 25 (0) ◽  
pp. 105 ◽  
Author(s):  
Tobias Søren Andersen

Seeing the talking face can influence the phoneme perceived from the voice. This facilitates speech perception in the natural case where the face and voice are congruent and can cause the McGurk illusion when they are not. The classical example of the McGurk illusion is when acoustic /aba/ is perceived as /ada/ when dubbed onto a face articulating /aga/. In order to fully understand the underlying process of integrating information across the senses we need a computational account with predictive power. The Fuzzy Logical Model of Perception is one computational account of audiovisual integration in speech perception. Here we describe alternative accounts in which integration is based on an early continuous internal representation on which the phonetic classes fall. We show that these alternative accounts can provide just as good a fit when corrected for the number of free parameters. We also show, using cross-validation, that they have greater, but not great, predictive power. Finally, we show that introducing a regularization term can amend the lack of predictive power. With regularization, models based on continuous representations have the highest predictive power.


Sign in / Sign up

Export Citation Format

Share Document