auditory feature
Recently Published Documents


TOTAL DOCUMENTS

64
(FIVE YEARS 6)

H-INDEX

15
(FIVE YEARS 0)

2021 ◽  
Vol 11 (2) ◽  
pp. 129-149
Author(s):  
Jingjing Xu ◽  
Robyn M. Cox

Recent research has established a connection between hearing aid (HA) users’ cognition and speech recognition performance with short and long compression release times (RT). Contradictive findings prevent researchers from using cognition to predict RT prescription. We hypothesized that the linguistic context of speech recognition test materials was one of the factors that accounted for the inconsistency. The present study was designed to examine the relationship between HA users’ cognition and their aided speech recognition performance with short and long RTs using materials with various linguistic contexts. Thirty-four older HA users’ cognitive abilities were quantified using a reading span test. They were fitted with behind-the-ear style HAs with adjustable RT settings. Three speech recognition tests were used: the word-in-noise (WIN) test, the American four alternative auditory feature (AFAAF) test, and the Bamford-Kowal-Bench speech-in-noise (BKB-SIN) test. The results showed that HA users with high cognitive abilities performed better on the AFAAF and the BKB-SIN than those with low cognitive abilities when using short RT. None of the speech recognition tests produced significantly different performance between the two RTs for either cognitive group. These findings did not support our hypothesis. The results suggest that cognition might not be important in prescribing RT.


2021 ◽  
Vol 12 (1) ◽  
Author(s):  
Yuanning Li ◽  
Claire Tang ◽  
Junfeng Lu ◽  
Jinsong Wu ◽  
Edward F. Chang

AbstractLanguages can use a common repertoire of vocal sounds to signify distinct meanings. In tonal languages, such as Mandarin Chinese, pitch contours of syllables distinguish one word from another, whereas in non-tonal languages, such as English, pitch is used to convey intonation. The neural computations underlying language specialization in speech perception are unknown. Here, we use a cross-linguistic approach to address this. Native Mandarin- and English- speaking participants each listened to both Mandarin and English speech, while neural activity was directly recorded from the non-primary auditory cortex. Both groups show language-general coding of speaker-invariant pitch at the single electrode level. At the electrode population level, we find language-specific distribution of cortical tuning parameters in Mandarin speakers only, with enhanced sensitivity to Mandarin tone categories. Our results show that speech perception relies upon a shared cortical auditory feature processing mechanism, which may be tuned to the statistics of a given language.


2021 ◽  
pp. 002221942098800
Author(s):  
Paula Virtala ◽  
Eino Partanen ◽  
Teija Kujala

Rules and regularities of language are typically processed in an implicit and effortless way in the human brain. Individuals with developmental dyslexia have problems in implicit learning of regularities in sequential stimuli but the neural basis of this deficit has not been studied. This study investigated extraction and utilization of a complex auditory rule at neural and perceptual levels in 18 adults with dyslexia and 20 typical readers. Mismatch negativity (MMN) and P3a responses to rule violations in speech stimuli, reflecting change detection and attention switch, respectively, were recorded with electroencephalogram. Both groups reported no or little explicit awareness of the rule, suggesting implicit processing. People with dyslexia showed deficient extraction of the rule evidenced by diminished MMNs estimated to originate particularly from the left perisylvian region. The group difference persisted in attentive condition after the participants were told about the rule, and behavioral detection of the rule violations was poor in people with dyslexia, possibly suggesting difficulties also in utilizing explicit information of the rule. Based on these results, the speech processing difficulties in dyslexia extend beyond phoneme discrimination and basic auditory feature extraction. Challenges in implicit extraction and effortless adoption of complex auditory rules may be central for language learning difficulties in dyslexia.


AI ◽  
2020 ◽  
Vol 1 (4) ◽  
pp. 487-509
Author(s):  
Sudarshan Ramenahalli

The natural environment and our interaction with it are essentially multisensory, where we may deploy visual, tactile and/or auditory senses to perceive, learn and interact with our environment. Our objective in this study is to develop a scene analysis algorithm using multisensory information, specifically vision and audio. We develop a proto-object-based audiovisual saliency map (AVSM) for the analysis of dynamic natural scenes. A specialized audiovisual camera with 360∘ field of view, capable of locating sound direction, is used to collect spatiotemporally aligned audiovisual data. We demonstrate that the performance of a proto-object-based audiovisual saliency map in detecting and localizing salient objects/events is in agreement with human judgment. In addition, the proto-object-based AVSM that we compute as a linear combination of visual and auditory feature conspicuity maps captures a higher number of valid salient events compared to unisensory saliency maps. Such an algorithm can be useful in surveillance, robotic navigation, video compression and related applications.


2019 ◽  
Vol 17 (2) ◽  
pp. 170-177
Author(s):  
Lei Deng ◽  
Yong Gao

In this paper, authors propose an auditory feature extraction algorithm in order to improve the performance of the speaker recognition system in noisy environments. In this auditory feature extraction algorithm, the Gammachirp filter bank is adapted to simulate the auditory model of human cochlea. In addition, the following three techniques are applied: cube-root compression method, Relative Spectral Filtering Technique (RASTA), and Cepstral Mean and Variance Normalization algorithm (CMVN).Subsequently, based on the theory of Gaussian Mixes Model-Universal Background Model (GMM-UBM), the simulated experiment was conducted. The experimental results implied that speaker recognition systems with the new auditory feature has better robustness and recognition performance compared to Mel-Frequency Cepstral Coefficients(MFCC), Relative Spectral-Perceptual Linear Predictive (RASTA-PLP),Cochlear Filter Cepstral Coefficients (CFCC) and gammatone Frequency Cepstral Coefficeints (GFCC)


Neuron ◽  
2018 ◽  
Vol 98 (2) ◽  
pp. 405-416.e4 ◽  
Author(s):  
Xiong Jiang ◽  
Mark A. Chevillet ◽  
Josef P. Rauschecker ◽  
Maximilian Riesenhuber
Keyword(s):  

Sign in / Sign up

Export Citation Format

Share Document