scholarly journals Neural tracking to go: auditory attention decoding and saliency detection with mobile EEG

Author(s):  
Lisa Straetmans ◽  
B. Holtze ◽  
Stefan Debener ◽  
Manuela Jaeger ◽  
Bojana Mirkovic

Abstract Objective. Neuro-steered assistive technologies have been suggested to offer a major advancement in future devices like neuro-steered hearing aids. Auditory attention decoding methods would in that case allow for identification of an attended speaker within complex auditory environments, exclusively from neural data. Decoding the attended speaker using neural information has so far only been done in controlled laboratory settings. Yet, it is known that ever-present factors like distraction and movement are reflected in the neural signal parameters related to attention. Approach. Thus, in the current study we applied a two-competing speaker paradigm to investigate performance of a commonly applied EEG-based auditory attention decoding (AAD) model outside of the laboratory during leisure walking and distraction. Unique environmental sounds were added to the auditory scene and served as distractor events. Main results. The current study shows, for the first time, that the attended speaker can be accurately decoded during natural movement. At a temporal resolution of as short as 5-seconds and without artifact attenuation, decoding was found to be significantly above chance level. Further, as hypothesized, we found a decrease in attention to the to-be-attended and the to-be-ignored speech stream after the occurrence of a salient event. Additionally, we demonstrate that it is possible to predict neural correlates of distraction with a computational model of auditory saliency based on acoustic features. Conclusion. Taken together, our study shows that auditory attention tracking outside of the laboratory in ecologically valid conditions is feasible and a step towards the development of future neural-steered hearing aids.

2021 ◽  
Vol 12 ◽  
Author(s):  
Ivine Kuruvila ◽  
Jan Muncke ◽  
Eghart Fischer ◽  
Ulrich Hoppe

Human brain performs remarkably well in segregating a particular speaker from interfering ones in a multispeaker scenario. We can quantitatively evaluate the segregation capability by modeling a relationship between the speech signals present in an auditory scene, and the listener's cortical signals measured using electroencephalography (EEG). This has opened up avenues to integrate neuro-feedback into hearing aids where the device can infer user's attention and enhance the attended speaker. Commonly used algorithms to infer the auditory attention are based on linear systems theory where cues such as speech envelopes are mapped on to the EEG signals. Here, we present a joint convolutional neural network (CNN)—long short-term memory (LSTM) model to infer the auditory attention. Our joint CNN-LSTM model takes the EEG signals and the spectrogram of the multiple speakers as inputs and classifies the attention to one of the speakers. We evaluated the reliability of our network using three different datasets comprising of 61 subjects, where each subject undertook a dual-speaker experiment. The three datasets analyzed corresponded to speech stimuli presented in three different languages namely German, Danish, and Dutch. Using the proposed joint CNN-LSTM model, we obtained a median decoding accuracy of 77.2% at a trial duration of 3 s. Furthermore, we evaluated the amount of sparsity that the model can tolerate by means of magnitude pruning and found a tolerance of up to 50% sparsity without substantial loss of decoding accuracy.


2019 ◽  
Author(s):  
Daniel J. Strauss ◽  
Farah I. Corona-Strauss ◽  
Andreas Schroeer ◽  
Philipp Flotho ◽  
Ronny Hannemann ◽  
...  

AbstractIt is commonly assumed that, unlike dogs and cats, we humans do not make ear movements when focusing our attention reflexively toward novel sounds or voluntarily toward those that are goal–relevant. In fact, it has been suggested that we do have a pinna–orienting system. Although this system became vestigial about 25 million years ago, it still exists as a “neural fossil” within the brain. Consistent with this hypothesis, we demonstrate for the first time that the direction of auditory attention is reflected in the sustained electrical activity of muscles within the vestigial auriculomotor system.Surface electromyograms (EMGs) were taken from muscles that either move the pinna or alter its shape. To assess reflexive, stimulus-driven attention we presented novel sounds from speakers at four different lateral locations while the participants silently read a boring text in front of them. To test voluntary, goal-directed attention we instructed participants to listen to a short story coming from one of these speakers, while ignoring a competing story from the corresponding speaker on the opposite side.In both experiments, EMG recordings showed larger activity at the ear on the side of the attended stimulus, but with slightly different patterns. Upward movement (perking) differed according to the lateral focus of attention only during voluntary orienting; rearward folding of the pinna’s upper-lateral edge exhibited such differences only during reflexive orienting. The existence of a pinna-orienting system in humans, one that is experimentally accessible, offers opportunities for basic as well as applied science. It could lead to a better understanding of the evolution of auditory attention and support the near real–time decoding of auditory attention in technical applications, for example, for attentionally controlled hearing aids that preferentially amplify sounds the user is attempting to listen to.


Author(s):  
M. N. Favorskaya ◽  
L. C. Jain

Introduction:Saliency detection is a fundamental task of computer vision. Its ultimate aim is to localize the objects of interest that grab human visual attention with respect to the rest of the image. A great variety of saliency models based on different approaches was developed since 1990s. In recent years, the saliency detection has become one of actively studied topic in the theory of Convolutional Neural Network (CNN). Many original decisions using CNNs were proposed for salient object detection and, even, event detection.Purpose:A detailed survey of saliency detection methods in deep learning era allows to understand the current possibilities of CNN approach for visual analysis conducted by the human eyes’ tracking and digital image processing.Results:A survey reflects the recent advances in saliency detection using CNNs. Different models available in literature, such as static and dynamic 2D CNNs for salient object detection and 3D CNNs for salient event detection are discussed in the chronological order. It is worth noting that automatic salient event detection in durable videos became possible using the recently appeared 3D CNN combining with 2D CNN for salient audio detection. Also in this article, we have presented a short description of public image and video datasets with annotated salient objects or events, as well as the often used metrics for the results’ evaluation.Practical relevance:This survey is considered as a contribution in the study of rapidly developed deep learning methods with respect to the saliency detection in the images and videos.


2021 ◽  
Vol 21 (1) ◽  
Author(s):  
Domenico Cuda ◽  
Sara Ghiselli ◽  
Alessandra Murri

Abstract Background Prevalence of hearing loss increases with age. Its estimated prevalence is 40–50 % in people over 75 years of age. Recent studies agree that declinein hearing threshold contribute to deterioration in sociality, sensitivity, cognition, and quality of life for elderly subjects. The aim of the study presented in this paper is to verify whether or not rehabilitation using first time applied Hearing Aids (HA) in a cohort of old people with hearing impairment improves both speech perception in a noisy environment over time and the overall health-related quality of life. Methods The monocentric, prospective, repeated measurements, single-subject, clinical observational study is to recruit 100 older adults, first-time HA recipients (≥ 65 years).The evaluation protocol is designed to analyze changes in specific measurement tools a year after the first HA usage in comparison with the evaluation before HA fitting. Evaluations will consist of multiparametric details collected through self-report questionnaires completed by the recipients and a series of commonly used audiometric measures and geriatric assessment tools. The primary indicator of changes in speech perception in noise to be used is the Italian version of Oldenburg Satz (OLSA) test whereas the indicator of changes in overall quality of life will be the Assessment of Quality of Life (AQoL) and Hearing Handicap Inventory for the Elderly (HHIE) questionnaires. The Montreal Cognitive Assessment (MoCA) will help in screening the cognitive state of the subjects. Discussion The protocol is designed to make use of measurement tools that have already been applied to the hearing-impaired population in order to compare the effects of HA rehabilitation in the older adults immediately before first HA usage (Pre) and after 1 year of experience (Post). This broad approach will lead to a greater understanding of how useful hearing influences the quality of life in older individuals, and therefore improves potentials for healthy aging. The data is to be analyzed by using an intrasubject endpoint comparison. Outcomes will be described and analyzed in detail. Trial registration This research was retrospectively registered underno. NCT04333043at ClinicalTrials.gov (http://www.clinicaltrials.gov/) on the 26 March 2020. This research has been registered with the Ethics Committee of the Area Vasta Emilia Nord under number 104, date of approval 17/07/2017.


Author(s):  
Angela Ryall ◽  
Lorienne M. Jenstad ◽  
John Pumford ◽  
Tami Howe ◽  
Garnet Grosjean

Abstract Background When dispensing hearing aids, audiologists must follow validated fitting and verification procedures to ensure that the hearing aids are properly fitted to the client's hearing. Real ear measurements (REMs) are best practice for verifying hearing aids. Prior literature regarding REMs has mainly focused on the clinicians' perspective. Purpose This study investigated informational counseling throughout REMs by gathering perspectives of first-time hearing aid users regarding the content and format of counseling. Research Design The study used an interpretive description approach with focus groups. Study Sample There were 16 adult participants (4 males, 12 females) who were first-time hearing aid users and who all had memory of REMs occurring during their own hearing aid verification. Intervention We investigated the addition of informational counseling during REM verification. Data Collection and Analysis Four focus groups were conducted to elicit feedback on a demonstration of informational counseling during REM hearing aid verification. The data from the focus groups were transcribed verbatim and analyzed using qualitative content analysis. Results Analysis revealed positive aspects, negative aspects, and suggested changes in relation to the verbal and visual information presented during the REM verification demonstration. These data fell into two broad categories: the interaction and transaction of informational counseling. Conclusion Most clients were interested in learning more about REMs if the information was accessible. Results provide recommendations for clinical audiologists and REM system manufacturers to make the information presented during informational counseling more client-friendly and individualized for client-centered care. To continue exploring this new inquiry, further experimental research is required to determine if there is any added value of incorporating informational counseling during REMs.


Author(s):  
Gemma Almond

Abstract This study explores the representation and use of Victorian visual aids, specifically focusing on how the design of spectacle and eyeglass frames shaped ideas of the ‘normal’ and ‘abnormal’ body. It contributes to our understanding of assistive technologies in the Victorian period by showcasing the usefulness of material evidence for exploring how an object was produced and perceived. By placing visual aids in their medical and cultural context for the first time, it will show how the study of spectacle and eyeglass frames develops our understanding of Victorian society more broadly. Contemporaries drew upon industrialization, increasing education, and the proliferation of print to explain a rise in refractive vision ‘errors’. Through exploring the design of three spectacle frames from the London Science Museum’s collections, this study will show how the representations and manufacture of visual aids transformed in response to these wider changes. The material evidence, as well as contemporary newspapers, periodicals, and medical texts, reveal that visual aids evolved from an unusual to a more mainstream device. It argues that visual aids are a unique assistive technology, one that is able to inform our understanding of how Victorians measured the body and constructed ideas of ‘normalcy’ and ‘abnormalcy’.


2013 ◽  
Vol 22 (2) ◽  
pp. 329-331 ◽  
Author(s):  
Louise Hickson ◽  
Ariane Laplante-Lévesque ◽  
Lena Wong

Purpose The authors address 3 questions: (a) What is evidence-based practice (EBP), and why is it important for adults with hearing impairment? (b) What is the evidence about intervention options for adults who fail a hearing screening and are identified with hearing impairment? (c) What intervention options do adults choose when identified with hearing impairment for the first time? Method The 5 steps of the EBP process are discussed in relation to a clinical question about whether hearing aids and communication programs reduce activity limitations and participation restrictions compared with no treatment for adults who fail a hearing screening and are identified with hearing impairment. Results Systematic reviews of the evidence indicate that both hearing aids and communication programs reduce activity limitations and participation restrictions for this population and are therefore appropriate options. A study is then described in which these options were presented to 153 clients identified with hearing impairment for the first time: 43% chose hearing aids, 18% chose communication programs, and the remaining 39% chose not to take any action. Conclusion EBP supports the offer of intervention options to adults who fail a hearing screening and are identified with hearing impairment.


Author(s):  
Gabrielle H. Saunders ◽  
Leslie Grush ◽  
Jay Vachhani ◽  
Katharina V. Echt ◽  
Susan Griest ◽  
...  

Abstract Background Age-related changes (both normal and pathological), and health literacy are relevant to audiological practice. Changes associated with the musculoskeletal, vascular, and nervous systems drive manual, visual, and cognitive function. These in turn affect the capabilities required for effective hearing aid (HA) skill acquisition, use, and management. Meanwhile, health literacy influences the ability to gain access to, understand, and use information, which is important for promoting and maintaining HA use and management. Understanding the interindividual variability of these variables can help audiologists characterize those individuals who might have suboptimal HA outcomes. This knowledge can then inform better clinical practices and guide implementation of processes to improve care quality and outcomes. Purpose The aim of the study is to assess the variation in manual, visual, and cognitive function, and health literacy, among community-dwelling older individuals, and to determine whether and which of these variables are associated with reported HA outcome and/or the knowledge and skill to manage HAs. Research Design Data presented here were collected as part of an efficacy trial of four variants of HA orientation. The data were collected at baseline (prior to HA fitting) and after 4 to 8 weeks of HA use. Study Sample The study sample consists of 265 U.S. Veterans aged 51 to 87 years with no previous HA experience who were scheduled to receive their first pair of HAs from the Veterans Administration. Data Collection and Analysis We assessed baseline measures of hand function, vision, cognition, and health literacy just prior to participants receiving their first pair of HAs. HA management skills and knowledge, and HA outcome were measured after 4 to 8 weeks of HA use using the Hearing Aid Skills and Knowledge (HASK) and International Outcomes Inventory for Hearing Aids (IOI-HA), respectively. Data collected here was compared with published norms to assess variation in baseline measures. Associations between baseline performance and outcomes data were examined using t-tests comparing participants who performed at or above age-based norms with those who performed below age-based norms. Results Participants' performance on the baseline measures was highly variable, with the proportions of individuals performing below norms varying by test measure. When combining data across the nine baseline measures, approximately 10% of participants performed below published norms on five or more measures, and 85% performed below norms on at least one measure. Poor manual dexterity, ability to learn a new task, and ability to draw inferences from spoken information negatively impacted HA management and outcome. Conclusion There was a considerable heterogeneity among a community-dwelling sample of first time HA users in terms of sensory, cognitive, and motor function. Clinicians should consider modifying their clinical practice to account for such heterogeneity and best support their patients in adapting to new HAs.


2014 ◽  
Vol 10 (1) ◽  
pp. 20130926 ◽  
Author(s):  
Tamás Faragó ◽  
Attila Andics ◽  
Viktor Devecseri ◽  
Anna Kis ◽  
Márta Gácsi ◽  
...  

Humans excel at assessing conspecific emotional valence and intensity, based solely on non-verbal vocal bursts that are also common in other mammals. It is not known, however, whether human listeners rely on similar acoustic cues to assess emotional content in conspecific and heterospecific vocalizations, and which acoustical parameters affect their performance. Here, for the first time, we directly compared the emotional valence and intensity perception of dog and human non-verbal vocalizations. We revealed similar relationships between acoustic features and emotional valence and intensity ratings of human and dog vocalizations: those with shorter call lengths were rated as more positive, whereas those with a higher pitch were rated as more intense. Our findings demonstrate that humans rate conspecific emotional vocalizations along basic acoustic rules, and that they apply similar rules when processing dog vocal expressions. This suggests that humans may utilize similar mental mechanisms for recognizing human and heterospecific vocal emotions.


Sign in / Sign up

Export Citation Format

Share Document