Binaural Source Localization and Spatial Audio Reproduction for Telepresence Applications

2007 ◽  
Vol 16 (5) ◽  
pp. 509-522 ◽  
Author(s):  
Fakheredine Keyrouz ◽  
Klaus Diepold

Telepresence is generally described as the feeling of being immersed in a remote environment, be it virtual or real. A multimodal telepresence environment, equipped with modalities such as vision, audition, and haptic, improves immersion and augments the overall perceptual presence. The present work focuses on acoustic telepresence at both the teleoperator and operator sites. On the teleoperator side, we build a novel binaural sound source localizer using generic Head Related Transfer Functions (HRTFs). This new localizer provides estimates for the direction of a single sound source given in terms of azimuth and elevation angles in free space by using only two microphones. It also uses an algorithm that is efficient compared to the currently known algorithms used in similar localization processes. On the operator side, the paper addresses the problem of spatially interpolating HRTFs for densely sampled high-fidelity 3D sound synthesis. In our telepresence application scenario the synthesized 3D sound is presented to the operator over headphones and shall achieve a high-fidelity acoustic immersion. Using measured HRTF data, we create interpolated HRTFs between the existing functions using a matrix-valued interpolation function. The comparison with existing interpolation methods reveals that our new method offers superior performance and is capable of achieving high-fidelity reconstructions of HRTFs.

2014 ◽  
Vol 41 (16) ◽  
pp. 7106-7113 ◽  
Author(s):  
Lucas Adams Seewald ◽  
Luiz Gonzaga ◽  
Mauricio Roberto Veronez ◽  
Vicente Peruffo Minotto ◽  
Cláudio Rosito Jung

2019 ◽  
Vol 9 (13) ◽  
pp. 2618 ◽  
Author(s):  
Tomasz Rudzki ◽  
Ignacio Gomez-Lanzaco ◽  
Jessica Stubbs ◽  
Jan Skoglund ◽  
Damian T. Murphy ◽  
...  

The increasing popularity of Ambisonics as a spatial audio format for streaming services poses new challenges to existing audio coding techniques. Immersive audio delivered to mobile devices requires an efficient bitrate compression that does not affect the spatial quality of the content. Good localizability of virtual sound sources is one of the key elements that must be preserved. This study was conducted to investigate the localization precision of virtual sound source presentations within Ambisonic scenes encoded with Opus low-bitrate compression at different bitrates and Ambisonic orders (1st, 3rd, and 5th). The test stimuli were reproduced over a 50-channel spherical loudspeaker configuration and binaurally using individually measured and generic Head-Related Transfer Functions (HRTFs). Participants were asked to adjust the position of a virtual acoustic pointer to match the position of virtual sound source within the bitrate-compressed Ambisonic scene. Results show that auditory localization in low-bitrate compressed Ambisonic scenes is not significantly affected by codec parameters. The key factors influencing localization are the rendering method and Ambisonic order truncation. This suggests that efficient perceptual coding might be successfully used for mobile spatial audio delivery.


2017 ◽  
Vol 29 (1) ◽  
pp. 72-82 ◽  
Author(s):  
Takuya Suzuki ◽  
◽  
Hiroaki Otsuka ◽  
Wataru Akahori ◽  
Yoshiaki Bando ◽  
...  

[abstFig src='/00290001/07.jpg' width='300' text='Six impulse response measurement signals' ] Two major functions, sound source localization and sound source separation, provided by robot audition open source software HARK exploit the acoustic transfer functions of a microphone array to improve the performance. The acoustic transfer functions are calculated from the measured acoustic impulse response. In the measurement, special signals such as Time Stretched Pulse (TSP) are used to improve the signal-to-noise ratio of the measurement signals. Recent studies have identified the importance of selecting a measurement signal according to the applications. In this paper, we investigate how six measurement signals – up-TSP, down-TSP, M-Series, Log-SS, NW-SS, and MN-SS – influence the performance of the MUSIC-based sound source localization provided by HARK. Experiments with simulated sounds, up to three simultaneous sound sources, demonstrate no significant difference among the six measurement signals in the MUSIC-based sound source localization.


Sign in / Sign up

Export Citation Format

Share Document