Model-based Head Orientation Estimation for Smart Devices

Author(s):  
Qiang Yang ◽  
Yuanqing Zheng

Voice interaction is friendly and convenient for users. Smart devices such as Amazon Echo allow users to interact with them by voice commands and become increasingly popular in our daily life. In recent years, research works focus on using the microphone array built in smart devices to localize the user's position, which adds additional context information to voice commands. In contrast, few works explore the user's head orientation, which also contains useful context information. For example, when a user says, "turn on the light", the head orientation could infer which light the user is referring to. Existing model-based works require a large number of microphone arrays to form an array network, while machine learning-based approaches need laborious data collection and training workload. The high deployment/usage cost of these methods is unfriendly to users. In this paper, we propose HOE, a model-based system that enables Head Orientation Estimation for smart devices with only two microphone arrays, which requires a lower training overhead than previous approaches. HOE first estimates the user's head orientation candidates by measuring the voice energy radiation pattern. Then, the voice frequency radiation pattern is leveraged to obtain the final result. Real-world experiments are conducted, and the results show that HOE can achieve a median estimation error of 23 degrees. To the best of our knowledge, HOE is the first model-based attempt to estimate the head orientation by only two microphone arrays without the arduous data training overhead.

Author(s):  
Rebecca C. Felsheim ◽  
Andreas Brendel ◽  
Patrick A. Naylor ◽  
Walter Kellermann

2019 ◽  
Vol 283 ◽  
pp. 04001
Author(s):  
Boquan Yang ◽  
Shengguo Shi ◽  
Desen Yang

Recently, spherical microphone arrays (SMA) have become increasingly significant for source localization and identification in three dimension due to its spherical symmetry. However, conventional Spherical Harmonic Beamforming (SHB) based on SMA has limitations, such as poor resolution and high side-lobe levels in image maps. To overcome these limitations, this paper employs the iterative generalized inverse beamforming algorithm with a virtual extrapolated open spherical microphone array. The sidelobes can be suppressed and the main-lobe can be narrowed by introducing the two iteration processes into the generalized inverse beamforming (GIB) algorithm. The instability caused by uncertainties in actual measurements, such as measurement noise and configuration problems in the process of GIB, can be minimized by iteratively redefining the form of regularization matrix and the corresponding GIB localization results. In addition, the poor performance of microphone arrays in the low-frequency range due to the array aperture can be improved by using a virtual extrapolated open spherical array (EA), which has a larger array aperture. The virtual array is obtained by a kind of data preprocessing method through the regularization matrix algorithm. Both results from simulations and experiments show the feasibility and accuracy of the method.


2015 ◽  
Vol 23 (6) ◽  
pp. 8061 ◽  
Author(s):  
J. Mateo ◽  
M. A. Losada ◽  
A. López

2017 ◽  
Vol 29 (1) ◽  
pp. 83-93
Author(s):  
Kouhei Sekiguchi ◽  
◽  
Yoshiaki Bando ◽  
Katsutoshi Itoyama ◽  
Kazuyoshi Yoshii

[abstFig src='/00290001/08.jpg' width='300' text='Optimizing robot positions for source separation' ] The active audition method presented here improves source separation performance by moving multiple mobile robots to optimal positions. One advantage of using multiple mobile robots that each has a microphone array is that each robot can work independently or as part of a big reconfigurable array. To determine optimal layout of the robots, we must be able to predict source separation performance from source position information because actual source signals are unknown and actual separation performance cannot be calculated. Our method thus simulates delay-and-sum beamforming from a possible layout to calculate gain theoretically, i.e., the expected ratio of a target sound source to other sound sources in the corresponding separated signal. Robots are moved into the layout with the highest average gain over target sources. Experimental results showed that our method improved the harmonic mean of signal-to-distortion ratios (SDRs) by 5.5 dB in simulation and by 3.5 dB in a real environment.


Sign in / Sign up

Export Citation Format

Share Document