robot audition
Recently Published Documents


TOTAL DOCUMENTS

111
(FIVE YEARS 11)

H-INDEX

15
(FIVE YEARS 2)

Sensors ◽  
2021 ◽  
Vol 21 (15) ◽  
pp. 5005
Author(s):  
Caleb Rascon

Beamforming is a type of audio array processing techniques used for interference reduction, sound source localization, and as pre-processing stage for audio event classification and speaker identification. The auditory scene analysis community can benefit from a systemic evaluation and comparison between different beamforming techniques. In this paper, five popular beamforming techniques are evaluated in two different acoustic environments, while varying the number of microphones, the number of interferences, and the direction-of-arrival error, by using the Acoustic Interactions for Robot Audition (AIRA) corpus and a common software framework. Additionally, a highly efficient phase-based frequency masking beamformer is also evaluated, which is shown to outperform all five techniques. Both the evaluation corpus and the beamforming implementations are freely available and provided for experiment repeatability and transparency. Raw results are also provided as a complement to this work to the reader, to facilitate an informed decision of which technique to use. Finally, the insights and tendencies observed from the evaluation results are presented.


Birds ◽  
2021 ◽  
Vol 2 (2) ◽  
pp. 158-172
Author(s):  
Shinji Sumitani ◽  
Reiji Suzuki ◽  
Takaya Arita ◽  
Kazuhiro Nakadai ◽  
Hiroshi G. Okuno

To understand the social interactions among songbirds, extracting the timing, position, and acoustic properties of their vocalizations is essential. We propose a framework for automatic and fine-scale extraction of spatial-spectral-temporal patterns of bird vocalizations in a densely populated environment. For this purpose, we used robot audition techniques to integrate information (i.e., the timing, direction of arrival, and separated sound of localized sources) from multiple microphone arrays (array of arrays) deployed in an environment, which is non-invasive. As a proof of concept of this framework, we examined the ability of the method to extract active vocalizations of multiple Zebra Finches in an outdoor mesh tent as a realistic situation in which they could fly and vocalize freely. We found that localization results of vocalizations reflected the arrangements of landmark spots in the environment such as nests or perches and some vocalizations were localized at non-landmark positions. We also classified their vocalizations as either songs or calls by using a simple method based on the tempo and length of the separated sounds, as an example of the use of the information obtained from the framework. Our proposed approach has great potential to understand their social interactions and the semantics or functions of their vocalizations considering the spatial relationships, although detailed understanding of the interaction would require analysis of more long-term recordings.


Author(s):  
Reiji Suzuki ◽  
Hao Zhao ◽  
Shinji Sumitani ◽  
Shiho Matsubayashi ◽  
Takaya Arita ◽  
...  
Keyword(s):  

Author(s):  
Shinji Sumitani ◽  
Reiji Suzuki ◽  
Shiho Matsubayashi ◽  
Takaya Arita ◽  
Kazuhiro Nakadai ◽  
...  

Sensors ◽  
2019 ◽  
Vol 19 (18) ◽  
pp. 3902 ◽  
Author(s):  
Caleb Rascon ◽  
Oscar Ruiz-Espitia ◽  
Jose Martinez-Carranza

Audio analysis over an Unmanned Aerial Systems (UAS) is of interest it is an essential step for on-board sound source localization and separation. This could be useful for search & rescue operations, as well as for detection of unauthorized drone operations. In this paper, an analysis of the previously introduced Acoustic Interactions for Robot Audition (AIRA)-UAS corpus is presented, which is a set of recordings produced by the ego-noise of a drone performing different aerial maneuvers and by other drones flying nearby. It was found that the recordings have a very low Signal-to-Noise Ratio (SNR), that the noise is dynamic depending of the drone’s movements, and that their noise signatures are highly correlated. Three popular filtering techniques were evaluated in this work in terms of noise reduction and signature extraction, which are: Berouti’s Non-Linear Noise Subtraction, Adaptive Quantile Based Noise Estimation, and Improved Minima Controlled Recursive Averaging. Although there was moderate success in noise reduction, no filter was able to keep intact the signature of the drone flying in parallel. These results are evidence of the challenge in audio processing over drones, implying that this is a field prime for further research.


2019 ◽  
Vol 43 (8) ◽  
pp. 2293-2317 ◽  
Author(s):  
Quan V. Nguyen ◽  
Francis Colas ◽  
Emmanuel Vincent ◽  
François Charpillet

Sign in / Sign up

Export Citation Format

Share Document