Audio Bank: A high-level acoustic signal representation for audio event recognition

Author(s):  
Tushar Sandhan ◽  
Sukanya Sonowal ◽  
Jin Young Choi
Author(s):  
Najla Bouarada Ghrab ◽  
Rania Rebai Boukhriss ◽  
Emna Fendri ◽  
Mohamed Hammami

Sensors ◽  
2021 ◽  
Vol 21 (19) ◽  
pp. 6622
Author(s):  
Barış Bayram ◽  
Gökhan İnce

Acoustic scene analysis (ASA) relies on the dynamic sensing and understanding of stationary and non-stationary sounds from various events, background noises and human actions with objects. However, the spatio-temporal nature of the sound signals may not be stationary, and novel events may exist that eventually deteriorate the performance of the analysis. In this study, a self-learning-based ASA for acoustic event recognition (AER) is presented to detect and incrementally learn novel acoustic events by tackling catastrophic forgetting. The proposed ASA framework comprises six elements: (1) raw acoustic signal pre-processing, (2) low-level and deep audio feature extraction, (3) acoustic novelty detection (AND), (4) acoustic signal augmentations, (5) incremental class-learning (ICL) (of the audio features of the novel events) and (6) AER. The self-learning on different types of audio features extracted from the acoustic signals of various events occurs without human supervision. For the extraction of deep audio representations, in addition to visual geometry group (VGG) and residual neural network (ResNet), time-delay neural network (TDNN) and TDNN based long short-term memory (TDNN–LSTM) networks are pre-trained using a large-scale audio dataset, Google AudioSet. The performances of ICL with AND using Mel-spectrograms, and deep features with TDNNs, VGG, and ResNet from the Mel-spectrograms are validated on benchmark audio datasets such as ESC-10, ESC-50, UrbanSound8K (US8K), and an audio dataset collected by the authors in a real domestic environment.


2013 ◽  
Vol 333-335 ◽  
pp. 504-509 ◽  
Author(s):  
Fei Hu ◽  
Chang Qing Shen ◽  
Fan Rang Kong

The noises with high level from the different positions of the train lead to the difficulty for extraction of the acoustic signal from the train bearings. Thus the de-noising should be carried out to eliminate the influence of the noises. A new de-noising method has been proposed based on the Doppler Shift. With the crazy climber detection algorithm, all the instantaneous frequencies could be extracted. According to the acoustic theory of Morse, the information of the position of these acoustic sources is revealed. Finally, the variable digital filtering was employed to remove the noises from the other positions of the train. A simulation on the signal containing with ten frequency components demonstrated the validity of this method proposed and the useful components of the signal were enhanced.


Author(s):  
Huy Phan ◽  
Lars Hertel ◽  
Marco Maass ◽  
Radoslaw Mazur ◽  
Alfred Mertins

2014 ◽  
Vol 21 (5) ◽  
pp. 507-524 ◽  
Author(s):  
Tong Lu ◽  
Gongyou Wang ◽  
Feng Su

Sign in / Sign up

Export Citation Format

Share Document