acoustic event
Recently Published Documents


TOTAL DOCUMENTS

222
(FIVE YEARS 61)

H-INDEX

20
(FIVE YEARS 4)

2021 ◽  
pp. 104980
Author(s):  
Luca Trani ◽  
Giuliano Andrea Pagani ◽  
João Paulo Pereira Zanetti ◽  
Camille Chapeland ◽  
Läslo Evers

2021 ◽  
Author(s):  
Lijian Gao ◽  
Qirong Mao ◽  
Jingjing Chen ◽  
Ming Dong ◽  
Ratna Chinnam ◽  
...  

Sensors ◽  
2021 ◽  
Vol 21 (19) ◽  
pp. 6622
Author(s):  
Barış Bayram ◽  
Gökhan İnce

Acoustic scene analysis (ASA) relies on the dynamic sensing and understanding of stationary and non-stationary sounds from various events, background noises and human actions with objects. However, the spatio-temporal nature of the sound signals may not be stationary, and novel events may exist that eventually deteriorate the performance of the analysis. In this study, a self-learning-based ASA for acoustic event recognition (AER) is presented to detect and incrementally learn novel acoustic events by tackling catastrophic forgetting. The proposed ASA framework comprises six elements: (1) raw acoustic signal pre-processing, (2) low-level and deep audio feature extraction, (3) acoustic novelty detection (AND), (4) acoustic signal augmentations, (5) incremental class-learning (ICL) (of the audio features of the novel events) and (6) AER. The self-learning on different types of audio features extracted from the acoustic signals of various events occurs without human supervision. For the extraction of deep audio representations, in addition to visual geometry group (VGG) and residual neural network (ResNet), time-delay neural network (TDNN) and TDNN based long short-term memory (TDNN–LSTM) networks are pre-trained using a large-scale audio dataset, Google AudioSet. The performances of ICL with AND using Mel-spectrograms, and deep features with TDNNs, VGG, and ResNet from the Mel-spectrograms are validated on benchmark audio datasets such as ESC-10, ESC-50, UrbanSound8K (US8K), and an audio dataset collected by the authors in a real domestic environment.


2021 ◽  
Vol 11 (18) ◽  
pp. 8581
Author(s):  
Yuzhuo Liu ◽  
Hangting Chen ◽  
Jian Wang ◽  
Pei Wang ◽  
Pengyuan Zhang

In recent years, the involvement of synthetic strongly labeled data, weakly labeled data, and unlabeled data has drawn much research attention in semi-supervised acoustic event detection (SAED). The classic self-training method carries out predictions for unlabeled data and then selects predictions with high probabilities as pseudo-labels for retraining. Such models have shown its effectiveness in SAED. However, probabilities are poorly calibrated confidence estimates, and samples with low probabilities are ignored. Hence, we introduce a confidence-based semi-supervised Acoustic event detection (C-SAED) framework. The C-SAED method learns confidence deliberately and retrains all data distinctly by applying confidence as weights. Additionally, we apply a power pooling function whose coefficient can be trained automatically and use weakly labeled data more efficiently. The experimental results demonstrate that the generated confidence is proportional to the accuracy of the predictions. Our C-SAED framework achieves a relative error rate reduction of 34% in contrast to the baseline model.


2021 ◽  
Author(s):  
Tatsuya Komatsu ◽  
Shinji Watanabe ◽  
Koichi Miyazaki ◽  
Tomoki Hayashi

2021 ◽  
Vol 178 ◽  
pp. 107970
Author(s):  
Yanhua Zhang ◽  
Ke Zhang ◽  
Jingyu Wang ◽  
Yu Su

2021 ◽  
Author(s):  
Alexander Iliev ◽  
Mayank Dewli ◽  
Muhsin Kalkan ◽  
Preeti Prakash Kudva ◽  
Rekha Turkar

2021 ◽  
pp. 115007
Author(s):  
Zixing Zhang ◽  
Ding Liu ◽  
Jing Han ◽  
Kun Qian ◽  
Björn W. Schuller

Sign in / Sign up

Export Citation Format

Share Document