scholarly journals Archiving the Sonic Ephemeral: Towards a classification of sound installation documentations through spatial audio features

Author(s):  
Nicolas D'Aleman Arango
2010 ◽  
Author(s):  
Matthew Black ◽  
Athanasios Katsamanis ◽  
Chi-Chun Lee ◽  
Adam C. Lammert ◽  
Brian R. Baucom ◽  
...  

Author(s):  
Yakaiah Potharaju ◽  
◽  
Manjunathachari Kamsali ◽  
Chennakesava Kesavari ◽  
◽  
...  

2012 ◽  
Vol 2012 ◽  
pp. 1-6 ◽  
Author(s):  
Fatemeh Safara ◽  
Shyamala Doraisamy ◽  
Azreen Azman ◽  
Azrul Jantan ◽  
Sri Ranga

Heart murmurs are the first signs of cardiac valve disorders. Several studies have been conducted in recent years to automatically differentiate normal heart sounds, from heart sounds with murmurs using various types of audio features. Entropy was successfully used as a feature to distinguish different heart sounds. In this paper, new entropy was introduced to analyze heart sounds and the feasibility of using this entropy in classification of five types of heart sounds and murmurs was shown. The entropy was previously introduced to analyze mammograms. Four common murmurs were considered including aortic regurgitation, mitral regurgitation, aortic stenosis, and mitral stenosis. Wavelet packet transform was employed for heart sound analysis, and the entropy was calculated for deriving feature vectors. Five types of classification were performed to evaluate the discriminatory power of the generated features. The best results were achieved by BayesNet with 96.94% accuracy. The promising results substantiate the effectiveness of the proposed wavelet packet entropy for heart sounds classification.


2021 ◽  
Author(s):  
Lam Pham ◽  
Alexander Schindler ◽  
Mina Schutz ◽  
Jasmin Lampert ◽  
Sven Schlarb ◽  
...  

In this paper, we present deep learning frameworks for audio-visual scene classification (SC) and indicate how individual visual and audio features as well as their combination affect SC performance.Our extensive experiments, which are conducted on DCASE (IEEE AASP Challenge on Detection and Classification of Acoustic Scenes and Events) Task 1B development dataset, achieve the best classification accuracy of 82.2\%, 91.1\%, and 93.9\% with audio input only, visual input only, and both audio-visual input, respectively.The highest classification accuracy of 93.9\%, obtained from an ensemble of audio-based and visual-based frameworks, shows an improvement of 16.5\% compared with DCASE baseline.


Author(s):  
Fika Hastarita Rachman ◽  
Riyanarto Sarno ◽  
Chastine Fatichah

Music has lyrics and audio. That’s components can be a feature for music emotion classification. Lyric features were extracted from text data and audio features were extracted from audio signal data.In the classification of emotions, emotion corpus is required for lyrical feature extraction. Corpus Based Emotion (CBE) succeed to increase the value of F-Measure for emotion classification on text documents. The music document has an unstructured format compared with the article text document. So it requires good preprocessing and conversion process before classification process. We used MIREX Dataset for this research. Psycholinguistic and stylistic features were used as lyrics features. Psycholinguistic feature was a feature that related to the category of emotion. In this research, CBE used to support the extraction process of psycholinguistic feature. Stylistic features related with usage of unique words in the lyrics, e.g. ‘ooh’, ‘ah’, ‘yeah’, etc. Energy, temporal and spectrum features were extracted for audio features.The best test result for music emotion classification was the application of Random Forest methods for lyrics and audio features. The value of F-measure was 56.8%.


Author(s):  
Jose Alvaro Luna-Gonzalez ◽  
Daniel Robles-Camarillo ◽  
Mariko Nakano-Miyatake ◽  
Humberto Lanz-Mendoza ◽  
Hector Perez-Meana

In this paper, a classification of mosquito’s specie is performed using mosquito wingbeats samples obtained by optical sensor. Six world-wide representative species of mosquitos, which are Aedes aegypti, Aedes albopictus, Anopheles arabiensis, Anopheles gambiae and Culex pipiens, Culex quinquefasciatus, are considered for classification. A total of 60,000 samples are divided equally in each specie mentioned above. In total, 25 audio feature extraction algorithms are applied to extract 39 feature values per sample. Further, each audio feature is transformed to a color image, which shows audio features presenting by different pixel values. We used a fully connected neural networks for audio features and a convolutional neural network (CNN) for image dataset generated from audio features. The CNN-based classifier shows 90.75% accuracy, which outperforms the accuracy of 87.18% obtained by the first classifier using directly audio features.


2020 ◽  
Vol 7 (1) ◽  
Author(s):  
Dinarte Vasconcelos ◽  
Nuno Jardim Nunes ◽  
João Gomes

Abstract As vectors of malaria, dengue, zika, and yellow fever, mosquitoes are considered one of the more severe worldwide health hazards. Widespread surveillance of mosquitoes is essential for understanding their complex ecology and behaviour, and also for predicting and formulating effective control strategies against mosquito-borne diseases. One technique involves using bioacoustics to automatically identify different species from their wing-beat sounds during flight. In this dataset, we collect sounds of three species of mosquitoes: Aedes Aegypti, Culex Quinquefasciatus & Pipiens, and Culiseta. These species were collected and reproduced in the laboratory of the Natural History Museum of Funchal, in Portugal, by entomologists trained to recognize and classify mosquitoes. For collecting the samples, we used a microcontroller and a mobile phone. The dataset presents audio samples collected with different sampling rates, where 34 audio features characterize each sound file, making it is possible to observe how mosquito populations vary heterogeneously. This dataset provides the basis for feature extraction and classification of flapping-wing flight sounds that could be used to identify different species.


2020 ◽  
Vol 12 (3) ◽  
pp. 57-67
Author(s):  
Chetna Dabas ◽  
Aditya Agarwal ◽  
Naman Gupta ◽  
Vaibhav Jain ◽  
Siddhant Pathak

Music genre classification has its own popularity index in the present times. Machine learning can play an important role in the music streaming task. This research article proposes a machine learning based model for the classification of music genre. The evaluation of the proposed model is carried out while considering different music genres as in blues, metal, pop, country, classical, disco, jazz and hip-hop. Different audio features utilized in this study include MFCC (Mel Frequency Spectral Coefficients), Delta, Delta-Delta and temporal aspects for processing the data. The implementation of the proposed model has been done in the Python language. The results of the proposed model reveal an accuracy SVM accuracy of 95%. The proposed algorithm has been compared with existing algorithms and the proposed algorithm performs better than the existing ones in terms of accuracy.


2020 ◽  
Vol 10 (17) ◽  
pp. 5956
Author(s):  
Sławomir K. Zieliński ◽  
Hyunkook Lee ◽  
Paweł Antoniuk ◽  
Oskar Dadan

The purpose of this paper is to compare the performance of human listeners against the selected machine learning algorithms in the task of the classification of spatial audio scenes in binaural recordings of music under practical conditions. The three scenes were subject to classification: (1) music ensemble (a group of musical sources) located in the front, (2) music ensemble located at the back, and (3) music ensemble distributed around a listener. In the listening test, undertaken remotely over the Internet, human listeners reached the classification accuracy of 42.5%. For the listeners who passed the post-screening test, the accuracy was greater, approaching 60%. The above classification task was also undertaken automatically using four machine learning algorithms: convolutional neural network, support vector machines, extreme gradient boosting framework, and logistic regression. The machine learning algorithms substantially outperformed human listeners, with the classification accuracy reaching 84%, when tested under the binaural-room-impulse-response (BRIR) matched conditions. However, when the algorithms were tested under the BRIR mismatched scenario, the accuracy obtained by the algorithms was comparable to that exhibited by the listeners who passed the post-screening test, implying that the machine learning algorithms capability to perform in unknown electro-acoustic conditions needs to be further improved.


Sign in / Sign up

Export Citation Format

Share Document