On classification of environmental acoustic data using crowds

Author(s):  
Shan Zhang ◽  
Aditya Vempaty ◽  
Susan E. Parks ◽  
Pramod K. Varshney
Keyword(s):  
Author(s):  
Dongshik Kang ◽  
◽  
Sigeru Omatu ◽  
Michifumi Yoshioka

An advanced neuro-classification of new and used bills using the spectral patterns is proposed. An acoustic spectral pattern is obtained from the output of the two-stage adaptive digital filters (ADFs) for time-series acoustic data. The acoustic spectral patterns are fed to a competitive neural network, and classified into some categories which show worn-out degrees of the bill. The proposed method is based on extension of an ADF, an individual adaptation (IA) algorithm, and a learning vector quantization (LVQ) algorithm. The experimental results show that the proposed method is useful to classify new and used bills.


2012 ◽  
Vol 69 (8) ◽  
pp. 1329-1339 ◽  
Author(s):  
Ronan Fablet ◽  
Paul Gay ◽  
Salvador Peraltilla ◽  
Cecilia Peña ◽  
Ramiro Castillo ◽  
...  

Whereas fisheries acoustics data processing mainly focused on the detection, characterization, and recognition of individual fish schools, here we addressed the characterization and discrimination of fish school clusters. The proposed scheme relied on the application of the Bags-of-Features (BoF) approach to acoustic echograms. This approach is widely exploited for pattern recognition issues and naturally applies here, considering fish schools as the relevant elementary objects. It relies on the extraction and categorization of fish schools in fisheries acoustic data. Echogram descriptors were computed per unit echogram length as the numbers of schools in different school categories. We applied this approach to the discrimination of juvenile and adult anchovy ( Engraulis ringens ) off Peru. Whereas the discrimination of individual schools is low (below 70%), the proposed BoF scheme achieved between 89% and 92% of correct classification of juvenile and adult echograms for different survey data sets and significantly outperformed classical school-based echogram characteristics (about 10% of improvement of the correct classification rate). We further illustrate the potential of the proposed scheme for the estimation of the spatial distribution of juvenile and adult anchovy populations.


2011 ◽  
Vol 57 (202) ◽  
pp. 267-276 ◽  
Author(s):  
Alec Van Herwijnen ◽  
Jürg Schweizer

AbstractIn snow, acoustic emissions originate from the breaking of bonds between snow crystals and the formation of cracks. Previous research has shown that acoustic signals emanate from a natural snowpack. The relation between these signals and the stability of the snowpack has thus far remained elusive. Studies on other hazardous gravitational processes suggest that damage accumulation precedes major failure. If increased cracking activity could be detected in snow this might be used for avalanche prediction. We report on the development of a seismic sensor array to continuously monitor acoustic emissions in an avalanche start zone. During three winters, over 1400 sensor days of continuous acoustic data were collected. With the aid of automatic cameras and a microphone the main types of background noise were identified. Seismic signals generated by avalanches were also identified. Spectrograms from seismic signals generated by avalanches exhibit a unique triangular shape unlike any source of background noise, suggesting that automatic detection and classification of events is possible. Furthermore, discriminating between loose-snow and snow-slab avalanches is possible. Thus far we have not identified precursor events for natural dry-snow slab avalanche release. Detailed investigation of one dry-snow slab avalanche showed that signals observed prior to the release originated from background noise or small loose-snow avalanches.


2016 ◽  
Vol 73 (8) ◽  
pp. 1998-2008 ◽  
Author(s):  
Niall G. Fallon ◽  
Sophie Fielding ◽  
Paul G. Fernandes

Abstract Target identification remains a challenge for acoustic surveys of marine fauna. Antarctic krill, Euphausia superba, are typically identified through a combination of expert scrutiny of echograms and analysis of differences in mean volume backscattering strengths (SV; dB re 1 m−1) measured at two or more echosounder frequencies. For commonly used frequencies, however, the differences for krill are similar to those for many co-occurring fish species that do not possess swimbladders. At South Georgia, South Atlantic, one species in particular, mackerel icefish, Champsocephalus gunnari, forms pelagic aggregations, which can be difficult to distinguish acoustically from large krill layers. Mackerel icefish are currently surveyed using bottom-trawls, but the resultant estimates of abundance may be biased because of the species' semi-pelagic distribution. An acoustic estimate of the pelagic component of the population could indicate the magnitude of this bias, but first a reliable target identification method is required. To address this, random forests (RFs) were generated using acoustic and net sample data collected during surveys. The final RF classified as krill, icefish, and mixed aggregations of weak scattering fish species with an overall estimated accuracy of 95%. Minimum SV, mean aggregation depth (m), mean distance from the seabed (m), and geographic positional data were most important to the accuracy of the RF. Time-of-day and the difference between SV at 120 kHz (SV 120) and that at 38 kHz (SV 38) were also important. The RF classification resulted in significantly higher estimates of backscatter apportioned to krill when compared with widely applied identification methods based on fixed and variable ranges of SV 120–SV 38. These results suggest that krill density is underestimated when those SV-differencing methods are used for target identification. RFs are an objective means for target identification and could enhance the utility of incidentally collected acoustic data.


2021 ◽  
Vol 9 (1) ◽  
Author(s):  
Andréa Thiebault ◽  
Chloé Huetz ◽  
Pierre Pistorius ◽  
Thierry Aubin ◽  
Isabelle Charrier

Abstract Background Studies on animal behaviour often involve the quantification of the occurrence and duration of various activities. When direct observations are challenging (e.g., at night, in a burrow, at sea), animal-borne devices can be used to remotely record the movement and behaviour of an animal (e.g., changing body posture and movement, geographical position) and/or its immediate surrounding environment (e.g., wet or dry, pressure, temperature, light). Changes in these recorded variables are related to different activities undertaken by the animal. Here we explored the use of animal-borne acoustic recorders to automatically infer activities in seabirds. Results We deployed acoustic recorders on Cape gannets and analysed sound data from 10 foraging trips. The different activities (flying, floating on water and diving) were associated with clearly distinguishable acoustic features. We developed a method to automatically identify the activities of equipped individuals, exclusively from animal-borne acoustic data. A random subset of four foraging trips was manually labelled and used to train a classification algorithm (k-nearest neighbour model). The algorithm correctly classified activities with a global accuracy of 98.46%. The model was then used to automatically assess the activity budgets on the remaining non-labelled data, as an illustrative example. In addition, we conducted a systematic review of studies that have previously used data from animal-borne devices to automatically classify animal behaviour (n = 61 classifications from 54 articles). The majority of studies (82%) used accelerometers (alone or in combination with other sensors, such as gyroscopes or magnetometers) for classifying activities, and to a lesser extent GPS, acoustic recorders or pressure sensors, all potentially providing a good accuracy of classification (> 90%). Conclusion This article demonstrates that acoustic data alone can be used to reconstruct activity budgets with very good accuracy. In addition to the animal’s activity, acoustic devices record the environment of equipped animals (biophony, geophony, anthropophony) that can be essential to contextualise the behaviour of animals. They hence provide a valuable alternative to the set of tools available to assess animals’ behaviours and activities in the wild.


Sensor Review ◽  
2021 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Dhanalakshmi M. ◽  
Nagarajan T. ◽  
Vijayalakshmi P.

Purpose Dysarthria is a neuromotor speech disorder caused by neuromuscular disturbances that affect one or more articulators resulting in unintelligible speech. Though inter-phoneme articulatory variations are well captured by formant frequency-based acoustic features, these variations are expected to be much higher for dysarthric speakers than normal. These substantial variations can be well captured by placing sensors in appropriate articulatory position. This study focuses to determine a set of articulatory sensors and parameters in order to assess articulatory dysfunctions in dysarthric speech. Design/methodology/approach The current work aims to determine significant sensors and parameters associated using motion path and correlation analyzes on the TORGO database of dysarthric speech. Among eight informative sensor channels and six parameters per channel in positional data, the sensors such as tongue middle, back and tip, lower and upper lips and parameters (y, z, φ) are found to contribute significantly toward capturing the articulatory information. Acoustic and positional data analyzes are performed to validate these identified significant sensors. Furthermore, a convolutional neural network-based classifier is developed for both phone-and word-level classification of dysarthric speech using acoustic and positional data. Findings The average phone error rate is observed to be lower, up to 15.54% for positional data when compared with acoustic-only data. Further, word-level classification using a combination of both acoustic and positional information is performed to study that the positional data acquired using significant sensors will boost the performance of classification even for severe dysarthric speakers. Originality/value The proposed work shows that the significant sensors and parameters can be used to assess dysfunctions in dysarthric speech effectively. The articulatory sensor data helps in better assessment than the acoustic data even for severe dysarthric speakers.


Sign in / Sign up

Export Citation Format

Share Document