scholarly journals Age group classification and gender recognition from speech with temporal convolutional neural networks

Author(s):  
Héctor A. Sánchez-Hevia ◽  
Roberto Gil-Pita ◽  
Manuel Utrilla-Manso ◽  
Manuel Rosa-Zurera

AbstractThis paper analyses the performance of different types of Deep Neural Networks to jointly estimate age and identify gender from speech, to be applied in Interactive Voice Response systems available in call centres. Deep Neural Networks are used, because they have recently demonstrated discriminative and representation capabilities in a wide range of applications, including speech processing problems based on feature extraction and selection. Networks with different sizes are analysed to obtain information on how performance depends on the network architecture and the number of free parameters. The speech corpus used for the experiments is Mozilla’s Common Voice dataset, an open and crowdsourced speech corpus. The results are really good for gender classification, independently of the type of neural network, but improve with the network size. Regarding the classification by age groups, the combination of convolutional neural networks and temporal neural networks seems to be the best option among the analysed, and again, the larger the size of the network, the better the results. The results are promising for use in IVR systems, with the best systems achieving a gender identification error of less than 2% and a classification error by age group of less than 20%.

Aerospace ◽  
2020 ◽  
Vol 7 (12) ◽  
pp. 171
Author(s):  
Anil Doğru ◽  
Soufiane Bouarfa ◽  
Ridwan Arizar ◽  
Reyhan Aydoğan

Convolutional Neural Networks combined with autonomous drones are increasingly seen as enablers of partially automating the aircraft maintenance visual inspection process. Such an innovative concept can have a significant impact on aircraft operations. Though supporting aircraft maintenance engineers detect and classify a wide range of defects, the time spent on inspection can significantly be reduced. Examples of defects that can be automatically detected include aircraft dents, paint defects, cracks and holes, and lightning strike damage. Additionally, this concept could also increase the accuracy of damage detection and reduce the number of aircraft inspection incidents related to human factors like fatigue and time pressure. In our previous work, we have applied a recent Convolutional Neural Network architecture known by MASK R-CNN to detect aircraft dents. MASK-RCNN was chosen because it enables the detection of multiple objects in an image while simultaneously generating a segmentation mask for each instance. The previously obtained F1 and F2 scores were 62.67% and 59.35%, respectively. This paper extends the previous work by applying different techniques to improve and evaluate prediction performance experimentally. The approach uses include (1) Balancing the original dataset by adding images without dents; (2) Increasing data homogeneity by focusing on wing images only; (3) Exploring the potential of three augmentation techniques in improving model performance namely flipping, rotating, and blurring; and (4) using a pre-classifier in combination with MASK R-CNN. The results show that a hybrid approach combining MASK R-CNN and augmentation techniques leads to an improved performance with an F1 score of (67.50%) and F2 score of (66.37%).


2018 ◽  
Vol 2 ◽  
pp. e25762 ◽  
Author(s):  
Lara Lloret ◽  
Ignacio Heredia ◽  
Fernando Aguilar ◽  
Elisabeth Debusschere ◽  
Klaas Deneudt ◽  
...  

Phytoplankton form the basis of the marine food web and are an indicator for the overall status of the marine ecosystem. Changes in this community may impact a wide range of species (Capuzzo et al. 2018) ranging from zooplankton and fish to seabirds and marine mammals. Efficient monitoring of the phytoplankton community is therefore essential (Edwards et al. 2002). Traditional monitoring techniques are highly time intensive and involve taxonomists identifying and counting numerous specimens under the light microscope. With the recent development of automated sampling devices, image analysis technologies and learning algorithms, the rate of counting and identification of phytoplankton can be increased significantly (Thyssen et al. 2015). The FlowCAM (Álvarez et al. 2013) is an imaging particle analysis system for the identification and classification of phytoplankton. Within the Belgian Lifewatch observatory, monthly phytoplankton samples are taken at nine stations in the Belgian part of the North Sea. These samples are run through the FlowCAM and each particle is photographed. Next, the particles are identified based on their morphology (and fluorescence) using state-of-the-art Convolutional Neural Networks (CNNs) for computer vision. This procedure requires learning sets of expert validated images. The CNNs are specifically designed to take advantage of the two dimensional structure of these images by finding local patterns, being easier to train and having many fewer parameters than a fully connected network with the same number of hidden units. In this work we present our approach to the use of CNNs for the identification and classification of phytoplankton, testing it on several benchmarks and comparing with previous classification techniques. The network architecture used is ResNet50 (He et al. 2016). The framework is fully written in Python using the TensorFlow (Abadi, M. et al. 2016) module for Deep Learning. Deployment and exploitation of the current framework is supported by the recently started European Union Horizon 2020 programme funded project DEEP-Hybrid-Datacloud (Grant Agreement number 777435), which supports the expensive training of the system needed to develop the application and provides the necessary computational resources to the users.


Author(s):  
Jayme Barbedo ◽  
Luciano Koenigkan ◽  
Patrícia Santos

The evolution in imaging technologies and artificial intelligence algorithms, coupled with improvements in UAV technology, has enabled the use of unmanned aircraft in a wide range of applications. The feasibility of this kind of approach for cattle monitoring has been demonstrated by several studies, but practical use is still challenging due to the particular characteristics of this application, such as the need to track mobile targets and the extensive areas that need to be covered in most cases. The objective of this study was to investigate the feasibility of using a tilted angle to increase the area covered by each image. Deep Convolutional Neural Networks (Xception architecture) were used to generate the models for the experiments, which covered aspects like ideal input dimensions, effect of the distance between animals and sensor, effect of classification error on the overall detection process, and impact of physical obstacles on the accuracy of the model. Experimental results indicate that oblique images can be successfully used under certain conditions, but some practical limitations need to be addressed in order to make this approach appealing.


Author(s):  
Anil Dogru ◽  
Soufiane Bouarfa ◽  
Ridwan Arizar ◽  
Reyhan Aydogan

Convolutional Neural Networks combined with autonomous drones are increasingly seen as enablers of partially automating the aircraft maintenance visual inspection process. Such an innovative concept can have a significant impact on aircraft operations. Through supporting aircraft maintenance engineers detect and classify a wide range of defects, the time spent on inspection can significantly be reduced. Examples of defects that can be automatically detected include aircraft dents, paint defects, cracks and holes, and lightning strike damage. Additionally, this concept could also increase the accuracy of damage detection and reduce the number of aircraft inspection incidents related to human factors like fatigue and time pressure. In our previous work, we have applied a recent Convolutional Neural Network architecture known by MASK R-CNN to detect aircraft dents. MASK-RCNN was chosen because it enables the detection of multiple objects in an image while simultaneously generating a segmentation mask for each instance. The previously obtained F1 and F2 scores were 62.67% and 59.35% respectively. This paper extends the previous work by applying different techniques to improve and evaluate prediction performance experimentally. The approaches uses include (1) Balancing the original dataset by adding images without dents; (2) Increasing data homogeneity by focusing on wing images only; (3) Exploring the potential of three augmentation techniques in improving model performance namely flipping, rotating, and blurring; and (4) using a pre-classifier in combination with MASK R-CNN. The results show that a hybrid approache combining MASK R-CNN and augmentation techniques leads to an improved performance with an F1 score of (67.50%) and F2 score of (66.37%)


2016 ◽  
Vol 807 ◽  
pp. 155-166 ◽  
Author(s):  
Julia Ling ◽  
Andrew Kurzawski ◽  
Jeremy Templeton

There exists significant demand for improved Reynolds-averaged Navier–Stokes (RANS) turbulence models that are informed by and can represent a richer set of turbulence physics. This paper presents a method of using deep neural networks to learn a model for the Reynolds stress anisotropy tensor from high-fidelity simulation data. A novel neural network architecture is proposed which uses a multiplicative layer with an invariant tensor basis to embed Galilean invariance into the predicted anisotropy tensor. It is demonstrated that this neural network architecture provides improved prediction accuracy compared with a generic neural network architecture that does not embed this invariance property. The Reynolds stress anisotropy predictions of this invariant neural network are propagated through to the velocity field for two test cases. For both test cases, significant improvement versus baseline RANS linear eddy viscosity and nonlinear eddy viscosity models is demonstrated.


Author(s):  
Sarah Badr AlSumairi ◽  
Mohamed Maher Ben Ismail

Pneumonia is an infectious disease of the lungs. About one third to one half of pneumonia cases are caused by bacteria. Early diagnosis is a critical factor for a successful treatment process. Typically, the disease can be diagnosed by a radiologist using chest X-ray images. In fact, chest X-rays are currently the best available method for diagnosing pneumonia. However, the recognition of pneumonia symptoms is a challenging task that relies on the availability of expert radiologists. Such “human” diagnosis can be inaccurate and subjective due to lack of clarity and erroneous decision. Moreover, the error can increase more if the physician is requested to analyze tens of X-rays within a short period of time. Therefore, Computer-Aided Diagnosis (CAD) systems were introduced to support and assist physicians and make their efforts more productive. In this paper, we investigate, design, implement and assess customized Convolutional Neural Networks to overcome the image-based Pneumonia classification problem. Namely, ResNet-50 and DenseNet-161 models were inherited to design customized deep network architecture and improve the overall pneumonia classification accuracy. Moreover, data augmentation was deployed and associated with standard datasets to assess the proposed models. Besides, standard performance measures were used to validate and evaluate the proposed system.


2020 ◽  
Author(s):  
B Wang ◽  
Y Sun ◽  
Bing Xue ◽  
Mengjie Zhang

© 2019, Springer Nature Switzerland AG. Image classification is a difficult machine learning task, where Convolutional Neural Networks (CNNs) have been applied for over 20 years in order to solve the problem. In recent years, instead of the traditional way of only connecting the current layer with its next layer, shortcut connections have been proposed to connect the current layer with its forward layers apart from its next layer, which has been proved to be able to facilitate the training process of deep CNNs. However, there are various ways to build the shortcut connections, it is hard to manually design the best shortcut connections when solving a particular problem, especially given the design of the network architecture is already very challenging. In this paper, a hybrid evolutionary computation (EC) method is proposed to automatically evolve both the architecture of deep CNNs and the shortcut connections. Three major contributions of this work are: Firstly, a new encoding strategy is proposed to encode a CNN, where the architecture and the shortcut connections are encoded separately; Secondly, a hybrid two-level EC method, which combines particle swarm optimisation and genetic algorithms, is developed to search for the optimal CNNs; Lastly, an adjustable learning rate is introduced for the fitness evaluations, which provides a better learning rate for the training process given a fixed number of epochs. The proposed algorithm is evaluated on three widely used benchmark datasets of image classification and compared with 12 peer Non-EC based competitors and one EC based competitor. The experimental results demonstrate that the proposed method outperforms all of the peer competitors in terms of classification accuracy.


2021 ◽  
Vol 2089 (1) ◽  
pp. 012013
Author(s):  
Priyadarshini Chatterjee ◽  
Dutta Sushama Rani

Abstract Automated diagnosis of diseases in the recent years have gain lots of advantages and potential. Specially automated screening of cancers has helped the clinicians over the time. Sometimes it is seen that the diagnosis of the clinicians is biased but automated detection can help them to come to a proper conclusion. Automated screening is implemented using either artificial inter connected system or convolutional inter connected system. As Artificial neural network is slow in computation, so Convolutional Neural Network has achieved lots of importance in the recent years. It is also seen that Convolutional Neural Network architecture requires a smaller number of datasets. This also provides them an edge over Artificial Neural Networks. Convolutional Neural Networks is used for both segmentation and classification. Image dissection is one of the important steps in the model used for any kind of image analysis. This paper surveys various such Convolutional Neural Networks that are used for medical image analysis.


Complexity ◽  
2019 ◽  
Vol 2019 ◽  
pp. 1-16 ◽  
Author(s):  
Alejandro Baldominos ◽  
Yago Saez ◽  
Pedro Isasi

Neuroevolution is the field of study that uses evolutionary computation in order to optimize certain aspect of the design of neural networks, most often its topology and hyperparameters. The field was introduced in the late-1980s, but only in the latest years the field has become mature enough to enable the optimization of deep learning models, such as convolutional neural networks. In this paper, we rely on previous work to apply neuroevolution in order to optimize the topology of deep neural networks that can be used to solve the problem of handwritten character recognition. Moreover, we take advantage of the fact that evolutionary algorithms optimize a population of candidate solutions, by combining a set of the best evolved models resulting in a committee of convolutional neural networks. This process is enhanced by using specific mechanisms to preserve the diversity of the population. Additionally, in this paper, we address one of the disadvantages of neuroevolution: the process is very expensive in terms of computational time. To lessen this issue, we explore the performance of topology transfer learning: whether the best topology obtained using neuroevolution for a certain domain can be successfully applied to a different domain. By doing so, the expensive process of neuroevolution can be reused to tackle different problems, turning it into a more appealing approach for optimizing the design of neural networks topologies. After evaluating our proposal, results show that both the use of neuroevolved committees and the application of topology transfer learning are successful: committees of convolutional neural networks are able to improve classification results when compared to single models, and topologies learned for one problem can be reused for a different problem and data with a good performance. Additionally, both approaches can be combined by building committees of transferred topologies, and this combination attains results that combine the best of both approaches.


2020 ◽  
Vol 143 ◽  
pp. 02015
Author(s):  
Li Zherui ◽  
Cai Huiwen

Sea ice classification is one of the important tasks of sea ice monitoring. Accurate extraction of sea ice types is of great significance on sea ice conditions assessment, smooth navigation and safty marine operations. Sentinel-2 is an optical satellite launched by the European Space Agency. High spatial resolution and wide range imaging provide powerful support for sea ice monitoring. However, traditional supervised classification method is difficult to achieve fine results for small sample features. In order to solve the problem, this paper proposed a sea ice extraction method based on deep learning and it was applied to Liaodong Bay in Bohai Sea, China. The convolutional neural network was used to extract and classify the feature of the image from Sentinel-2. The results showed that the overall accuracy of the algorithm was 85.79% which presented a significant improvement compared with the tranditional algorithms, such as minimum distance method, maximum likelihood method, Mahalanobis distance method, and support vector machine method. The method proposed in this paper, which combines convolutional neural networks and high-resolution multispectral data, provides a new idea for remote sensing monitoring of sea ice.


Sign in / Sign up

Export Citation Format

Share Document