Deep neural network to extract high-level features and labels in multi-label classification problems

2020 ◽  
Vol 413 ◽  
pp. 259-270
Author(s):  
Marilyn Bello ◽  
Gonzalo Nápoles ◽  
Ricardo Sánchez ◽  
Rafael Bello ◽  
Koen Vanhoof
Electronics ◽  
2021 ◽  
Vol 10 (21) ◽  
pp. 2687
Author(s):  
Eun-Hun Lee ◽  
Hyeoncheol Kim

The significant advantage of deep neural networks is that the upper layer can capture the high-level features of data based on the information acquired from the lower layer by stacking layers deeply. Since it is challenging to interpret what knowledge the neural network has learned, various studies for explaining neural networks have emerged to overcome this problem. However, these studies generate the local explanation of a single instance rather than providing a generalized global interpretation of the neural network model itself. To overcome such drawbacks of the previous approaches, we propose the global interpretation method for the deep neural network through features of the model. We first analyzed the relationship between the input and hidden layers to represent the high-level features of the model, then interpreted the decision-making process of neural networks through high-level features. In addition, we applied network pruning techniques to make concise explanations and analyzed the effect of layer complexity on interpretability. We present experiments on the proposed approach using three different datasets and show that our approach could generate global explanations on deep neural network models with high accuracy and fidelity.


2020 ◽  
Vol 10 (6) ◽  
pp. 1265-1273
Author(s):  
Lili Chen ◽  
Huoyao Xu

Sleep apnea (SA) is a common sleep disorders affecting the sleep quality. Therefore the automatic SA detection has far-reaching implications for patients and physicians. In this paper, a novel approach is developed based on deep neural network (DNN) for automatic diagnosis SA. To this end, five features are extracted from electrocardiogram (ECG) signals through wavelet decomposition and sample entropy. The deep neural network is constructed by two-layer stacked sparse autoencoder (SSAE) network and one softmax layer. The softmax layer is added at the top of the SSAE network for diagnosing SA. Afterwards, the SSAE network can get more effective high-level features from raw features. The experimental results reveal that the performance of deep neural network can accomplish an accuracy of 96.66%, a sensitivity of 96.25%, and a specificity of 97%. In addition, the performance of deep neural network outperforms the comparison models including support vector machine (SVM), random forest (RF), and extreme learning machine (ELM). Finally, the experimental results reveal that the proposed method can be valid applied to automatic SA event detection.


eLife ◽  
2018 ◽  
Vol 7 ◽  
Author(s):  
Iris IA Groen ◽  
Michelle R Greene ◽  
Christopher Baldassano ◽  
Li Fei-Fei ◽  
Diane M Beck ◽  
...  

Inherent correlations between visual and semantic features in real-world scenes make it difficult to determine how different scene properties contribute to neural representations. Here, we assessed the contributions of multiple properties to scene representation by partitioning the variance explained in human behavioral and brain measurements by three feature models whose inter-correlations were minimized a priori through stimulus preselection. Behavioral assessments of scene similarity reflected unique contributions from a functional feature model indicating potential actions in scenes as well as high-level visual features from a deep neural network (DNN). In contrast, similarity of cortical responses in scene-selective areas was uniquely explained by mid- and high-level DNN features only, while an object label model did not contribute uniquely to either domain. The striking dissociation between functional and DNN features in their contribution to behavioral and brain representations of scenes indicates that scene-selective cortex represents only a subset of behaviorally relevant scene information.


2020 ◽  
Author(s):  
Song Tong ◽  
Xuefeng Liang ◽  
Takatsune Kumada ◽  
Sunao Iwaki

Empirical evidence has shown that there is an ideal arrangement of facial features (ideal ratios) that can optimize the attractiveness of a person's face. These putative ratios define facial attractiveness in terms of spatial relations and provide important rules for measuring the attractiveness of a face. In this paper, we show that a deep neural network (DNN) model can learn putative ratios based only on categorical annotation when no annotated facial features for attractiveness are explicitly given. To this end, we conducted three experiments. In Experiment 1, we trained a DNN model to recognize facial attractiveness using four category-specific neurons (female/male $\times$ high/low attractiveness). In Experiment 2, face-like images were generated by reversing the DNN model (e.g., deconvolution). These images depict the intuitive attributes of the four categories of facial attractiveness and reveal certain consistencies with reported evidence on the putative ratios of facial attractiveness. In Experiment 3, simulated psychophysical experiments on facial images with varying ratios of features reveal changes in the activity of the category-specific neurons that are remarkably similar to those of human judgements reported in a previous study. These results show that the trained DNN model can learn putative ratios as key features for the representation of facial attractiveness. These findings advance our understanding of facial attractiveness and high-level human perception.


2017 ◽  
Author(s):  
Michael F. Bonner ◽  
Russell A. Epstein

ABSTRACTBiologically inspired deep convolutional neural networks (CNNs), trained for computer vision tasks, have been found to predict cortical responses with remarkable accuracy. However, the complex internal operations of these models remain poorly understood, and the factors that account for their success are unknown. Here we developed a set of techniques for using CNNs to gain insights into the computational mechanisms underlying cortical responses. We focused on responses in the occipital place area (OPA), a scene-selective region of dorsal occipitoparietal cortex. In a previous study, we showed that fMRI activation patterns in the OPA contain information about the navigational affordances of scenes: that is, information about where one can and cannot move within the immediate environment. We hypothesized that this affordance information could be extracted using a set of purely feedforward computations. To test this idea, we examined a deep CNN with a feedforward architecture that had been previously trained for scene classification. We found that the CNN was highly predictive of OPA representations, and, importantly, that it accounted for the portion of OPA variance that reflected the navigational affordances of scenes. The CNN could thus serve as an image-computable candidate model of affordance-related responses in the OPA. We then ran a series of in silico experiments on this model to gain insights into its internal computations. These analyses showed that the computation of affordance-related features relied heavily on visual information at high-spatial frequencies and cardinal orientations, both of which have previously been identified as low-level stimulus preferences of scene-selective visual cortex. These computations also exhibited a strong preference for information in the lower visual field, which is consistent with known retinotopic biases in the OPA. Visualizations of feature selectivity within the CNN suggested that affordance-based responses encoded features that define the layout of the spatial environment, such as boundary-defining junctions and large extended surfaces. Together, these results map the sensory functions of the OPA onto a fully quantitative model that provides insights into its visual computations. More broadly, they advance integrative techniques for understanding visual cortex across multiple level of analysis: from the identification of cortical sensory functions to the modeling of their underlying algorithmic implementations.AUTHOR SUMMARYHow does visual cortex compute behaviorally relevant properties of the local environment from sensory inputs? For decades, computational models have been able to explain only the earliest stages of biological vision, but recent advances in the engineering of deep neural networks have yielded a breakthrough in the modeling of high-level visual cortex. However, these models are not explicitly designed for testing neurobiological theories, and, like the brain itself, their complex internal operations remain poorly understood. Here we examined a deep neural network for insights into the cortical representation of the navigational affordances of visual scenes. In doing so, we developed a set of high-throughput techniques and statistical tools that are broadly useful for relating the internal operations of neural networks with the information processes of the brain. Our findings demonstrate that a deep neural network with purely feedforward computations can account for the processing of navigational layout in high-level visual cortex. We next performed a series of experiments and visualization analyses on this neural network, which characterized a set of stimulus input features that may be critical for computing navigationally related cortical representations and identified a set of high-level, complex scene features that may serve as a basis set for the cortical coding of navigational layout. These findings suggest a computational mechanism through which high-level visual cortex might encode the spatial structure of the local navigational environment, and they demonstrate an experimental approach for leveraging the power of deep neural networks to understand the visual computations of the brain.


2017 ◽  
Author(s):  
Iris I. A. Groen ◽  
Michelle R. Greene ◽  
Christopher Baldassano ◽  
Li Fei-Fei ◽  
Diane M. Beck ◽  
...  

AbstractInherent correlations between visual and semantic features in real-world scenes make it difficult to determine how different scene properties contribute to neural representations. Here, we assessed the contributions of multiple properties to scene representation by partitioning the variance explained in human behavioral and brain measurements by three feature models whose inter-correlations were minimized a priori through stimulus preselection. Behavioral assessments of scene similarity reflected unique contributions from a functional feature model indicating potential actions in scenes as well as high-level visual features from a deep neural network (DNN). In contrast, similarity of cortical responses in scene-selective areas was uniquely explained by mid- and high-level DNN features only, while an object label model did not contribute uniquely to either domain. The striking dissociation between functional and DNN features in their contribution to behavioral and brain representations of scenes indicates that scene-selective cortex represents only a subset of behaviorally relevant scene information.


Author(s):  
V. Nisha Jenipher ◽  
S. Princy Suganthi Bai ◽  
A. Venkatesh ◽  
K. Ravindran ◽  
Adlin Sheeba

Author(s):  
Siti Noorain Mohmad Yousoff ◽  
‘Amirah Baharin ◽  
Afnizanfaizal Abdullah

<p>Bio-based chemical products such as xylitol have achieved remarkable attentions both in pharmaceutical and food industries due to their several advantages such as sugar substitute that can help diabetic patients and help in preventing tooth decay problem. To produce xylitol, recently, microbial host such as E. Coli often used as it is predicted that E. Coli can produce high level of xylitol. Therefore, metabolic engineering need to be done towards E. Coli and powerful tools are needed to manipulate, simulate and analyse the E. Coli metabolic pathway. Artificial intelligence methods such as deep neural network offer an efficient and powerful approach to be used to analyse the xylitol production value and at the same time to predict which genes and pathway that give biggest effect in the process to produce xylitol in E. Coli. Results show that, with an absence of genes pgi, tkt and tala, xylitol production can be boosted up to the higher level.</p>


Sign in / Sign up

Export Citation Format

Share Document