scholarly journals Efficient Learning of Contextual Mappings by Context-Dependent Neural Nets

Author(s):  
Piotr Ciskowski
2004 ◽  
Vol 15 (6) ◽  
pp. 1367-1377 ◽  
Author(s):  
P. Ciskowski ◽  
E. Rafajlowicz

Author(s):  
BAO WANG ◽  
STAN J. OSHER

Improving the accuracy and robustness of deep neural nets (DNNs) and adapting them to small training data are primary tasks in deep learning (DL) research. In this paper, we replace the output activation function of DNNs, typically the data-agnostic softmax function, with a graph Laplacian-based high-dimensional interpolating function which, in the continuum limit, converges to the solution of a Laplace–Beltrami equation on a high-dimensional manifold. Furthermore, we propose end-to-end training and testing algorithms for this new architecture. The proposed DNN with graph interpolating activation integrates the advantages of both deep learning and manifold learning. Compared to the conventional DNNs with the softmax function as output activation, the new framework demonstrates the following major advantages: First, it is better applicable to data-efficient learning in which we train high capacity DNNs without using a large number of training data. Second, it remarkably improves both natural accuracy on the clean images and robust accuracy on the adversarial images crafted by both white-box and black-box adversarial attacks. Third, it is a natural choice for semi-supervised learning. This paper is a significant extension of our earlier work published in NeurIPS, 2018. For reproducibility, the code is available at https://github.com/BaoWangMath/DNN-DataDependentActivation.


1995 ◽  
Vol 06 (03) ◽  
pp. 273-282 ◽  
Author(s):  
LLUIS GARRIDO ◽  
VICENS GAITAN ◽  
MIQUEL SERRA-RICART ◽  
XAVIER CALBET

We present a new method based on multilayer feedforward neural nets for displaying an n-dimensional distribution in a projected space of 1, 2 or 3 dimensions. A fully nonlinear net with several hidden layers is used. Efficient learning is achieved using multi-seed backpropagation. As a principal component analysis (PCA), the proposed method is useful for extracting information on the structure of the data set, but unlike the PCA, the transformation between the original distribution and the projected one is not restricted to be linear. Artificial examples and a real application are presented in order to show the reliability and potential of the method.


2012 ◽  
Vol 24 (8) ◽  
pp. 1967-2006 ◽  
Author(s):  
Ruslan Salakhutdinov ◽  
Geoffrey Hinton

We present a new learning algorithm for Boltzmann machines that contain many layers of hidden variables. Data-dependent statistics are estimated using a variational approximation that tends to focus on a single mode, and data-independent statistics are estimated using persistent Markov chains. The use of two quite different techniques for estimating the two types of statistic that enter into the gradient of the log likelihood makes it practical to learn Boltzmann machines with multiple hidden layers and millions of parameters. The learning can be made more efficient by using a layer-by-layer pretraining phase that initializes the weights sensibly. The pretraining also allows the variational inference to be initialized sensibly with a single bottom-up pass. We present results on the MNIST and NORB data sets showing that deep Boltzmann machines learn very good generative models of handwritten digits and 3D objects. We also show that the features discovered by deep Boltzmann machines are a very effective way to initialize the hidden layers of feedforward neural nets, which are then discriminatively fine-tuned.


2014 ◽  
Vol 45 (3) ◽  
pp. 153-163 ◽  
Author(s):  
Sanne Nauts ◽  
Oliver Langner ◽  
Inge Huijsmans ◽  
Roos Vonk ◽  
Daniël H. J. Wigboldus

Asch’s seminal research on “Forming Impressions of Personality” (1946) has widely been cited as providing evidence for a primacy-of-warmth effect, suggesting that warmth-related judgments have a stronger influence on impressions of personality than competence-related judgments (e.g., Fiske, Cuddy, & Glick, 2007 ; Wojciszke, 2005 ). Because this effect does not fit with Asch’s Gestalt-view on impression formation and does not readily follow from the data presented in his original paper, the goal of the present study was to critically examine and replicate the studies of Asch’s paper that are most relevant to the primacy-of-warmth effect. We found no evidence for a primacy-of-warmth effect. Instead, the role of warmth was highly context-dependent, and competence was at least as important in shaping impressions as warmth.


Author(s):  
Alp Aslan ◽  
Anuscheh Samenieh ◽  
Tobias Staudigl ◽  
Karl-Heinz T. Bäuml

Changing environmental context during encoding can influence episodic memory. This study examined the memorial consequences of environmental context change in children. Kindergartners, first and fourth graders, and young adults studied two lists of items, either in the same room (no context change) or in two different rooms (context change), and subsequently were tested on the two lists in the room in which the second list was encoded. As expected, in adults, the context change impaired recall of the first list and improved recall of the second. Whereas fourth graders showed the same pattern of results as adults, in both kindergartners and first graders no memorial effects of the context change arose. The results indicate that the two effects of environmental context change develop contemporaneously over middle childhood and reach maturity at the end of the elementary school days. The findings are discussed in light of both retrieval-based and encoding-based accounts of context-dependent memory.


Sign in / Sign up

Export Citation Format

Share Document