scholarly journals Constrained Ensemble Langevin Monte Carlo

2021 ◽  
Vol 0 (0) ◽  
pp. 0
Author(s):  
Zhiyan Ding ◽  
Qin Li

<p style='text-indent:20px;'>The classical Langevin Monte Carlo method looks for samples from a target distribution by descending the samples along the gradient of the target distribution. The method enjoys a fast convergence rate. However, the numerical cost is sometimes high because each iteration requires the computation of a gradient. One approach to eliminate the gradient computation is to employ the concept of "ensemble." A large number of particles are evolved together so the neighboring particles provide gradient information to each other. In this article, we discuss two algorithms that integrate the ensemble feature into LMC, and the associated properties.</p><p style='text-indent:20px;'>In particular, we find that if one directly surrogates the gradient using the ensemble approximation, the algorithm, termed Ensemble Langevin Monte Carlo, is unstable due to a high variance term. If the gradients are replaced by the ensemble approximations only in a constrained manner, to protect from the unstable points, the algorithm, termed Constrained Ensemble Langevin Monte Carlo, resembles the classical LMC up to an ensemble error but removes most of the gradient computation.</p>

Author(s):  
Galiya Z. Lotova ◽  
Guennady A. Mikhailov

AbstractAlgorithms of Monte Carlo method for estimating probabilistic moments of the parameter of time exponential asymptotics of particle flux with multiplication in a random medium are constructed. It was analytically and numerically shown that the asymptotics of the mean number of particles is close to an exponential one with the factor containing a summand proportional to square of time.


1991 ◽  
Vol 05 (01n02) ◽  
pp. 113-118
Author(s):  
A. Moreo

We study the behavior of the mean number of particles <n> as a function of the chemical potential μ in the two dimensional Hubbard model with both attractive and repulsive interaction, using a quantum Monte Carlo method. Working at U/t=10, 4 and −4 on lattices with 4×4, 6×6 and 8×8 sites, we do not find evidence of phase separation.


Author(s):  
Jinghui Chen ◽  
Dongruo Zhou ◽  
Yiqi Tang ◽  
Ziyan Yang ◽  
Yuan Cao ◽  
...  

Adaptive gradient methods, which adopt historical gradient information to automatically adjust the learning rate, despite the nice property of fast convergence, have been observed to generalize worse than stochastic gradient descent (SGD) with momentum in training deep neural networks. This leaves how to close the generalization gap of adaptive gradient methods an open problem. In this work, we show that adaptive gradient methods such as Adam, Amsgrad, are sometimes "over adapted". We design a new algorithm, called Partially adaptive momentum estimation method, which unifies the Adam/Amsgrad with SGD by introducing a partial adaptive parameter $p$, to achieve the best from both worlds. We also prove the convergence rate of our proposed algorithm to a stationary point in the stochastic nonconvex optimization setting. Experiments on standard benchmarks show that our proposed algorithm can maintain fast convergence rate as Adam/Amsgrad while generalizing as well as SGD in training deep neural networks. These results would suggest practitioners pick up adaptive gradient methods once again for faster training of deep neural networks.


Author(s):  
Xiaomei Mo ◽  
Jie Xu

This paper studies the convergence rate and consistency of Empirical Risk Minimization algorithm, where the samples need not be independent and identically distributed (i.i.d.) but can come from uniformly ergodic Markov chain (u.e.M.c.). We firstly establish the generalization bounds of Empirical Risk Minimization algorithm with u.e.M.c. samples. Then we deduce that the Empirical Risk Minimization algorithm on the base of u.e.M.c. samples is consistent and owns a fast convergence rate.


Sign in / Sign up

Export Citation Format

Share Document