Parallel Version of the Mirror Descent Algorithm for the Two-Armed Bandit Problem

Author(s):  
Alexander Kolnogorov ◽  
Dmitry Shiyan
2021 ◽  
Vol 13 (2) ◽  
pp. 9-39
Author(s):  
Александр Валерианович Колногоров ◽  
Alexander Kolnogorov ◽  
Александр Викторович Назин ◽  
Alexander Nazin ◽  
Дмитрий Николаевич Шиян ◽  
...  

We consider the minimax setup for the two-armed bandit problem as applied to data processing if there are two alternative processing methods with different a priori unknown efficiencies. One should determine the most efficient method and provide its predominant application. To this end, we use the mirror descent algorithm (MDA). It is well-known that corresponding minimax risk has the order of $N^{1/2$ with $N$ being the number of processed data and this bound is unimprovable in order. We propose a batch version of the MDA which allows processing data by packets that is especially important if parallel data processing can be provided. In this case, the processing time is determined by the number of  batches rather than by the total number of data. Unexpectedly, it turned out that the batch version behaves unlike the ordinary one even if the number of packets is large. Moreover, the batch version provides significantly smaller value of the minimax risk, i.e., it considerably improves a control performance. We explain this result by considering another batch modification of the MDA which behavior is close to behavior of the ordinary version and minimax risk is close as well. Our estimates use invariant descriptions of the algorithms based on Gaussian approximations of incomes in batches of data in the domain of ``close'' distributions and are obtained by Monte-Carlo simulations.


2005 ◽  
Vol 41 (4) ◽  
pp. 368-384 ◽  
Author(s):  
A. B. Juditsky ◽  
A. V. Nazin ◽  
A. B. Tsybakov ◽  
N. Vayatis

2014 ◽  
Vol 75 (6) ◽  
pp. 1010-1016
Author(s):  
A. V. Nazin ◽  
S. V. Anulova ◽  
A. A. Tremba

2017 ◽  
Vol 29 (3) ◽  
pp. 825-860 ◽  
Author(s):  
Yunwen Lei ◽  
Ding-Xuan Zhou

We study the convergence of the online composite mirror descent algorithm, which involves a mirror map to reflect the geometry of the data and a convex objective function consisting of a loss and a regularizer possibly inducing sparsity. Our error analysis provides convergence rates in terms of properties of the strongly convex differentiable mirror map and the objective function. For a class of objective functions with Hölder continuous gradients, the convergence rates of the excess (regularized) risk under polynomially decaying step sizes have the order [Formula: see text] after [Formula: see text] iterates. Our results improve the existing error analysis for the online composite mirror descent algorithm by avoiding averaging and removing boundedness assumptions, and they sharpen the existing convergence rates of the last iterate for online gradient descent without any boundedness assumptions. Our methodology mainly depends on a novel error decomposition in terms of an excess Bregman distance, refined analysis of self-bounding properties of the objective function, and the resulting one-step progress bounds.


Sign in / Sign up

Export Citation Format

Share Document