scholarly journals Lipschitz Continuity and Approximate Equilibria

Algorithmica ◽  
2020 ◽  
Vol 82 (10) ◽  
pp. 2927-2954
Author(s):  
Argyrios Deligkas ◽  
John Fearnley ◽  
Paul Spirakis

Abstract In this paper, we study games with continuous action spaces and non-linear payoff functions. Our key insight is that Lipschitz continuity of the payoff function allows us to provide algorithms for finding approximate equilibria in these games. We begin by studying Lipschitz games, which encompass, for example, all concave games with Lipschitz continuous payoff functions. We provide an efficient algorithm for computing approximate equilibria in these games. Then we turn our attention to penalty games, which encompass biased games and games in which players take risk into account. Here we show that if the penalty function is Lipschitz continuous, then we can provide a quasi-polynomial time approximation scheme. Finally, we study distance biased games, where we present simple strongly polynomial time algorithms for finding best responses in $$L_1$$ L 1 and $$L_2^2$$ L 2 2 biased games, and then use these algorithms to provide strongly polynomial algorithms that find 2/3 and 5/7 approximate equilibria for these norms, respectively.

Author(s):  
Dan Dadush ◽  
László A. Végh ◽  
Giacomo Zambelli

We present a new class of polynomial-time algorithms for submodular function minimization (SFM) as well as a unified framework to obtain strongly polynomial SFM algorithms. Our algorithms are based on simple iterative methods for the minimum-norm problem, such as the conditional gradient and Fujishige–Wolfe algorithms. We exhibit two techniques to turn simple iterative methods into polynomial-time algorithms. First, we adapt the geometric rescaling technique, which has recently gained attention in linear programming, to SFM and obtain a weakly polynomial bound [Formula: see text]. Second, we exhibit a general combinatorial black box approach to turn [Formula: see text]-approximate SFM oracles into strongly polynomial exact SFM algorithms. This framework can be applied to a wide range of combinatorial and continuous algorithms, including pseudo-polynomial ones. In particular, we can obtain strongly polynomial algorithms by a repeated application of the conditional gradient or of the Fujishige–Wolfe algorithm. Combined with the geometric rescaling technique, the black box approach provides an [Formula: see text] algorithm. Finally, we show that one of the techniques we develop in the paper can also be combined with the cutting-plane method of Lee et al., yielding a simplified variant of their [Formula: see text] algorithm.


2016 ◽  
Vol 26 (2) ◽  
pp. 159-171
Author(s):  
Mehdi Ghiyasvand

In this paper, a new problem on a directed network is presented. Let D be a feasible network such that all arc capacities are equal to U. Given a t > 0, the network D with arc capacities U - t is called the t-network. The goal of the problem is to compute the largest t such that the t-network is feasible. First, we present a weakly polynomial time algorithm to solve this problem, which runs in O(log(nU)) maximum flow computations, where n is the number of nodes. Then, an O(m2n) time approach is presented, where m is the number of arcs. Both the weakly and strongly polynomial algorithms are inspired by McCormick and Ervolina (1994).


Author(s):  
Buvanesh Pandian V

Reinforcement learning is a mathematical framework for agents to interact intelligently with their environment. Unlike supervised learning, where a system learns with the help of labeled data, reinforcement learning agents learn how to act by trial and error only receiving a reward signal from their environments. A field where reinforcement learning has been prominently successful is robotics [3]. However, real-world control problems are also particularly challenging because of the noise and high- dimensionality of input data (e.g., visual input). In recent years, in the field of supervised learning, deep neural networks have been successfully used to extract meaning from this kind of data. Building on these advances, deep reinforcement learning was used to solve complex problems like Atari games and Go. Mnih et al. [1] built a system with fixed hyper parameters able to learn to play 49 different Atari games only from raw pixel inputs. However, in order to apply the same methods to real-world control problems, deep reinforcement learning has to be able to deal with continuous action spaces. Discretizing continuous action spaces would scale poorly, since the number of discrete actions grows exponentially with the dimensionality of the action. Furthermore, having a parametrized policy can be advantageous because it can generalize in the action space. Therefore with this thesis we study state-of-the-art deep reinforcement learning algorithm, Deep Deterministic Policy Gradients. We provide a theoretical comparison to other popular methods, an evaluation of its performance, identify its limitations and investigate future directions of research. The remainder of the thesis is organized as follows. We start by introducing the field of interest, machine learning, focusing our attention of deep learning and reinforcement learning. We continue by describing in details the two main algorithms, core of this study, namely Deep Q-Network (DQN) and Deep Deterministic Policy Gradients (DDPG). We then provide implementatory details of DDPG and our test environment, followed by a description of benchmark test cases. Finally, we discuss the results of our evaluation, identifying limitations of the current approach and proposing future avenues of research.


2019 ◽  
Vol 36 (02) ◽  
pp. 1940003
Author(s):  
Liwei Zhang ◽  
Shengzhe Gao ◽  
Saoyan Guo

In this paper, we study the stability of stochastic second-order programming when the probability measure is perturbed. Under the Lipschitz continuity of the objective function and metric regularity of the feasible set-valued mapping, the outer semicontinuity of the optimal solution set and Lipschitz continuity of optimal values are demonstrated. Moreover, we prove that, if the constraint non-degeneracy condition and strong second-order sufficient condition hold at a local minimum point of the original problem, there exists a Lipschitz continuous solution path satisfying the Karush–Kuhn–Tucker conditions.


2006 ◽  
Vol 08 (01) ◽  
pp. 1-8 ◽  
Author(s):  
MINGBAO SUN ◽  
XIAOPING YANG

For a Carnot group G of step two, we prove that H-convex functions are locally bounded from above. Therefore, H-convex functions on a Carnot group G of step two are locally Lipschitz continuous by using recent results by Magnani.


2019 ◽  
Vol 2019 ◽  
pp. 1-11
Author(s):  
Zhou Yu ◽  
Xiao Bing

Let D be the unit disk in the complex plane C and denote T=∂D. Write Hom+T,∂Ω for the class of all sense-preserving homeomorphism of T onto the boundary of a C2 convex Jordan domain Ω. In this paper, five equivalent conditions for the solutions of triharmonic equations ∂z∂z¯3ω=ff∈CD¯ with Dirichlet boundary value conditions ωzz¯zz¯T=γ2∈CT,ωzz¯T=γ1∈CT and ωT=γ0∈Hom+T,∂Ω to be Lipschitz continuous are presented.


Sign in / Sign up

Export Citation Format

Share Document