convex optimization problems
Recently Published Documents


TOTAL DOCUMENTS

269
(FIVE YEARS 95)

H-INDEX

25
(FIVE YEARS 3)

Author(s):  
Yang Zheng ◽  
Giovanni Fantuzzi

AbstractWe prove decomposition theorems for sparse positive (semi)definite polynomial matrices that can be viewed as sparsity-exploiting versions of the Hilbert–Artin, Reznick, Putinar, and Putinar–Vasilescu Positivstellensätze. First, we establish that a polynomial matrix P(x) with chordal sparsity is positive semidefinite for all $$x\in \mathbb {R}^n$$ x ∈ R n if and only if there exists a sum-of-squares (SOS) polynomial $$\sigma (x)$$ σ ( x ) such that $$\sigma P$$ σ P is a sum of sparse SOS matrices. Second, we show that setting $$\sigma (x)=(x_1^2 + \cdots + x_n^2)^\nu $$ σ ( x ) = ( x 1 2 + ⋯ + x n 2 ) ν for some integer $$\nu $$ ν suffices if P is homogeneous and positive definite globally. Third, we prove that if P is positive definite on a compact semialgebraic set $$\mathcal {K}=\{x:g_1(x)\ge 0,\ldots ,g_m(x)\ge 0\}$$ K = { x : g 1 ( x ) ≥ 0 , … , g m ( x ) ≥ 0 } satisfying the Archimedean condition, then $$P(x) = S_0(x) + g_1(x)S_1(x) + \cdots + g_m(x)S_m(x)$$ P ( x ) = S 0 ( x ) + g 1 ( x ) S 1 ( x ) + ⋯ + g m ( x ) S m ( x ) for matrices $$S_i(x)$$ S i ( x ) that are sums of sparse SOS matrices. Finally, if $$\mathcal {K}$$ K is not compact or does not satisfy the Archimedean condition, we obtain a similar decomposition for $$(x_1^2 + \cdots + x_n^2)^\nu P(x)$$ ( x 1 2 + ⋯ + x n 2 ) ν P ( x ) with some integer $$\nu \ge 0$$ ν ≥ 0 when P and $$g_1,\ldots ,g_m$$ g 1 , … , g m are homogeneous of even degree. Using these results, we find sparse SOS representation theorems for polynomials that are quadratic and correlatively sparse in a subset of variables, and we construct new convergent hierarchies of sparsity-exploiting SOS reformulations for convex optimization problems with large and sparse polynomial matrix inequalities. Numerical examples demonstrate that these hierarchies can have a significantly lower computational complexity than traditional ones.


Author(s):  
Erdal KORKMAZ ◽  
Abdulhamit Ozdemir

In this paper, we investigate the asymptotic stability of solutions for a class of nonlinear fractional neutral differential systems with time dependent delays when the given delays are unbounded. An example is used to show the efficacy of the theorems. The LMI tool box was used to calculate the solutions to the convex optimization problems.


Author(s):  
Thierno M. M. Sow

In this paper, a new iterative method  for solving  convex minimization  problems over the set of common fixed points of quasi-nonexpansive and demicontractive mappings is constructed. Convergence theorems are also proved in Hilbert spaces without any compactness assumption. As an application, we shall utilize our results to solve quadratic optimization  problems involving bounded linear operator. Our theorems are significant improvements on several important recent results.


Author(s):  
Lucas Brynte ◽  
Viktor Larsson ◽  
José Pedro Iglesias ◽  
Carl Olsson ◽  
Fredrik Kahl

AbstractWhy is it that semidefinite relaxations have been so successful in numerous applications in computer vision and robotics for solving non-convex optimization problems involving rotations? In studying the empirical performance, we note that there are few failure cases reported in the literature, in particular for estimation problems with a single rotation, motivating us to gain further theoretical understanding. A general framework based on tools from algebraic geometry is introduced for analyzing the power of semidefinite relaxations of problems with quadratic objective functions and rotational constraints. Applications include registration, hand–eye calibration, and rotation averaging. We characterize the extreme points and show that there exist failure cases for which the relaxation is not tight, even in the case of a single rotation. We also show that some problem classes are always tight given an appropriate parametrization. Our theoretical findings are accompanied with numerical simulations, providing further evidence and understanding of the results.


Author(s):  
V. Kungurtsev ◽  
F. Rinaldi

AbstractIn this paper, we consider stochastic weakly convex optimization problems, however without the existence of a stochastic subgradient oracle. We present a derivative free algorithm that uses a two point approximation for computing a gradient estimate of the smoothed function. We prove convergence at a similar rate as state of the art methods, however with a larger constant, and report some numerical results showing the effectiveness of the approach.


Sign in / Sign up

Export Citation Format

Share Document