scholarly journals Randomized sketch descent methods for non-separable linearly constrained optimization

Author(s):  
Ion Necoara ◽  
Martin Takáč

Abstract In this paper we consider large-scale smooth optimization problems with multiple linear coupled constraints. Due to the non-separability of the constraints, arbitrary random sketching would not be guaranteed to work. Thus, we first investigate necessary and sufficient conditions for the sketch sampling to have well-defined algorithms. Based on these sampling conditions we develop new sketch descent methods for solving general smooth linearly constrained problems, in particular, random sketch descent (RSD) and accelerated random sketch descent (A-RSD) methods. To our knowledge, this is the first convergence analysis of RSD algorithms for optimization problems with multiple non-separable linear constraints. For the general case, when the objective function is smooth and non-convex, we prove for the non-accelerated variant sublinear rate in expectation for an appropriate optimality measure. In the smooth convex case, we derive for both algorithms, non-accelerated and A-RSD, sublinear convergence rates in the expected values of the objective function. Additionally, if the objective function satisfies a strong convexity type condition, both algorithms converge linearly in expectation. In special cases, where complexity bounds are known for some particular sketching algorithms, such as coordinate descent methods for optimization problems with a single linear coupled constraint, our theory recovers the best known bounds. Finally, we present several numerical examples to illustrate the performances of our new algorithms.

Acta Numerica ◽  
2005 ◽  
Vol 14 ◽  
pp. 299-361 ◽  
Author(s):  
Nick Gould ◽  
Dominique Orban ◽  
Philippe Toint

Recent developments in numerical methods for solving large differentiable nonlinear optimization problems are reviewed. State-of-the-art algorithms for solving unconstrained, bound-constrained, linearly constrained and non-linearly constrained problems are discussed. As well as important conceptual advances and theoretical aspects, emphasis is also placed on more practical issues, such as software availability.


Author(s):  
T. E. Potter ◽  
K. D. Willmert ◽  
M. Sathyamoorthy

Abstract Mechanism path generation problems which use link deformations to improve the design lead to optimization problems involving a nonlinear sum-of-squares objective function subjected to a set of linear and nonlinear constraints. Inclusion of the deformation analysis causes the objective function evaluation to be computationally expensive. An optimization method is presented which requires relatively few objective function evaluations. The algorithm, based on the Gauss method for unconstrained problems, is developed as an extension of the Gauss constrained technique for linear constraints and revises the Gauss nonlinearly constrained method for quadratic constraints. The derivation of the algorithm, using a Lagrange multiplier approach, is based on the Kuhn-Tucker conditions so that when the iteration process terminates, these conditions are automatically satisfied. Although the technique was developed for mechanism problems, it is applicable to any optimization problem having the form of a sum of squares objective function subjected to nonlinear constraints.


2020 ◽  
Author(s):  
Qing Tao

The extrapolation strategy raised by Nesterov, which can accelerate the convergence rate of gradient descent methods by orders of magnitude when dealing with smooth convex objective, has led to tremendous success in training machine learning tasks. In this paper, we theoretically study its strength in the convergence of individual iterates of general non-smooth convex optimization problems, which we name \textit{individual convergence}. We prove that Nesterov's extrapolation is capable of making the individual convergence of projected gradient methods optimal for general convex problems, which is now a challenging problem in the machine learning community. In light of this consideration, a simple modification of the gradient operation suffices to achieve optimal individual convergence for strongly convex problems, which can be regarded as making an interesting step towards the open question about SGD posed by Shamir \cite{shamir2012open}. Furthermore, the derived algorithms are extended to solve regularized non-smooth learning problems in stochastic settings. {\color{blue}They can serve as an alternative to the most basic SGD especially in coping with machine learning problems, where an individual output is needed to guarantee the regularization structure while keeping an optimal rate of convergence.} Typically, our method is applicable as an efficient tool for solving large-scale $l_1$-regularized hinge-loss learning problems. Several real experiments demonstrate that the derived algorithms not only achieve optimal individual convergence rates but also guarantee better sparsity than the averaged solution.


Author(s):  
Ashok V. Kumar ◽  
David C. Gossard

Abstract A sequential approximation technique for non-linear programming is presented here that is particularly suited for problems in engineering design and structural optimization, where the number of variables are very large and function and sensitivity evaluations are computationally expensive. A sequence of sub-problems are iteratively generated using a linear approximation for the objective function and setting move limits on the variables using a barrier method. These sub-problems are strictly convex. Computation per iteration is significantly reduced by not solving the sub-problems exactly. Instead at each iteration, a few Newton-steps are taken for the sub-problem. A criteria for moving the move limit, is described that reduces or eliminates stepsize reduction during line search. The method was found to perform well for unconstrained and linearly constrained optimization problems. It requires very few function evaluations, does not require the hessian of the objective function and evaluates its gradient only once per iteration.


Author(s):  
Ehsan Kazemi ◽  
Liqiang Wang

Nonconvex and nonsmooth problems have recently attracted considerable attention in machine learning. However, developing efficient methods for the nonconvex and nonsmooth optimization problems with certain performance guarantee remains a challenge. Proximal coordinate descent (PCD) has been widely used for solving optimization problems, but the knowledge of PCD methods in the nonconvex setting is very limited. On the other hand, the asynchronous proximal coordinate descent (APCD) recently have received much attention in order to solve large-scale problems. However, the accelerated variants of APCD algorithms are rarely studied. In this paper, we extend APCD method to the accelerated algorithm (AAPCD) for nonsmooth and nonconvex problems that satisfies the sufficient descent property, by comparing between the function values at proximal update and a linear extrapolated point using a delay-aware momentum value. To the best of our knowledge, we are the first to provide stochastic and deterministic accelerated extension of APCD algorithms for general nonconvex and nonsmooth problems ensuring that for both bounded delays and unbounded delays every limit point is a critical point. By leveraging Kurdyka-Łojasiewicz property, we will show linear and sublinear convergence rates for the deterministic AAPCD with bounded delays. Numerical results demonstrate the practical efficiency of our algorithm in speed.


2020 ◽  
Vol 2020 ◽  
pp. 1-13
Author(s):  
Sha Lu ◽  
Zengxin Wei

Proximal point algorithm is a type of method widely used in solving optimization problems and some practical problems such as machine learning in recent years. In this paper, a framework of accelerated proximal point algorithm is presented for convex minimization with linear constraints. The algorithm can be seen as an extension to G u ¨ ler’s methods for unconstrained optimization and linear programming problems. We prove that the sequence generated by the algorithm converges to a KKT solution of the original problem under appropriate conditions with the convergence rate of O 1 / k 2 .


2015 ◽  
Vol 2015 ◽  
pp. 1-7
Author(s):  
Sakineh Tahmasebzadeh ◽  
Hamidreza Navidi ◽  
Alaeddin Malek

This paper proposes three numerical algorithms based on Karmarkar’s interior point technique for solving nonlinear convex programming problems subject to linear constraints. The first algorithm uses the Karmarkar idea and linearization of the objective function. The second and third algorithms are modification of the first algorithm using the Schrijver and Malek-Naseri approaches, respectively. These three novel schemes are tested against the algorithm of Kebiche-Keraghel-Yassine (KKY). It is shown that these three novel algorithms are more efficient and converge to the correct optimal solution, while the KKY algorithm fails in some cases. Numerical results are given to illustrate the performance of the proposed algorithms.


2018 ◽  
Vol 16 (05) ◽  
pp. 741-755 ◽  
Author(s):  
Qin Fang ◽  
Min Xu ◽  
Yiming Ying

The problem of minimizing a separable convex function under linearly coupled constraints arises from various application domains such as economic systems, distributed control, and network flow. The main challenge for solving this problem is that the size of data is very large, which makes usual gradient-based methods infeasible. Recently, Necoara, Nesterov and Glineur [Random block coordinate descent methods for linearly constrained optimization over networks, J. Optim. Theory Appl. 173(1) (2017) 227–254] proposed an efficient randomized coordinate descent method to solve this type of optimization problems and presented an appealing convergence analysis. In this paper, we develop new techniques to analyze the convergence of such algorithms, which are able to greatly improve the results presented in the above. This refined result is achieved by extending Nesterov’s second technique [Efficiency of coordinate descent methods on huge-scale optimization problems, SIAM J. Optim. 22 (2012) 341–362] to the general optimization problems with linearly coupled constraints. A novel technique in our analysis is to establish the basis vectors for the subspace of the linear constraints.


2020 ◽  
pp. 103-111
Author(s):  
Ольга Михайловна Прохорова

In the article we carried out a detailed analysis of the results obtained by J.-L. Lagrange in his first. The theory of extrema of functions of many variables, as part of mathematical analysis, refers to the mathematical foundations of the study of operations. In turn, many optimization problems are actually problems on the conditional extremum of the function of many variables. The relevance of this topic is determined by the fact that the methods for solving problems on the extremum of the function of many variables obtained in the mid 18th - early 20th centuries are used in solving modern problems. A special place here is occupied by L. Euler and J.-L. Lagrange. The aim of the article is to study the conditions for the maximum and minimum functions of many variables obtained by J.-L. Lagrange, and a comparison of its results with the presentation of this topic in modern textbooks on higher mathematics and mathematical analysis. It was established that in his first printed work he first formulated and proved sufficient conditions for the existence of an extremum of the function of many variables by actually establishing a criterion for the positive (negative) definiteness of quadratic forms, long before it appeared in J. Sylvester in the mid-19th century. A comparative analysis of the results of L. Euler and J.-L. Lagrange. It was found that sufficient conditions for the existence of an extremum of functions of many variables obtained in the first printed work of the young Lagrange are included in all modern textbooks in virtually the same form. The examples shown illustrate his theory. These are tasks of geometric and physical content. Special cases are considered in detail: functions of two and three variables. It is noted that this article became programmatic for the young Lagrange, although it remained unnoticed by his contemporaries. Subsequently, based on the method he obtained, he created the variational calculus, using the principle of least action and the theory of extrema, derived the basic laws of mechanics, the rule of factors for finding the conditional extremum of functions of many variables, which is named after him.


1999 ◽  
Vol 122 (3) ◽  
pp. 271-277 ◽  
Author(s):  
Ashok V. Kumar

A sequential approximation algorithm is presented here that is particularly suited for problems in engineering design and structural optimization, where the number of variables is very large and function and sensitivity evaluations are computationally expensive. A sequence of sub-problems are generated using a linear approximation for the objective function and setting move limits on the variables using a barrier method. These sub-problems are strictly convex and computation per iteration is significantly reduced by not solving the sub-problems exactly. Instead a few Newton-steps are taken for each sub-problem generated. A criterion, for setting the move limit, is described that reduces or eliminates step size reduction during line search. The method was found to perform well for unconstrained and linearly constrained optimization problems. It is particularly suitable for application to design of optimal shape and topology of structures by minimizing their compliance since it requires very few function evaluations, does not require the hessian of the objective function and evaluates its gradient only once for every sub-problem generated. [S1050-0472(00)01603-2]


Sign in / Sign up

Export Citation Format

Share Document