scholarly journals A Generic Approach for Accelerating Stochastic Zeroth-Order Convex Optimization

Author(s):  
Xiaotian Yu ◽  
Irwin King ◽  
Michael R. Lyu ◽  
Tianbao Yang

In this paper, we propose a generic approach for accelerating the convergence of existing algorithms to solve the problem of stochastic zeroth-order convex optimization (SZCO). Standard techniques for accelerating the convergence of stochastic zeroth-order algorithms are by exploring multiple functional evaluations (e.g., two-point evaluations), or by exploiting global conditions of the problem (e.g., smoothness and strong convexity). Nevertheless, these classic acceleration techniques are necessarily restricting the applicability of newly developed algorithms. The key of our proposed generic approach is to explore a local growth condition  (or called local error bound condition) of the objective function in SZCO. The benefits of the proposed acceleration technique are: (i) it is applicable to both settings with one-point evaluation and two-point evaluations; (ii) it does not necessarily require strong convexity or smoothness condition of the objective function; (iii) it yields an improvement on convergence for a broad family of problems. Empirical studies in various settings demonstrate the effectiveness of the proposed acceleration approach.

2005 ◽  
Vol 14 (01n02) ◽  
pp. 281-301 ◽  
Author(s):  
R. G. GORE ◽  
J. LI ◽  
M. T. MANRY ◽  
L. M. LIU ◽  
C. YU ◽  
...  

A method which modifies the objective function used far designing neural network classifiers is presented. The classical mean-square error criteria is relaxed by introducing two types of local error bias which are treated like free parameters. Open and closed form solutions are given for finding these bias parameters. The new objective function is seamlessly integrated into existing training algorithms such as back propagation (BP), output weight optimization (OWO), and hidden weight optimization (HWO). The resulting algorithms are successfully applied in training neural net classifiers having a linear final layer. Classifiers are trained and tested on several data sets from pattern recognition applications. Improvement over classical iterative regression methods is clearly demonstrated.


2015 ◽  
Vol 2015 ◽  
pp. 1-7
Author(s):  
Sakineh Tahmasebzadeh ◽  
Hamidreza Navidi ◽  
Alaeddin Malek

This paper proposes three numerical algorithms based on Karmarkar’s interior point technique for solving nonlinear convex programming problems subject to linear constraints. The first algorithm uses the Karmarkar idea and linearization of the objective function. The second and third algorithms are modification of the first algorithm using the Schrijver and Malek-Naseri approaches, respectively. These three novel schemes are tested against the algorithm of Kebiche-Keraghel-Yassine (KKY). It is shown that these three novel algorithms are more efficient and converge to the correct optimal solution, while the KKY algorithm fails in some cases. Numerical results are given to illustrate the performance of the proposed algorithms.


Author(s):  
Xiaoxiao Ma ◽  
Xiaojuan Chen

Because the traditional method of solving nonlinear equations takes a long time, an optimal path analysis method for solving nonlinear equations with limited local error is designed. Firstly, according to the finite condition of local error, the optimization objective function of nonlinear equations is established. Secondly, set the constraints of the objective function, solve the optimal solution of the nonlinear equation under the condition of limited local error, and obtain the optimal path of the nonlinear equation system. Finally, experiments show that the optimal path analysis method for solving nonlinear equations with limited local error takes less time than other methods, and can be effectively applied to practice


2020 ◽  
Vol 54 (3) ◽  
pp. 873-882
Author(s):  
Pedro Henrique González Silva ◽  
Ana Flávia U. S. Macambira ◽  
Renan Vicente Pinto ◽  
Luidi Simonetti ◽  
Nelson Maculan ◽  
...  

Given a solid T, represented by a compact set in ℝ3, the aim of this work is to find a covering of T by a finite set of spheres of different radii. Some level of intersection between the spheres is necessary to cover the solid. Moreover, the volume occupied by the spheres on the outside of T is limited. This problem has an application in the planning of a radio-surgery treatment known by Gamma Knife and can be formulated as a non-convex optimization problem with quadratic constraints and linear objective function. In this work, two new linear mathematical formulations with binary variables and a hybrid method are proposed. The hybrid method combines heuristic, data mining and an exact method. Computational results show that the proposed methods outperform the ones presented in the literature.


Author(s):  
Yurii Nesterov

AbstractIn this paper we develop new tensor methods for unconstrained convex optimization, which solve at each iteration an auxiliary problem of minimizing convex multivariate polynomial. We analyze the simplest scheme, based on minimization of a regularized local model of the objective function, and its accelerated version obtained in the framework of estimating sequences. Their rates of convergence are compared with the worst-case lower complexity bounds for corresponding problem classes. Finally, for the third-order methods, we suggest an efficient technique for solving the auxiliary problem, which is based on the recently developed relative smoothness condition (Bauschke et al. in Math Oper Res 42:330–348, 2017; Lu et al. in SIOPT 28(1):333–354, 2018). With this elaboration, the third-order methods become implementable and very fast. The rate of convergence in terms of the function value for the accelerated third-order scheme reaches the level $$O\left( {1 \over k^4}\right) $$O1k4, where k is the number of iterations. This is very close to the lower bound of the order $$O\left( {1 \over k^5}\right) $$O1k5, which is also justified in this paper. At the same time, in many important cases the computational cost of one iteration of this method remains on the level typical for the second-order methods.


Author(s):  
Afrooz Jalilzadeh ◽  
Angelia Nedić ◽  
Uday V. Shanbhag ◽  
Farzad Yousefian

Classical theory for quasi-Newton schemes has focused on smooth, deterministic, unconstrained optimization, whereas recent forays into stochastic convex optimization have largely resided in smooth, unconstrained, and strongly convex regimes. Naturally, there is a compelling need to address nonsmoothness, the lack of strong convexity, and the presence of constraints. Accordingly, this paper presents a quasi-Newton framework that can process merely convex and possibly nonsmooth (but smoothable) stochastic convex problems. We propose a framework that combines iterative smoothing and regularization with a variance-reduced scheme reliant on using an increasing sample size of gradients. We make the following contributions. (i) We develop a regularized and smoothed variable sample-size BFGS update (rsL-BFGS) that generates a sequence of Hessian approximations and can accommodate nonsmooth convex objectives by utilizing iterative regularization and smoothing. (ii) In strongly convex regimes with state-dependent noise, the proposed variable sample-size stochastic quasi-Newton (VS-SQN) scheme admits a nonasymptotic linear rate of convergence, whereas the oracle complexity of computing an [Formula: see text]-solution is [Formula: see text], where [Formula: see text] denotes the condition number and [Formula: see text]. In nonsmooth (but smoothable) regimes, using Moreau smoothing retains the linear convergence rate for the resulting smoothed VS-SQN (or sVS-SQN) scheme. Notably, the nonsmooth regime allows for accommodating convex constraints. To contend with the possible unavailability of Lipschitzian and strong convexity parameters, we also provide sublinear rates for diminishing step-length variants that do not rely on the knowledge of such parameters. (iii) In merely convex but smooth settings, the regularized VS-SQN scheme rVS-SQN displays a rate of [Formula: see text] with an oracle complexity of [Formula: see text]. When the smoothness requirements are weakened, the rate for the regularized and smoothed VS-SQN scheme rsVS-SQN worsens to [Formula: see text]. Such statements allow for a state-dependent noise assumption under a quadratic growth property on the objective. To the best of our knowledge, the rate results are among the first available rates for QN methods in nonsmooth regimes. Preliminary numerical evidence suggests that the schemes compare well with accelerated gradient counterparts on selected problems in stochastic optimization and machine learning with significant benefits in ill-conditioned regimes.


2013 ◽  
Vol 2013 ◽  
pp. 1-7 ◽  
Author(s):  
Jie Shen ◽  
Li-Ping Pang ◽  
Dan Li

An implementable algorithm for solving a nonsmooth convex optimization problem is proposed by combining Moreau-Yosida regularization and bundle and quasi-Newton ideas. In contrast with quasi-Newton bundle methods of Mifflin et al. (1998), we only assume that the values of the objective function and its subgradients are evaluated approximately, which makes the method easier to implement. Under some reasonable assumptions, the proposed method is shown to have a Q-superlinear rate of convergence.


2003 ◽  
Vol 9 (5) ◽  
pp. 353-361 ◽  
Author(s):  
Eun Seok Lee ◽  
George S. Dulikravich ◽  
Brian H. Dennis

An axial turbine rotor cascade-shape optimization with unsteady passing wakes was performed to obtain an improved aerodynamic performance using an unsteady flow, Reynolds-averaged Navier-Stokes equations solver that was based on explicit, finite difference; Runge-Kutta multistage time marching; and the diagonalized alternating direction implicit scheme. The code utilized Baldwin-Lomax algebraic andk-εturbulence modeling. The full approximation storage multigrid method and preconditioning were implemented as iterative convergence-acceleration techniques. An implicit dual-time stepping method was incorporated in order to simulate the unsteady flow fields. The objective function was defined as minimization of total pressure loss and maximization of lift, while the mass flow rate was fixed during the optimization. The design variables were several geometric parameters characterizing airfoil leading edge, camber, stagger angle, and inter-row spacing. The genetic algorithm was used as an optimizer, and the penalty method was introduced for combining the constraints with the objective function. Each individual's objective function was computed simultaneously by using a 32-processor distributedmemory computer. The optimization results indicated that only minor improvements are possible in unsteady rotor/stator aerodynamics by varying these geometric parameters.


Metals ◽  
2020 ◽  
Vol 10 (6) ◽  
pp. 781
Author(s):  
Jelena Srnec Novak ◽  
Francesco De Bona ◽  
Denis Benasciutti

In numerical simulations of components subjected to low-cycle fatigue loading, the material cyclic plasticity behavior must be modelled until complete stabilization, which occurs approximately at half the number of cycles to failure. If the plastic strain per cycle is small, a huge number of cycles must be simulated, which results into a huge and thus unaffordable simulation time. Acceleration techniques for shortening this time are useful, although their accuracy needs to be checked. This work aims to compare different approaches (nonlinear kinematic with “initial” and “stabilized” parameters and combined nonlinear kinematic and isotropic with the speed of stabilization fictitiously increased). It considers two benchmarks taken from the literature, in which the material has opposite cyclic behaviors (hardening, softening). A plane finite element model can be used in both benchmarks, thus permitting a simulation up to complete stabilization. Results confirm that the common approach of considering only the kinematic model (calibrated on “initial” or “stabilized” material state) from the very first cycle could lead to relevant errors. The acceleration technique based on a fictitious increase in the speed of stabilization leads to accurate results. Guidelines for calibrating this technique on a material’s hardening or softening behavior are, finally, proposed.


Sign in / Sign up

Export Citation Format

Share Document