Nonlinear Optimal Design of Dynamic Mechanical Systems

Author(s):  
T. E. Potter ◽  
K. D. Willmert ◽  
M. Sathyamoorthy

Abstract Mechanism path generation problems which use link deformations to improve the design lead to optimization problems involving a nonlinear sum-of-squares objective function subjected to a set of linear and nonlinear constraints. Inclusion of the deformation analysis causes the objective function evaluation to be computationally expensive. An optimization method is presented which requires relatively few objective function evaluations. The algorithm, based on the Gauss method for unconstrained problems, is developed as an extension of the Gauss constrained technique for linear constraints and revises the Gauss nonlinearly constrained method for quadratic constraints. The derivation of the algorithm, using a Lagrange multiplier approach, is based on the Kuhn-Tucker conditions so that when the iteration process terminates, these conditions are automatically satisfied. Although the technique was developed for mechanism problems, it is applicable to any optimization problem having the form of a sum of squares objective function subjected to nonlinear constraints.

Author(s):  
Ion Necoara ◽  
Martin Takáč

Abstract In this paper we consider large-scale smooth optimization problems with multiple linear coupled constraints. Due to the non-separability of the constraints, arbitrary random sketching would not be guaranteed to work. Thus, we first investigate necessary and sufficient conditions for the sketch sampling to have well-defined algorithms. Based on these sampling conditions we develop new sketch descent methods for solving general smooth linearly constrained problems, in particular, random sketch descent (RSD) and accelerated random sketch descent (A-RSD) methods. To our knowledge, this is the first convergence analysis of RSD algorithms for optimization problems with multiple non-separable linear constraints. For the general case, when the objective function is smooth and non-convex, we prove for the non-accelerated variant sublinear rate in expectation for an appropriate optimality measure. In the smooth convex case, we derive for both algorithms, non-accelerated and A-RSD, sublinear convergence rates in the expected values of the objective function. Additionally, if the objective function satisfies a strong convexity type condition, both algorithms converge linearly in expectation. In special cases, where complexity bounds are known for some particular sketching algorithms, such as coordinate descent methods for optimization problems with a single linear coupled constraint, our theory recovers the best known bounds. Finally, we present several numerical examples to illustrate the performances of our new algorithms.


2015 ◽  
Vol 2015 ◽  
pp. 1-7
Author(s):  
Sakineh Tahmasebzadeh ◽  
Hamidreza Navidi ◽  
Alaeddin Malek

This paper proposes three numerical algorithms based on Karmarkar’s interior point technique for solving nonlinear convex programming problems subject to linear constraints. The first algorithm uses the Karmarkar idea and linearization of the objective function. The second and third algorithms are modification of the first algorithm using the Schrijver and Malek-Naseri approaches, respectively. These three novel schemes are tested against the algorithm of Kebiche-Keraghel-Yassine (KKY). It is shown that these three novel algorithms are more efficient and converge to the correct optimal solution, while the KKY algorithm fails in some cases. Numerical results are given to illustrate the performance of the proposed algorithms.


2021 ◽  
Author(s):  
Yixuan Wang ◽  
Faruk Alpak ◽  
Guohua Gao ◽  
Chaohui Chen ◽  
Jeroen Vink ◽  
...  

Abstract Although it is possible to apply traditional optimization algorithms to determine the Pareto front of a multi-objective optimization problem, the computational cost is extremely high, when the objective function evaluation requires solving a complex reservoir simulation problem and optimization cannot benefit from adjoint-based gradients. This paper proposes a novel workflow to solve bi-objective optimization problems using the distributed quasi-Newton (DQN) method, which is a well-parallelized and derivative-free optimization (DFO) method. Numerical tests confirm that the DQN method performs efficiently and robustly. The efficiency of the DQN optimizer stems from a distributed computing mechanism which effectively shares the available information discovered in prior iterations. Rather than performing multiple quasi-Newton optimization tasks in isolation, simulation results are shared among distinct DQN optimization tasks or threads. In this paper, the DQN method is applied to the optimization of a weighted average of two objectives, using different weighting factors for different optimization threads. In each iteration, the DQN optimizer generates an ensemble of search points (or simulation cases) in parallel and a set of non-dominated points is updated accordingly. Different DQN optimization threads, which use the same set of simulation results but different weighting factors in their objective functions, converge to different optima of the weighted average objective function. The non-dominated points found in the last iteration form a set of Pareto optimal solutions. Robustness as well as efficiency of the DQN optimizer originates from reliance on a large, shared set of intermediate search points. On the one hand, this set of searching points is (much) smaller than the combined sets needed if all optimizations with different weighting factors would be executed separately; on the other hand, the size of this set produces a high fault tolerance. Even if some simulations fail at a given iteration, DQN’s distributed-parallel information-sharing protocol is designed and implemented such that the optimization process can still proceed to the next iteration. The proposed DQN optimization method is first validated on synthetic examples with analytical objective functions. Then, it is tested on well location optimization problems, by maximizing the oil production and minimizing the water production. Furthermore, the proposed method is benchmarked against a bi-objective implementation of the MADS (Mesh Adaptive Direct Search) method, and the numerical results reinforce the auspicious computational attributes of DQN observed for the test problems. To the best of our knowledge, this is the first time that a well-parallelized and derivative-free DQN optimization method has been developed and tested on bi-objective optimization problems. The methodology proposed can help improve efficiency and robustness in solving complicated bi-objective optimization problems by taking advantage of model-based search optimization algorithms with an effective information-sharing mechanism.


Author(s):  
Busaba Phruksaphanrat ◽  
◽  
Ario Ohsato ◽  

This research proposes an effective linear calculational method based on convex cone concept for solving non-linear optimization problems with a convex polyhedral objective function and linear constraints. One familiar type of convex polyhedral objective functions is a triangular type, which can be normally solved by weighted goal programming (WGP). The necessary preference information of WGP is weights of positive and negative deviational variables. Alternatively, the linear calculational method prefers the other operational way in designing of an objective function by deviational constants, which is practical for the decision maker. For convex polyhedral type objective function problems, conventionally separable convex programming and goal programming (GP) are applied. By separable convex programming, it needs to separate the objective function into line segments before solving the problem, which means increasing of variables and constraints. In case of GP each breakpoint is determined as a goal so the number of constraints and the deviational variables are drastically increased. By the effective linear calculational method proposed in this paper, the problem could be simply formulated to the linear programming problem, which is easy for the decision maker to apply. Moreover this method has lower number of constraints and variables than existing methods so the calculational time can also be reduced.


2018 ◽  
Vol 2018 ◽  
pp. 1-27 ◽  
Author(s):  
Shuangqing Chen ◽  
Yang Liu ◽  
Lixin Wei ◽  
Bing Guan

Particle swarm optimization (PSO) and fireworks algorithm (FWA) are two recently developed optimization methods which have been applied in various areas due to their simplicity and efficiency. However, when being applied to high-dimensional optimization problems, PSO algorithm may be trapped in the local optima owing to the lack of powerful global exploration capability, and fireworks algorithm is difficult to converge in some cases because of its relatively low local exploitation efficiency for noncore fireworks. In this paper, a hybrid algorithm called PS-FW is presented, in which the modified operators of FWA are embedded into the solving process of PSO. In the iteration process, the abandonment and supplement mechanism is adopted to balance the exploration and exploitation ability of PS-FW, and the modified explosion operator and the novel mutation operator are proposed to speed up the global convergence and to avoid prematurity. To verify the performance of the proposed PS-FW algorithm, 22 high-dimensional benchmark functions have been employed, and it is compared with PSO, FWA, stdPSO, CPSO, CLPSO, FIPS, Frankenstein, and ALWPSO algorithms. Results show that the PS-FW algorithm is an efficient, robust, and fast converging optimization method for solving global optimization problems.


2020 ◽  
Vol 2020 ◽  
pp. 1-9
Author(s):  
Yufeng Yao ◽  
Yan Ding ◽  
Shan Zhong ◽  
Zhiming Cui ◽  
Chenxi Huang

In the field of brain-computer interfaces, it is very common to use EEG signals for disease diagnosis. In this study, a style regularized least squares support vector machine based on multikernel learning is proposed and applied to the recognition of epilepsy abnormal signals. The algorithm uses the style conversion matrix to represent the style information contained in the sample, regularizes it in the objective function, optimizes the objective function through the commonly used alternative optimization method, and simultaneously updates the style conversion matrix and classifier during the iteration process parameter. In order to use the learned style information in the prediction process, two new rules are added to the traditional prediction method, and the style conversion matrix is used to standardize the sample style before classification.


2000 ◽  
Vol 122 (2) ◽  
pp. 155-163 ◽  
Author(s):  
Wei Chen ◽  
Atul Sahai ◽  
Achille Messac ◽  
Glynn J. Sundararaj

Computational optimization for design is effective only to the extent that the aggregate objective function adequately captures designer’s preference. Physical programming is an optimization method that captures the designer’s physical understanding of the desired design outcome in forming the aggregate objective function. Furthermore, to be useful, a resulting optimal design must be sufficiently robust/insensitive to known and unknown variations that to different degrees affect the design’s performance. This paper explores the effectiveness of the physical programming approach in explicitly addressing the issue of design robustness. Specifically, we synergistically integrate methods that had previously and independently been developed by the authors, thereby leading to optimal—robust—designs. We show how the physical programming method can be used to effectively exploit designer preference in making tradeoffs between the mean and variation of performance, by solving a bi-objective robust design problem. The work documented in this paper establishes the general superiority of physical programming over other conventional methods (e.g., weighted sum) in solving multiobjective optimization problems. It also illustrates that the physical programming method is among the most effective multicriteria mathematical programming techniques for the generation of Pareto solutions that belong to both convex and non-convex efficient frontiers. [S1050-0472(00)00902-8]


Author(s):  
Pengfei (Taylor) Li ◽  
Peirong (Slade) Wang ◽  
Farzana Chowdhury ◽  
Li Zhang

Traditional formulations for transportation optimization problems mostly build complicating attributes into constraints while keeping the succinctness of objective functions. A popular solution is the Lagrangian decomposition by relaxing complicating constraints and then solving iteratively. Although this approach is effective for many problems, it generates intractability in other problems. To address this issue, this paper presents an alternative formulation for transportation optimization problems in which the complicating attributes of target problems are partially or entirely built into the objective function instead of into the constraints. Many mathematical complicating constraints in transportation problems can be efficiently modeled in dynamic network loading (DNL) models based on the demand–supply equilibrium, such as the various road or vehicle capacity constraints or “IF–THEN” type constraints. After “pre-building” complicating constraints into the objective functions, the objective function can be approximated well with customized high-fidelity DNL models. Three types of computing benefits can be achieved in the alternative formulation: ( a) the original problem will be kept the same; ( b) computing complexity of the new formulation may be significantly reduced because of the disappearance of hard constraints; ( c) efficiency loss on the objective function side can be mitigated via multiple high-performance computing techniques. Under this new framework, high-fidelity and problem-specific DNL models will be critical to maintain the attributes of original problems. Therefore, the authors’ recent efforts in enhancing the DNL’s fidelity and computing efficiency are also described in the second part of this paper. Finally, a demonstration case study is conducted to validate the new approach.


Electronics ◽  
2021 ◽  
Vol 10 (12) ◽  
pp. 1452
Author(s):  
Cristian Mateo Castiblanco-Pérez ◽  
David Esteban Toro-Rodríguez ◽  
Oscar Danilo Montoya ◽  
Diego Armando Giral-Ramírez

In this paper, we propose a new discrete-continuous codification of the Chu–Beasley genetic algorithm to address the optimal placement and sizing problem of the distribution static compensators (D-STATCOM) in electrical distribution grids. The discrete part of the codification determines the nodes where D-STATCOM will be installed. The continuous part of the codification regulates their sizes. The objective function considered in this study is the minimization of the annual operative costs regarding energy losses and installation investments in D-STATCOM. This objective function is subject to the classical power balance constraints and devices’ capabilities. The proposed discrete-continuous version of the genetic algorithm solves the mixed-integer non-linear programming model that the classical power balance generates. Numerical validations in the 33 test feeder with radial and meshed configurations show that the proposed approach effectively minimizes the annual operating costs of the grid. In addition, the GAMS software compares the results of the proposed optimization method, which allows demonstrating its efficiency and robustness.


Sign in / Sign up

Export Citation Format

Share Document