Dynamic programming and the principle of optimality: A systematic approach

1978 ◽  
Vol 1 (4) ◽  
pp. 183-190 ◽  
Author(s):  
Moshe Sniedovich
Author(s):  
R. Giancarlo

In this Chapter we present some general algorithmic techniques that have proved to be useful in speeding up the computation of some families of dynamic programming recurrences which have applications in sequence alignment, paragraph formation and prediction of RNA secondary structure. The material presented in this chapter is related to the computation of Levenshtein distances and approximate string matching that have been discussed in the previous three chapters. Dynamic programming is a general technique for solving discrete optimization (minimization or maximization) problems that can be represented by decision processes and for which the principle of optimality holds. We can view a decision process as a directed graph in which nodes represent the states of the process and edges represent decisions. The optimization problem at hand is represented as a decision process by decomposing it into a set of subproblems of smaller size. Such recursive decomposition is continued until we get only trivial subproblems, which can be solved directly. Each node in the graph corresponds to a subproblem and each edge (a, b) indicates that one way to solve subproblem a optimally is to solve first subproblem b optimally. Then, an optimal solution, or policy, is typically given by a path on the graph that minimizes or maximizes some objective function. The correctness of this approach is guaranteed by the principle of optimality which must be satisfied by the optimization problem: An optimal policy has the property that whatever the initial node (state) and initial edge (decision) are, the remaining edges (decisions) must be an optimal policy with regard to the node (state) resulting from the first transition. Another consequence of the principle of optimality is that we can express the optimal cost (and solution) of a subproblem in terms of optimal costs (and solutions) of problems of smaller size. That is, we can express optimal costs through a recurrence relation. This is a key component of dynamic programming, since we can compute the optimal cost of a subproblem only once, store the result in a table, and look it up when needed.


1994 ◽  
Vol 4 (1) ◽  
pp. 33-69 ◽  
Author(s):  
Oege De Moor

Dynamic programming is a strategy for solving optimisation problems. In this paper, we show how many problems that may be solved by dynamic programming are instances of the same abstract specification. This specification is phrased using the calculus of relations offered by topos theory. The main theorem underlying dynamic programming can then be proved by straightforward equational reasoning.The generic specification of dynamic programming makes use of higher-order operators on relations, akin to the fold operators found in functional programming languages. In the present context, a data type is modelled as an initial F-algebra, where F is an endofunctor on the topos under consideration. The mediating arrows from this initial F-algebra to other F-algebras are instances of fold – but only for total functions. For a regular category ε, it is possible to construct a category of relations Rel(ε). When a functor between regular categories is a so-called relator, it can be extended (in some canonical way) to a functor between the corresponding categories of relations. Applied to an endofunctor on a topos, this process of extending functors preserves initial algebras, and hence fold can be generalised from functions to relations.It is well-known that the use of dynamic programming is governed by the principle of optimality. Roughly, the principle of optimality says that an optimal solution is composed of optimal solutions to subproblems. In a first attempt, we formalise the principle of optimality as a distributivity condition. This distributivity condition is elegant, but difficult to check in practice. The difficulty arises because we consider minimum elements with respect to a preorder, and therefore minimum elements are not unique. Assuming that we are working in a Boolean topos, it can be proved that monotonicity implies distributivity, and this monotonicity condition is easy to verify in practice.


1988 ◽  
Vol 18 (9) ◽  
pp. 1118-1122 ◽  
Author(s):  
Greg J. Arthaud ◽  
W. David Klemperer

A four state-descriptor dynamic programming model was used to seek economically optimal thinning regimes for high and low thinning in loblolly pine (Pinustaeda L.). Given the assumptions of the study, low thinning generated higher present values than high thinning. Growth equations were estimated from data generated by a stochastic growth simulator. Reasons for the occurrence of slightly suboptimal solutions are discussed. The paper reviews situations in which the principle of optimality might be violated when thinning problems are being solved with forward recursion dynamic programming.


Author(s):  
Xingyong Song ◽  
Mohd Azrin Mohd Zulkefli ◽  
Zongxuan Sun ◽  
Hsu-Chiang Miao

Clutch fill control is critical for automotive transmission performance and fuel economy, including both automatic and hybrid transmissions. The traditional approach, by which the clutch fill pressure command is manually calibrated, has a couple of limitations. First, the pressure profile is not optimized to reduce the peak clutch fill flow demand. Moreover, it is not systematically designed to account for uncertainties in the system, such as variations of solenoid valve time delay and parameters of the clutch assembly. In this paper, we present a systematic approach to evaluate the clutch fill dynamics and synthesize the optimal pressure profile. First, a clutch fill dynamic model is constructed and analyzed. Second, the applicability of the conventional numerical Dynamic Programming (DP) algorithm to the clutch fill control problem is explored and shown to be ineffective. Thus we developed a new customized DP method to obtain the optimal and robust pressure profile subject to specified constraints. After a series of simulations and case studies, the new customized DP approach is demonstrated to be effective, efficient, and robust for solving the clutch fill optimal control problem.


1966 ◽  
Vol 70 (664) ◽  
pp. 469-476 ◽  
Author(s):  
P. Dyer ◽  
A. R. M. Noton ◽  
D. Rutherford

SummaryIn the past decade considerable attention has been given to the possibility of using self-adaptive control systems to cope with the problem of providing stability augmentation to aircraft over extended flight envelopes. On the other hand, it has sometimes been considered that insufficient attention has been given to the design of autostabilisers with fixed parameters. The authors have applied recent developments In dynamic programming to the design of aircraft autostabilisers. The methods provide a highly systematic approach to the design of multi-loop systems which is especially useful for control of the lateral mode. Furthermore, the resulting optimum multiple feedback coefficients vary little throughout the flight envelope and the use of fixed parameters for the whole of the flight envelope appears to be possible. Aerodynamic data for two aircraft were used: a supersonic VTOL aircraft and a supersonic transport.


1987 ◽  
Vol 10 (3) ◽  
pp. 597-607 ◽  
Author(s):  
J. N. Kapur ◽  
Vinod Kumar ◽  
Uma Kumar

The principle of optimality of dynamic programming is used to prove three major inequalities due to Shannon, Renyi and Holder. The inequalities are then used to obtain some useful results in information theory. In particular measures are obtained to measure the mutual divergence among two or more probability distributions.


Sign in / Sign up

Export Citation Format

Share Document