Iterative local dynamic programming
- 1 March 2009
- conference paper
- conference paper
- Published by Institute of Electrical and Electronics Engineers (IEEE)
Abstract
We develop an iterative local dynamic programming method (iLDP) applicable to stochastic optimal control problems in continuous high-dimensional state and action spaces. Such problems are common in the control of biological movement, but cannot be handled by existing methods. iLDP can be considered a generalization of differential dynamic programming, in as much as: (a) we use general basis functions rather than quadratics to approximate the optimal value function; (b) we introduce a collocation method that dispenses with explicit differentiation of the cost and dynamics and ties iLDP to the unscented Kalman filter; (c) we adapt the local function approximator to the propagated state covariance, thus increasing accuracy at more likely states. Convergence is similar to quasi-Newton methods. We illustrate iLDP on several problems including the ldquoswimmerrdquo dynamical system which has 14 state and 4 control variables.Keywords
This publication has 3 references indexed in Scilit:
- Optimal feedback control as a theory of motor coordinationNature Neuroscience, 2002
- A new method for the nonlinear transformation of means and covariances in filters and estimatorsIEEE Transactions on Automatic Control, 2000
- Differential dynamic programming and Newton's methodInternational Journal of Control, 1988