Loading [MathJax]/extensions/MathMenu.js
Fast stochastic motion planning with optimality guarantees using local policy reconfiguration | IEEE Conference Publication | IEEE Xplore

Fast stochastic motion planning with optimality guarantees using local policy reconfiguration


Abstract:

This work presents a framework for fast reconfiguration of local control policies for a stochastic system to satisfy a high-level task specification. The motion of the sy...Show More

Abstract:

This work presents a framework for fast reconfiguration of local control policies for a stochastic system to satisfy a high-level task specification. The motion of the system is abstracted to a class of uncertain Markov models known as bounded-parameter Markov decision processes (BMDPs). During the abstraction, an efficient sampling-based method for stochastic optimal control is used to construct several policies within a discrete region of the state space in order for the system to transit between neighboring regions. A BMDP is then used to find an optimal strategy over the local policies by maximizing a continuous reward function; a new policy can be computed quickly if the reward function changes. The efficacy of the framework is demonstrated using a sequence of online tasks, showing that highly desirable policies can be obtained by reconfiguring existing local policies in just a few seconds.
Date of Conference: 31 May 2014 - 07 June 2014
Date Added to IEEE Xplore: 29 September 2014
Electronic ISBN:978-1-4799-3685-4
Print ISSN: 1050-4729
Conference Location: Hong Kong, China
References is not available for this document.

I. Introduction

The objective for traditional robotic motion planning is to compute a trajectory that will move the robot between two valid poses while respecting all physical constraints [1] – [3]. In practice, however, robots suffer from unexpected events like noisy actuation of a wheeled base when navigating uneven terrain, imperfect observations due to unfavorable sensing conditions, or an incorrect map if objects have moved around a room. These kinds of disturbances force the robot to deviate from its current course into a state from which there may be no clear path to the goal. Replanning is one option to address these detours, but this can be computationally prohibitive if the deviations are frequent or constant. A more robust strategy to combat motion planning under uncertainty is to model the problem as a stochastic decision process where the solution is not a single trajectory, but rather a control policy over all possible states of the system to maximize an objective function [3], [4].

Select All
1.
J.-C. Latombe, Robot Motion Planning, Boston, MA:Kluwer Academic Publishers, 1991.
2.
H. Choset, K. M. Lynch, S. Hutchinson, G. Kantor, W. Burgard, L. E. Kavraki, et al., Principles of Robot Motion: Theory Algorithms and Implementations, MIT Press, 2005.
3.
S. M. LaValle, Planning Algorithms, Cambridge University Press, 2006.
4.
S. Thrun, W. Burgard and D. Fox, Probabilistic Robotics, Cambridge, MA:MIT press, 2005.
5.
V. D. Blondel and J. N. Tsitsiklis, "A survey of computational complexity results in systems and control", Automatica, vol. 36, no. 9, pp. 1249-1274, 2000.
6.
T. Dean, L. P. Kaelbling, J. Kirman and A. Nicholson, "Planning under time constraints in stochastic domains", Artificial Intelligence, vol. 76, no. 1-2, pp. 35-74, 1995.
7.
J. Burlet, O. Aycard and T. Fraichard, "Robust motion planning using Markov decision processes and quadtree decomposition", IEEE Int'l. Conference on Robotics and Automation, vol. 3, pp. 2820-2825, 2004.
8.
R. Alterovitz, T. Siméon and K. Y. Goldberg, "The stochastic motion roadmap: a sampling framework for planning with markov motion uncertainty", Robotics: Science and Systems 2007.
9.
S. C. W. Ong, S. W. Png, D. Hsu and W. S. Lee, "Planning under uncertainty for robotic tasks with mixed observability", Int'l Journal of Robotics Research, vol. 29, no. 8, pp. 1053-1068, July 2010.
10.
V. A. Huynh, S. Karaman and E. Frazzoli, "An incremental sampling-based algorithm for stochastic optimal control", IEEE Int'l. Conference on Robotics and Automation, pp. 2865-2872, May 2012.
11.
D. Hsu, J.-C. Latombe and R. Motwani, "Path planning in expansive configuration spaces", Int'l Journal of Computational Geometry and Applications, vol. 9, no. 4-5, pp. 495-512, 1999.
12.
S. M. LaValle and J. J. Kuffner, "Randomized kinodynamic planning", Intl. J. of Robotics Research, vol. 20, no. 5, pp. 378-400, May 2001.
13.
I. A. Sucan and L. E. Kavraki, "A sampling-based tree planner for systems with complex dynamics", IEEE Trans. on Robotics, vol. 28, no. 1, pp. 116-131, 2012.
14.
R. Tedrake, I. R. Manchester, M. Tobenkin and J. W. Roberts, "LQR-trees: Feedback motion planning via sums-of-squares verification", Int'l Journal of Robotics Research, vol. 29, no. 8, pp. 1038-1052, July 2010.
15.
S. Chakravorty and S. Kumar, "Generalized sampling-based motion planners", IEEE Transactions on Systems Man and Cybernetics Part B: Cybernetics, vol. 41, no. 3, pp. 855-866, June 2011.
16.
R. Givan, S. Leach and T. Dean, "Bounded-parameter Markov decision processes", Artificial Intelligence, vol. 122, no. 1-2, pp. 71-109, 2000.
17.
D. Wu and X. Koutsoukos, "Reachability analysis of uncertain systems using bounded-parameter Markov decision processes", Artificial Intelligence, vol. 172, no. 8-9, pp. 945-954, 2008.
18.
H. J. Kushner and P. Dupuis, Numerical methods for stochastic control problems in continuous time, Springer, vol. 24, 2001.
19.
J. R. Shewchuk, "Delaunay refinement algorithms for triangular mesh generation", Comp Geometry, vol. 22, no. 1-3, pp. 21-74, 2002.
20.
J. G. Kemeny and J. L. Snell, Finite Markov Chains, Springer-Verlag, 1976.
21.
I. A. Sucan, M. Moll and L. E. Kavraki, "The Open Motion Planning Library", IEEE Robotics & Automation Magazine, vol. 19, no. 4, pp. 72-82, Dec. 2012.

Contact IEEE to Subscribe

References

References is not available for this document.