Loading [MathJax]/extensions/MathMenu.js
Policy Optimization to Learn Adaptive Motion Primitives in Path Planning With Dynamic Obstacles | IEEE Journals & Magazine | IEEE Xplore

Policy Optimization to Learn Adaptive Motion Primitives in Path Planning With Dynamic Obstacles


Abstract:

This letter addresses the kinodynamic motion planning for non-holonomic robots in dynamic environments with both static and dynamic obstacles – a challenging problem that...Show More

Abstract:

This letter addresses the kinodynamic motion planning for non-holonomic robots in dynamic environments with both static and dynamic obstacles – a challenging problem that lacks a universal solution yet. One of the promising approaches to solve it is decomposing the problem into the smaller sub-problems and combining the local solutions into the global one. The crux of any planning method for non-holonomic robots is the generation of motion primitives that generates solutions to local planning sub-problems. In this work we introduce a novel learnable steering function (policy), which takes into account kinodynamic constraints of the robot and both static and dynamic obstacles. This policy is efficiently trained via the policy optimization. Empirically, we show that our steering function generalizes well to unseen problems. We then plug in the trained policy into the sampling-based and lattice-based planners, and evaluate the resultant POLAMP algorithm (Policy Optimization that Learns Adaptive Motion Primitives) in a range of challenging setups that involve a car-like robot operating in the obstacle-rich parking-lot environments. We show that POLAMP is able to plan collision-free kinodynamic trajectories with success rates higher than 92%, when 50 simultaneously moving obstacles populate the environment showing better performance than the state-of-the-art competitors.
Published in: IEEE Robotics and Automation Letters ( Volume: 8, Issue: 2, February 2023)
Page(s): 824 - 831
Date of Publication: 30 December 2022

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

Autonomous robotic systems have become one of the most popular research topics in recent years due to their pronounced potential social benefits. In particular, autonomous driving is developing rapidly and at the same time requires efficient motion planning in complex and highly dynamic environments, meanwhile taking into account the kinodynamic constraints of an non-holonomic autonomous vehicle. Often, the planners that address the first aspect of the problem, i.e. dynamic environment, like the ones presented in [1], [2], [3] do not take into account the kinodynamic constraints. On the other hand, kinodynamic planners often do not explicitly reason about the future changes of the environments, even if these changes are for-seen, e.g. predicted by the control system of the robot. In this work we want to enrich the kinodynamic planning methods with the ability to take the dynamics of the environment as well (at the planning stage).

Select All
1.
M. Otte and E. Frazzoli, "RRTX: Asymptotically optimal single-query sampling-based motion planning with quick replanning", Int. J. Robot. Res., vol. 35, no. 7, pp. 797-822, 2016.
2.
M. Phillips and M. Likhachev, "SIPP: Safe interval path planning for dynamic environments", Proc. IEEE Int. Conf. Robot. Automat., pp. 5628-5635, 2011.
3.
K. Yakovlev and A. Andreychuk, "Towards time-optimal any-angle path planning with dynamic obstacles", Proc. 31st Int. Conf. Automated Plan. Scheduling, pp. 405-414, 2021.
4.
M. Likhachev and D. Ferguson, "Planning long dynamically feasible maneuvers for autonomous vehicles", Int. J. Robot. Res., vol. 28, no. 8, pp. 933-945, 2009.
5.
P. E. Hart, N. J. Nilsson and B. Raphael, "A formal basis for the heuristic determination of minimum cost paths", IEEE Trans. Syst. Sci. Cybern., vol. 4, no. 2, pp. 100-107, Jul. 1968.
6.
S. M. LaValle, "Rapidly-exploring random trees: A. new tool for path planning", 1998.
7.
S. Karaman, M. Walter, A. Perez, E. Frazzoli and S. Teller, "Anytime motion planning using the RRT", Proc. IEEE Int. Conf. Robot. Automat., pp. 1478-1483, 2011.
8.
H. T. L. Chiang, J. Hsu, M. Fiser, L. Tapia and A. Faust, "RL-RRT: Kinodynamic motion planning via learning reachability estimators from RL policies", IEEE Robot. Automat. Lett., vol. 4, no. 4, pp. 4298-4305, Oct. 2019.
9.
A. Faust et al., "PRM-RL: Long-range robotic navigation tasks by combining reinforcement learning and sampling-based planning", Proc. IEEE Int. Conf. Robot. Automat., pp. 5113-5120, 2018.
10.
D. González, J. Pérez, V. Milanés and F. Nashashibi, "A review of motion planning techniques for automated vehicles", IEEE Trans. Intell. Transp. Syst., vol. 17, no. 4, pp. 1135-1145, Apr. 2016.
11.
J. hwan Jeon, S. Karaman and E. Frazzoli, "Anytime computation of time-optimal off-road vehicle maneuvers using the RRT", Proc. 50th IEEE Conf. Decis. Control Eur. Control Conf., pp. 3276-3282, 2011.
12.
D. J. Webb and J. Van Den Berg, "Kinodynamic RRT*: Asymptotically optimal motion planning for robots with linear dynamics", Proc. IEEE Int. Conf. Robot. Automat., pp. 5054-5061, 2013.
13.
C. Xie, B. J. Van Den, S. Patil and P. Abbeel, "Toward asymptotically optimal motion planning for kinodynamic systems using a two-point boundary value problem solver", Proc. IEEE Int. Conf. Robot. Automat., pp. 4187-4194, 2015.
14.
G. Vailland, V. Gouranton and M. Babel, "Cubic bezier local path planner for non-holonomic feasible and comfortable path generation", Proc. IEEE Int. Conf. Robot. Automat., pp. 7894-7900, 2021.
15.
Z. H. Y. Chen and S. Li, "Horizon-based lazy optimal RRT for fast efficient replanning in dynamic environment", Auton Robot, vol. 43, pp. 2271-2292, 2019.
16.
J. Lin, T. Zhou, D. Zhu, J. Liu and M. Q.-H. Meng, "Search-based online trajectory planning for car-like robots in highly dynamic environments", Proc. IEEE Int. Conf. Robot. Automat., pp. 8151-8157, 2021.
17.
M. Rufli and R. Siegwart, "On the design of deformable input-/state-lattice graphs", Proc. IEEE Int. Conf. Robot. Automat., pp. 3071-3077, 2010.
18.
J. Ziegler and C. Stiller, "Spatiotemporal state lattices for fast trajectory planning in dynamic on-road driving scenarios", Proc. IEEE/RSJ Int. Conf. Intell. Robots Syst., pp. 1879-1884.
19.
C. Pérez-D'Arpino, C. Liu, P. Goebel, R. Martín-Martín and S. Savarese, "Robot navigation in constrained pedestrian environments using reinforcement learning", Proc. IEEE Int. Conf. Robot. Automat., pp. 1140-1146, 2021.
20.
A. I. Panov, "Simultaneous learning and planning in a hierarchical control system for a cognitive agent", Automat. Remote Control, vol. 83, no. 6, pp. 869-883, 2022.
21.
G. P. Kontoudis and K. G. Vamvoudakis, "Kinodynamic motion planning with continuous-time Q-learning: An online model-free and safe navigation framework", IEEE Trans. Neural Netw. Learn. Syst., vol. 30, no. 12, pp. 3803-3817, Dec. 2019.
22.
M. Jamal and A. Panov, "Adaptive maneuver planning for autonomous vehicles using behavior tree on apollo platform", Proc. Artif. Intell. XXXVIII., vol. 13101, pp. 327-340, 2021.
23.
B. Paden, M. Čáp, S. Z. Yong, D. Yershov and E. Frazzoli, "A survey of motion planning and control techniques for self-driving urban vehicles", IEEE Trans. Intell. Veh., vol. 1, no. 1, pp. 33-55, Mar. 2016.
24.
J. Schulman, F. Wolski, P. Dhariwal, A. Radford and O. Klimov, "Proximal policy optimization algorithms", 2017.
25.
T. Haarnoja, A. Zhou, P. Abbeel and S. Levine, "Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor", Proc. 35th Int. Conf. Mach. Learn., vol. 80, pp. 1861-1870, 2018.
26.
A. Abdolmaleki, J. T. Springenberg, Y. Tassa, R. Munos, N. Heess and M. Riedmiller, "Maximum a posteriori policy optimisation", Proc. 6th Int. Conf. Learn. Representations, 2018.
27.
M. Hessel et al., "Muesli: Combining improvements in policy optimization", Proc. 38th Int. Conf. Mach. Learn., vol. 139, pp. 4214-4226, 2021, [online] Available: http://arxiv.org/abs/2104.06159.
28.
M.-R. Hsieh, Y.-L. Lin and W. H. Hsu, "Drone-based object counting by spatially regularized regional proposal network", Proc. 2017 IEEE Int. Conf. Comput. Vis., pp. 4165-4173, 2017.
29.
A. Astolfi, "Exponential stabilization of a wheeled mobile robot via discontinuous control", J. Dyn. Syst. Meas. Control, vol. 121, no. 1, pp. 121-126, 1999.
30.
Y. Li, Z. Littlefield and K. E. Bekris, "Asymptotically optimal sampling-based kinodynamic planning", Int. J. Robot. Res., vol. 35, no. 5, pp. 528-564, 2016.

Contact IEEE to Subscribe

References

References is not available for this document.