Policy Optimization to Learn Adaptive Motion Primitives in Path Planning with Dynamic Obstacles


Panov A. Yakovlev K.


This letter addresses the kinodynamic motion planning for non-holonomic robots in dynamic environments with both static and dynamic obstacles – a challenging problem that lacks a universal solution yet. One of the promising approaches to solve it is decomposing the problem into the smaller sub-problems and combining the local solutions into the global one. The crux of any planning method for non-holonomic robots is the generation of motion primitives that generates solutions to local planning sub-problems. In this work we introduce a novel learnable steering function (policy), which takes into account kinodynamic constraints of the robot and both static and dynamic obstacles. This policy is efficiently trained via the policy optimization. Empirically, we show that our steering function generalizes well to unseen problems. We then plug in the trained policy into the sampling-based and lattice-based planners, and evaluate the resultant POLAMP algorithm (Policy Optimization that Learns Adaptive Motion Primitives) in a range of challenging setups that involve a car-like robot operating in the obstacle-rich parking-lot environments. We show that POLAMP is able to plan collision-free kinodynamic trajectories with success rates higher than 92%, when 50 simultaneously moving obstacles populate the environment showing better performance than the state-of-the-art competitors.

External links

DOI: 10.1109/LRA.2022.3233261

Download PDF from the arXiv repository: https://arxiv.org/abs/2212.14307

ResearchGate: https://www.researchgate.net/publication/366718264_Policy_Optimization_to_Learn_Adaptive_Motion_Primitives_in_Path_Planning_with_Dynamic_Obstacles

Reference link

Brian Angulo, Aleksandr Panov, Konstantin Yakovlev. Policy Optimization to Learn Adaptive Motion Primitives in Path Planning with Dynamic Obstacles // IEEE Robotics and Automation Letters, Volume 8, Issue 2, February 2023. pp. 824-831.