Loading [MathJax]/extensions/MathZoom.js
HELSA: Hierarchical Reinforcement Learning with Spatiotemporal Abstraction for Large-Scale Multi-Agent Path Finding | IEEE Conference Publication | IEEE Xplore

HELSA: Hierarchical Reinforcement Learning with Spatiotemporal Abstraction for Large-Scale Multi-Agent Path Finding


Abstract:

The Multi-Agent Path Finding (MAPF) problem is a critical challenge in dynamic multi-robot systems. Recent studies have revealed that multi-agent reinforcement learning (...Show More

Abstract:

The Multi-Agent Path Finding (MAPF) problem is a critical challenge in dynamic multi-robot systems. Recent studies have revealed that multi-agent reinforcement learning (MARL) is a promising approach to solving MAPF problems in a fully decentralized manner. However, as the size of the multi-robot system increases, sample inefficiency becomes a major impediment to learning-based methods. This paper presents a hierarchical reinforcement learning (HRL) framework for large-scale multi-agent path finding, featuring applying spatial and temporal abstraction to capture intermediate reward and thus encourage efficient exploration. Specifically, we introduce a meta controller that partitions the map into interconnected regions and optimizes agents' region-wise paths towards globally better solutions. Additionally, we design a lower-level controller that efficiently solves each sub-problem by incorporating heuristic guidance and an inter-agent communication mechanism with RL-based policies. Our empirical results on test instances of various scales demonstrate that our method outperforms existing approaches in terms of both success rate and makespan.
Date of Conference: 01-05 October 2023
Date Added to IEEE Xplore: 13 December 2023
ISBN Information:

ISSN Information:

Conference Location: Detroit, MI, USA

Funding Agency:

References is not available for this document.

I. Introduction

In multi-robot systems, the ability to navigate effectively and efficiently is crucial. Such systems may entail thousands of automated mobile robots (AMRs) that have to find collision-free paths in a collaborative manner [1]. The Problem of multi-agent path finding is a challenging NP-hard problem [2] with numerous variants [3], and it has been of interest to researchers in the field of multi-agent systems ever since its inception.

Select All
1.
P. R. Wurman, R. D’ Andrea and M. Mountz, "Coordinating hundreds of cooperative autonomous vehicles in warehouses", AI magazine, vol. 29, no. 1, pp. 9-9, 2008.
2.
J. Yu and S. M. LaValle, "Structure and intractability of optimal multi-robot path planning on graphs", Twenty-Seventh AAAI Conference on Artificial Intelligence, 2013.
3.
R. Stern, N. R. Sturtevant, A. Felner et al., "Multi-agent pathfinding: Definitions variants and benchmarks", Twelfth Annual Symposium on Combinatorial Search, 2019.
4.
G. Sartoretti, J. Kerr, Y. Shi et al., "Primal: Pathfinding via reinforcement and imitation multi-agent learning", IEEE Robotics and Automation Letters, vol. 4, no. 3, pp. 2378-2385, 2019.
5.
Z. Liu, B. Chen, H. Zhou et al., "Mapper: Multi-agent path planning with evolutionary reinforcement learning in mixed dynamic environments", 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 11 748-11 754, 2020.
6.
Z. Ma, Y. Luo and H. Ma, "Distributed heuristic multi-agent path finding with communication", 2021 IEEE International Conference on Robotics and Automation (ICRA), pp. 8699-8705, 2021.
7.
Z. Ma, Y. Luo and J. Pan, "Learning selective communication for multi-agent path finding", IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 1455-1462, 2021.
8.
S. Ren, K. He, R. Girshick et al., "Faster r-cnn: Towards real-time object detection with region proposal networks", Advances in neural information processing systems, vol. 28, 2015.
9.
R. S. Sutton, D. Precup and S. Singh, "Between mdps and semi-mdps: A framework for temporal abstraction in reinforcement learning", Artificial intelligence, vol. 112, no. 1–2, pp. 181-211, 1999.
10.
E. Boyarski, A. Felner, R. Stern et al., "Icbs: Improved conflict-based search algorithm for multi-agent pathfinding", Twenty-Fourth International Joint Conference on Artificial Intelligence, 2015.
11.
C. Amato, G. D. Konidaris and L. P. Kaelbling, "Planning with macro-actions in decentralized pomdps", Proceedings of the 2014 International Conference on Autonomous Agents and Multi-Agent Systems, pp. 12731280, 2014.
12.
D. Horgan, J. Quan, D. Budden et al., "Distributed prioritized experience replay", International Conference on Learning Representations, 2018.
13.
Q. Li, W. Lin, Z. Liu et al., "Message-aware graph attention networks for large-scale multi-robot path planning", IEEE Robotics and Automation Letters, vol. 6, no. 3, pp. 5533-5540, 2021.
14.
Y. Liu, W. Wang, Y. Hu et al., "Multi-agent game abstraction via graph attention neural network", Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, no. 05, pp. 7211-7218, 2020.
15.
M. Damani, Z. Luo, E. Wenzel et al., "Primal _2: Pathfinding via reinforcement and imitation multi-agent learning-lifelong", IEEE Robotics and Automation Letters, vol. 6, no. 2, pp. 2666-2673, 2021.

Contact IEEE to Subscribe

References

References is not available for this document.