Loading [MathJax]/extensions/MathMenu.js
Deep Multi-Agent Reinforcement Learning for Highway On-Ramp Merging in Mixed Traffic | IEEE Journals & Magazine | IEEE Xplore

Deep Multi-Agent Reinforcement Learning for Highway On-Ramp Merging in Mixed Traffic


Abstract:

On-ramp merging is a challenging task for autonomous vehicles (AVs), especially in mixed traffic where AVs coexist with human-driven vehicles (HDVs). In this paper, we fo...Show More

Abstract:

On-ramp merging is a challenging task for autonomous vehicles (AVs), especially in mixed traffic where AVs coexist with human-driven vehicles (HDVs). In this paper, we formulate the mixed-traffic highway on-ramp merging problem as a multi-agent reinforcement learning (MARL) problem, where the AVs (on both merge lane and through lane) collaboratively learn a policy to adapt to HDVs to maximize the traffic throughput. We develop an efficient and scalable MARL framework that can be used in dynamic traffic where the communication topology could be time-varying. Parameter sharing and local rewards are exploited to foster inter-agent cooperation while achieving great scalability. An action masking scheme is employed to improve learning efficiency by filtering out invalid/unsafe actions at each step. In addition, a novel priority-based safety supervisor is developed to significantly reduce collision rate and greatly expedite the training process. A gym-like simulation environment is developed and open-sourced with three different levels of traffic densities. We exploit curriculum learning to efficiently learn harder tasks from trained models under simpler settings. Comprehensive experimental results show the proposed MARL framework consistently outperforms several state-of-the-art benchmarks.
Published in: IEEE Transactions on Intelligent Transportation Systems ( Volume: 24, Issue: 11, November 2023)
Page(s): 11623 - 11638
Date of Publication: 22 June 2023

ISSN Information:

Funding Agency:

References is not available for this document.

Select All
1.
Future of Driving, Mar. 2021, [online] Available: https://www.tesla.com/autopilot.
2.
Apollo Open Platform, Mar. 2021, [online] Available: https://apollo.auto/developer.html.
3.
V. V. Dixit, S. Chand and D. J. Nair, "Autonomous vehicles: Disengagements accidents and reaction times", PLoS ONE, vol. 11, no. 12, Dec. 2016.
4.
F. M. Favaro, N. Nader, S. O. Eurich, M. Tripp and N. Varadaraju, "Examining accident reports involving autonomous vehicles in California", PLoS ONE, vol. 12, no. 9, Sep. 2017.
5.
D. Ni and J. D. Leonard, "A simplified kinematic wave model at a merge bottleneck", Appl. Math. Model., vol. 29, no. 11, pp. 1054-1072, Nov. 2005.
6.
L. Leclercq, J. A. Laval and N. Chiabaut, "Capacity drops at merges: An endogenous model", Proc.-Social Behav. Sci., vol. 17, pp. 12-26, Jan. 2011.
7.
Z. H. Khattak, B. L. Smith, H. Park and M. D. Fontaine, "Cooperative lane control application for fully connected and automated vehicles at multilane freeways", Transp. Res. C Emerg. Technol., vol. 111, pp. 294-317, Feb. 2020.
8.
D. Bevly et al., "Lane change and merge maneuvers for connected and automated vehicles: A survey", IEEE Trans. Intell. Vehicles, vol. 1, no. 1, pp. 105-120, Mar. 2016.
9.
J. Rios-Torres and A. A. Malikopoulos, "A survey on the coordination of connected and automated vehicles at intersections and merging at highway on-ramps", IEEE Trans. Intell. Transp. Syst., vol. 18, no. 5, pp. 1066-1077, May 2017.
10.
L. N. Jacobson, K. C. Henry and O. Mehyar, Real-Time Metering Algorithm for Centralized Control, Transportation Research Board, vol. 1232, pp. 17-26, 1989, [online] Available: https://trid.trb.org/view/308666.
11.
J. Hourdakis and P. G. Michalopoulos, "Evaluation of ramp control effectiveness in two twin cities freeways", Transp. Res. Rec. J. Transp. Res. Board, vol. 1811, no. 1, pp. 21-29, Jan. 2002.
12.
Y. Lin, J. McPhee and N. L. Azad, "Anti-jerk on-ramp merging using deep reinforcement learning", Proc. IEEE Intell. Vehicles Symp. (IV), pp. 7-14, Oct. 2020.
13.
W. Cao, M. Mukai, T. Kawabe, H. Nishira and N. Fujiki, "Cooperative vehicle path generation during merging using model predictive control with real-time optimization", Control Eng. Pract., vol. 34, pp. 98-105, Jan. 2015.
14.
J. B. Rawlings, D. Q. Mayne and M. M. Diehl, Model Predictive Control: Theory Computation and Design, Portland, OR, USA:Nob Hill Publishing, 2017.
15.
M. Papageorgiou, C. Kiakaki, V. Dinopoulou, A. Kotsialos and Y. Wang, "Review of road traffic control strategies", Proc. IEEE, vol. 91, no. 12, pp. 2043-2067, Dec. 2003.
16.
M. Papageorgiou and A. Kotsialos, "Freeway ramp metering: An overview", IEEE Trans. Intell. Transp. Syst., vol. 3, no. 4, pp. 271-281, Dec. 2002.
17.
I. Papamichail and M. Papageorgiou, "Traffic-responsive linked ramp-metering control", IEEE Trans. Intell. Transp. Syst., vol. 9, no. 1, pp. 111-121, Mar. 2008.
18.
Y. Hou, P. Edara and C. Sun, "Modeling mandatory lane changing using Bayes classifier and decision trees", IEEE Trans. Intell. Transp. Syst., vol. 15, no. 2, pp. 647-655, Apr. 2014.
19.
F. Marczak, W. Daamen and C. Buisson, "Key variables of merging behaviour: Empirical comparison between two sites and assessment of gap acceptance theory", Proc.-Social Behav. Sci., vol. 80, pp. 678-697, Jun. 2013.
20.
Z. E. A. Kherroubi, S. Aknine and R. Bacha, "Novel decision-making strategy for connected and autonomous vehicles in highway on-ramp merging", IEEE Trans. Intell. Transp. Syst., vol. 23, no. 8, pp. 12490-12502, Aug. 2022.
21.
J. Lubars et al., "Combining reinforcement learning with model predictive control for on-ramp merging", arXiv:2011.08484, 2020.
22.
T. P. Lillicrap et al., "Continuous control with deep reinforcement learning", arXiv:1509.02971, 2015.
23.
J. Wang, T. Shi, Y. Wu, L. Miranda-Moreno and L. Sun, "Multi-agent graph reinforcement learning for connected automated driving", Proc. 37th Int. Conf. Mach. Learn., pp. 1-6, 2020.
24.
P. Palanisamy, "Multi-agent connected autonomous driving using deep reinforcement learning", Proc. Int. Joint Conf. Neural Netw. (IJCNN), pp. 1-7, Jul. 2020.
25.
M. Kaushik and K. M. Krishna, "Parameter sharing reinforcement learning architecture for multi agent driving behaviors", arXiv:1811.07214, 2018.
26.
P. Young Joun Ha, S. Chen, J. Dong, R. Du, Y. Li and S. Labi, "Leveraging the capabilities of connected and autonomous vehicles and multi-agent reinforcement learning to mitigate highway bottleneck congestion", arXiv:2010.05436, 2020.
27.
S. Bhalla, S. G. Subramanian and M. Crowley, "Deep multi agent reinforcement learning for autonomous driving", Proc. Can. Conf. Artif. Intell., pp. 67-78, 2020.
28.
J. Dong, S. Chen, P. Y. J. Ha, Y. Li and S. Labi, "A DRL-based multiagent cooperative control framework for CAV networks: A graphic convolution Q network", arXiv:2010.05437, 2020.
29.
C. Yu et al., "Distributed multiagent coordinated learning for autonomous driving in highways based on dynamic coordination graphs", IEEE Trans. Intell. Transp. Syst., vol. 21, no. 2, pp. 735-748, Feb. 2020.
30.
V. Mnih et al., "Asynchronous methods for deep reinforcement learning", Proc. Int. Conf. Mach. Learn., pp. 1928-1937, 2016.
Contact IEEE to Subscribe

References

References is not available for this document.