Communication-Efficient and Resilient Distributed Deep Reinforcement Learning for Multi-Agent Systems | IEEE Conference Publication | IEEE Xplore

Communication-Efficient and Resilient Distributed Deep Reinforcement Learning for Multi-Agent Systems


Abstract:

In this paper, we address the challenge of developing a resilient Deep Q-Network (DQN) algorithm for multi-agent systems (MASs) under Byzantine attacks. Traditional Q-lea...Show More

Abstract:

In this paper, we address the challenge of developing a resilient Deep Q-Network (DQN) algorithm for multi-agent systems (MASs) under Byzantine attacks. Traditional Q-learning methods suffer from significant communication overhead and limited capacity to handle high-dimensional state spaces, which restricts their scalability and effectiveness in large-scale MASs. To mitigate these issues, we propose an event-triggered resilient DQN (ET-RDQN) algorithm that reduces unnecessary communication while maintaining robustness under the environment exists adversarial agents. Our approach extends the DQN algorithm to MASs and employs the Mean-Subsequence Reduced (MSR) algorithm to filter out extreme values, ensuring that normal agents can still approximately estimate the optimal state-action values and achieve consensus even in the presence of malicious agents. The algorithm incorporates an event-triggered mechanism that triggers communication only when the triggering conditions are met, thereby reducing communication overhead. Theoretical analysis and extensive simulations demonstrate that our ET-RDQN algorithm significantly improves the resilience and efficiency of MASs in the presence of Byzantine attacks, providing a scalable and effective solution for distributed reinforcement learning in adversarial environments. Finally, we conclude this paper and outline potential directions for future work.
Date of Conference: 18-20 October 2024
Date Added to IEEE Xplore: 22 January 2025
ISBN Information:

ISSN Information:

Conference Location: Nanjing, China

Funding Agency:


I. Introduction

Multi-agent systems (MASs) have garnered significant attention due to their applicability in diverse domains such as formation control [1], autonomous vehicles [2] and smart grid [3]. In multi-agent reinforcement learning (MARL), several agents engage with their surroundings to acquire rewards and develop best practices to fulfill extended objectives. [4]. The incorporation of reinforcement learning (RL) techniques has enabled MASs to autonomously learn and adapt to complex environments, making them highly effective in dynamic and uncertain scenarios [5], [6]. However, MASs are vulnerable to Byzantine attacks, where some agents may behave arbitrarily due to faults or malicious intentions. These attacks can severely disrupt the learning process and degrade system performance. To mitigate the impact of Byzantine attacks, various methods have been proposed, focusing on enhancing consensus protocols and improving fault tolerance. Techniques such as majority voting [7], reputation-based systems [8], optimization algorithm [9], [10] and resilient consensus algorithms [11]–[13] have been explored to ensure system reliability in the presence of faulty or malicious agents. While these methods provide some level of robustness, they often come with significant communication overhead or computational complexity, which can limit their scalability and effectiveness in large-scale MASs. To address these limitations, researchers have explored alternative approaches that reduce communication or computational demands. Current studies in event-triggered mechanisms for multi-agent systems primarily aim to develop rules for triggering that ascertain the timing of such events and avoid Zeno phenomena, defined by endlessly frequent triggers within a limited time frame [14], [15]. Therefore, it is imperative to develop an algorithm capable of efficiently handling high-dimensional state spaces, maintaining robust performance in the presence of Byzantine attacks, and operating with low communication overhead.

Contact IEEE to Subscribe

References

References is not available for this document.