1 Introduction
Recently, significant and rapid advancements in deep learning (e.g., object detection algorithms for onboard cameras) have facilitated their use in autonomous vehicles (AVs) [1]. These learning-enabled systems (LESs) are safety critical [19], where their failure can result in loss of life, injuries, and financial damage.
This paper uses the term LESs to refer to AI-based systems whose behaviors are optimized based on training experiences (e.a., object detection algorithms).
LESs may not operate as expected if the training data does not sufficiently capture run-time contexts [22]. For example, an AV may be trained with external agents such as pedestrians or other vehicles, whose training behavior conforms to defined rules [39]. In contrast, human agents may exhibit unexpected behaviors (e.g., suddenly braking, unexpected lane changes, etc.), thereby introducing a new source of uncertainty for AVs. This paper uses the term “ego vehicle” to describe the AV under study and “non-ego vehicle” to describe other vehicles. We propose a non-cooperative game theoryNon-cooperative game theory describes a game theory setup where players are only aware of and attempt to optimize their individual objectives [30].
[30] and reinforcement learning (RL) [37] framework to discover unexpected behavior exhibited by a trained (selfish) non-ego vehicle and the unexpected responses from a naïveBasic driving behavior with no experience with uncertainty.
ego vehicle, which can be used to enable the robustification of the ego vehicle to prevent and/or mitigate unsafe situations.