1. Introduction
Understanding how human drivers and autonomous vehicles make decisions is essential to ensure safe and reliable operation in various real-world scenarios. Neural networks are powerful tools used for automated learning in the field of self-driving cars [2], [8], [25], [28], [30], [34], [37]. However, one significant challenge associated with deep neural networks is their nature as black-box models, which hinders the interpretability of their decision-making process. This paper proposes to address this challenge by applying concept bottleneck models for explaining driving scenarios. Concept bottleneck models incorporate vision-based human-defined concepts within a bottleneck in the model architecture [22], [31]. By encoding driving and scenario-related concepts into the decision-making process, our objective is to provide interpretable and explainable insights into the factors that influence the actions of both drivers and autonomous vehicles. Previous research has demonstrated the effectiveness of learning vehicle controls for autonomous driving [6], [24], [39], [42], [42], but the lack interpretability poses challenges to trust, safety, and regulatory compliance. The development of interpretable and explainable models has thus gained significant attention in the research community, aiming to bridge the gap between the performance and interpretability of deep learning models.