I. Introduction
Prediction and decision-making are critical for automated driving. Automated vehicles (AV) must continue to make decisions under dynamic uncertainties induced by the complex, intertwined, and unknown environment [1]. By emulating human driver functionalities, an AV algorithm contains modules to predict the movements of surrounding traffic participants (TPs) and make decisions afterward [2]. However, this task is challenging, even for human drivers, due to the uncertainties associated with TPs. The potential future movements of human TPs are inherently multimodal [3]; that is, all forecast future trajectories within physical limits are reasonable. This characteristic adds complexity to decision-making for both humans and machines in such uncertain environments. While recent advancements in deep learning have significantly enhanced fields like language models and computer vision, these improvements also boost the development of self-driving algorithms, particularly in higher-level modules such as perception, prediction, and decision-making. Nonetheless, many implementation issues remain unresolved in this dynamic, uncertain, and highly personalized context. We aim to address high-level uncertainties via personalized prediction and dynamic online learning. Additionally, unlike conventional methods that select the prediction with the highest probability for PnD, our approach incorporates a novel traffic entropy reward in decision-making, easing both future predictions and decisions.