I. Introduction
While autonomous driving systems have significantly advanced, their promise of full autonomy has not been achieved yet. For example, The key reason is the presence of challenging edge cases, especially ones not commonly observed during training of the machine learning models [9], [23]. The standard strategy of improving the accuracy has been to design larger and more compute-intensive models [16]. Running such models requires specialized processors called graphical processing units (GPUs) with larger memories. However, such GPUs with large memories tend to be significantly more expensive, as shown in Fig. 1. Thus, a safer and more efficient strategy for enabling autonomous driving is needed.