I. Introduction
Automated driving represents the forefront of future transportation technologies, promising enhanced safety and efficiency. Semantic segmentation is a pivotal task in automated driving perception. It enables vehicles to comprehend driving scenes, detect obstacles, and navigate accordingly by assigning a specific class label to each pixel in an image. Therefore, achieving robust scene understanding is crucial for safe and accurate driving decisions. However, training a reliable semantic segmentation model demands extensive data, which can be time-consuming and expensive to collect and annotate [1]. Additionally, certain objects encountered on the road may be rare or uncommon, leading to insufficient information for these classes in the training dataset. Consequently, the model may struggle to provide accurate predictions for such objects [2]. To address these challenges, researchers have been exploring approaches to complement current research, where they utilize limited annotated examples of previously unknown classes to detect those novel object classes using Few-Shot Semantic Segmentation (FSSS), initiated by the work One-Shot Learning for Semantic Segmentation (OSLSM) [3]. While significant progress has been made in FSSS research, the existing work has primarily focused on object-centric datasets like Pascal VOC [4] or COCO [5], which often contain a restricted complexity of scenes compared to the road scene datasets like Cityscapes [1] and Mapillary [6].