I. Introduction
Recently, Convolutional Neural Network (CNN)-based methods have achieved promising performance in medical image segmentation, however, they heavily rely on extensive supervised data collected from multiple medical sites (i.e., hospitals) for model training [1]. Nevertheless, acquiring these multi-site datasets simultaneously and aggregating them into a unified set is impractical due to high storage costs and privacy concerns [2]. It is desirable to use an incremental data stream for model training, i.e., Continual Learning (CL), where the model is encouraged to learn consecutively from sequentially-arriving sites without catastrophic forgetting on the previously-learned yet currently-inaccessible sites [3]. Despite its success, existing CL typically performs in the fully-supervised training setting, which is extremely expensive to manually annotate all data in each arriving site. Yet, in clinical practice, most medical sites often only afford partial data labeling, as it saves financial budgets and circumvents the high demand for annotators with comprehensive expertise [4]. How to leverage these partially-labeled medical sites in CL is a realistic question to be solved.