1. Introduction
Semi-supervised learning (SSL) [7] has shown to be promising for leveraging unlabeled data to reduce the cost of constructing labeled data [4], [5], [36], [40], [57] and even boost the performance at scale [29], [49], [67], [68]. The common approach of these algorithms is to produce pseudo-labels for unlabeled data based on model's predictions and utilize them for regularizing model training [29], [38], [57]. Although adopted in a variety of tasks, these algorithms often assume class-balanced data, while many real-world datasets exhibit long-tailed distributions [3], [18], [3]1,[3]2. With classimbalanced data, the class distribution of pseudo-labels from unlabeled data becomes severely biased to the majority classes due to confirmation bias [2]. Such biased pseudo-labels can further bias the model during training.
Glimpse of the DASO framework. DASO reduces the overall bias in pseudo-labels (PL) from unlabeled data by blending two complementary pls from different classifiers. Note that bias is conceptually illustrated as relative PL size (rel. PL size), meaning that pseudo-label size is normalized by actual label size.