I. Introduction
Streaming data classification consists of a routine where a model is trained on the historical data and then used to classify upcoming samples. When the labels of the newly arrived samples are available, they become a part of the training data. Concept drift refers to inconsistencies in data generation at different times, which means the training data and the testing data have different distributions [1]–[3]. Drift detection aims to identify these differences with a statistical guarantee through what is, typically, a four-step process [4]: 1) cut data stream into chunks as training/testing sets; 2) abstract the datasets into a comparable model; 3) develop a test statistical or similarity measurement to quantify the distance between the models; and 4) design a hypothesis test to investigate the null hypothesis (most often, the null hypothesis is that there is no concept drift).