I. Introduction
Gaussian mixture models (GMMs) were extensively studied with applications in many domains, such as density estimation [1], clustering [2], [3], classification [4], image registration [5], [6], and regression [7], [8]. There are two main issues in the application of mixture models. The first is the estimation of model parameters. Parameter estimation is generally based on the maximum likelihood (ML) or maximum a posteriori (MAP) expectation maximization (EM) algorithm [9]–[11] or its variational extensions [12], [13]. The second issue is the choice of the number of mixture components. There are cases where the number of components is known a priori, (e.g., some classification problems). In the majority of applications, this number is, however, unknown [14]–[16], [18].