I. Introduction
The concept of stacking models in machine learning is an interesting application as it leverages the learning capabilities from the base models and enables a higher-level model to predict with a higher degree of accuracy [1]. Model stacking is a winning application among practitioners and competition winners [2]. Decision tree-based algorithms perform well on structured data. XGBoost one of the popular ensemble techniques based on gradient-boosting learning is preferred in many scenarios for predictive analysis due to its high accuracy of prediction and its feature of regularization [3] as compared to other ensemble techniques. XGBoost can work well for unbalanced datasets and also has a lesser requirement in terms of hyperparameter tunings as compared to Random Forest [23].In recent years LightGBM [4] has emerged as a promising competitor to XGBoost because of its speed and relatively higher efficiency. LightGBM with its inherent capability to split the tree leaf-wise as the best fit minimizes the loss and thus achieves higher accuracy. Also, it is quite popular that Lasso Regression [14] has emerged as a model for small datasets complemented by a very powerful feature selection capability. Thus at Level-0 of stacking, we have employed XGBoost, LightGBM, and Lasso Regression as the primary learners of best features contributing to more accurate results. Deep learning algorithms have been the best performers in terms of applications that demand higher amounts of accuracy. However, Deep Learning algorithms perform well on larger datasets i.e. they need more amount of training data. Also, the best part of Deep Learning algorithms is their ability to learn the features of the dataset used in the experimentation. The DeepMLP as the Level-1 model performs regression tasks for diabetes progression [5]. Based on these facts the research questions of the experimentation in this paper are as follows: