1 Introduction
SUPERVISED learning typically requires large volumes of labelled data. Training of sophisticated deep neural networks (DNNs) often involves learning from thousands (MNIST [1], CIFAR [2]) (sometimes millions, e.g., ImageNet [3]) of data samples. Despite their ability to train complex models, these training datasets pose practical challenges. These datasets (i) are often huge in size (e.g., ImageNet [3]), (ii) are proprietary, and (iii) involve privacy concerns (e.g., biometric, healthcare data). Hence, in practice, public access to the data samples used for training may not always be feasible. Instead, the resulting trained models can be made available relatively easily. For instance, Facebook’s Deepface [4] model is trained over 4M confidential face images.