I. Introduction:
Machine learning models heavily rely on large and diverse datasets for training. However, acquiring labeled data for training purposes is often expensive and time-consuming. Data augmentation emerges as a potent solution to mitigate the data scarcity dilemma [1]. By artificially enlarging the dataset, machine learning practitioners provide models with a richer and more varied set of examples, essential for robust learning. Traditional augmentation techniques like rotation, translation, and flipping have been the go-to methods, yet they possess limitations in capturing the intricacies and nuances of complex data distributions [1].