1. Introduction
Image-based virtual try-on aims at generating natural, distortion and artifacts-free images of a person wearing a selected garment. Image synthesis via GANs [7] has been widely used in applications like image editing [16], [23], [26], style-transfer [2], [8], [33] and image generation [13], [30], [34]. However, simply using synthesis methods that holistically change the image does not result in the desired quality in virtual try-on setting. Existing methods adopt a scheme where the garment is first warped to meet the target person pose requirements. A GAN based generator network then fuses the warped garment and the target person images to generate a final try-on image. Traditionally, the warping is either done by a Thin Plate Spline (TPS) warp [3, 5, 10, 12, 15, 25, 29], or a dense flow fields based warp [1], [4], [9], [11], [14], or a combination of both [28]. In any case, the warping is inherently not capable of modeling all the changes that a garment undergoes (e.g occlusions) when it fits on a target person. And forcing it to do so, results in artifacts such as texture squeezing, stretching, and garment tear, etc.