1. Introduction
Diffusion models are a class of generative models that exhibit remarkable performance across a broad range of tasks, including but not limited to image generation [14], [24], [8], [2], [29], [4], [15], [38], super-resolution [33], [39], [6], inpainting [22], [31], and text-to-image generation [25], [32], [27], [10]. These models utilize the diffusion process to gradually introduce noise into the input data until it conforms to a Gaussian distribution. They then learn the reversal of this process to restore the data from sampled noise. Consequently, they achieve exact likelihood computation and excellent sample quality. However, one major drawback of diffusion models is their slow generation process. For instance, on a V100 GPU, generating a 256 × 256 image with StyleGAN [16] only takes 0.015s, whereas the ADM model requires multiple time steps for denoising during generation, leading to a significantly longer generation time of 14.75s.