Poster
Distilling Parallel Gradients for Fast ODE Solvers of Diffusion Models
Beier Zhu · Ruoyu Wang · Tong Zhao · Hanwang Zhang · Chi Zhang
Diffusion models (DMs) have achieved state-of-the-art generative performance but suffer from high sampling latency due to their sequential denoising nature. Existing solver-based acceleration methods often face image quality degradation under a low-latency budget. In this paper, we propose the Ensemble Parallel Direction solver (dubbed as EPD-Solver), a novel ODE solver that mitigates truncation errors by incorporating multiple parallel gradient evaluations in each ODE step. Importantly, since the additional gradient computations are independent, they can be fully parallelized, preserving low-latency sampling. Our method optimizes a small set of learnable parameters in a distillation fashion, ensuring minimal training overhead. Extensive experiments on various image synthesis benchmarks demonstrate the effectiveness of our EPD-Solver in achieving high-quality and low-latency sampling. For example, at the same latency level of 5 NFE, EPD achieves an FID of 5.26 on CIFAR-10, 8.74 on FFHQ, 7.95 on ImageNet, and 7.79 on LSUN Bedroom, surpassing existing learning-based solvers by a significant margin.
Live content is unavailable. Log in and register to view live content