Quick answer
AI Summary: Proves definitively that diffusion models, when properly architected and guided by classifiers, surpass the visual fidelity of state-of-the-art GANs, effectively kicking off the modern generative art revolution.
AI Summary: Proves definitively that diffusion models, when properly architected and guided by classifiers, surpass the visual fidelity of state-of-the-art GANs, effectively kicking off the modern generative art revolution.
We show that diffusion models can achieve image sample quality superior to the current state-of-the-art generative models. We achieve this on unconditional image synthesis by finding a better architecture through a series of ablations. For conditional image synthesis, we further improve sample quality with classifier guidance: using gradients from a classifier to guide a diffusion model during sampling. We achieve an FID of 2.77 on ImageNet 256x256, substantially outperforming the current state-of-the-art GANs (like BigGAN). This work proves that likelihood-based models can scale to high-fidelity, high-resolution image synthesis without the training instabilities associated with adversarial networks.
Share your opinion to help other learners triage faster.
Write a reviewInvite someone by email to share an invited review for Diffusion Models Beat GANs on Image Synthesis.