← Home

Quick answer

AI Summary: Proves definitively that diffusion models, when properly architected and guided by classifiers, surpass the visual fidelity of state-of-the-art GANs, effectively kicking off the modern generative art revolution.

Claim

Diffusion Models Beat GANs on Image Synthesis

Prafulla Dhariwal·
Alex Nichol

ABSTRACT

We show that diffusion models can achieve image sample quality superior to the current state-of-the-art generative models. We achieve this on unconditional image synthesis by finding a better architecture through a series of ablations. For conditional image synthesis, we further improve sample quality with classifier guidance: using gradients from a classifier to guide a diffusion model during sampling. We achieve an FID of 2.77 on ImageNet 256x256, substantially outperforming the current state-of-the-art GANs (like BigGAN). This work proves that likelihood-based models can scale to high-fidelity, high-resolution image synthesis without the training instabilities associated with adversarial networks.

Review Snapshot

Explore ratings

4.6
★★★★★
5 ratings
5 star
60%
4 star
40%
3 star
0%
2 star
0%
1 star
0%

Recommendation

100%

recommend this content.

Review this content

Share your opinion to help other learners triage faster.

Write a review

Invite a reviewer

Invite someone by email to share an invited review for Diffusion Models Beat GANs on Image Synthesis.

Author Inquiries

Public questions about this content. Attendemia will route your question to the author. Vote on the most important ones. No guarantee of response.
Post an inquiry
Sort by: Most helpful