← Home

Quick answer

AI Summary: Introduces the VQ-VAE, a breakthrough generative model that learns discrete, highly compressed latent representations, solving the posterior collapse problem in standard autoencoders.

Claim

Neural Discrete Representation Learning

Aaron van den Oord·
Oriol Vinyals·
Koray Kavukcuoglu

ABSTRACT

Learning useful representations without supervision remains a key challenge in machine learning. We propose the Vector Quantised-Variational AutoEncoder (VQ-VAE), a simple yet powerful generative model that learns discrete representations of data. Unlike standard VAEs that use continuous latent spaces, VQ-VAE utilizes vector quantisation to overcome the 'posterior collapse' problem, where latents are ignored when paired with a powerful autoregressive decoder. We show that VQ-VAE can learn highly compressed, discrete latent representations of images, audio, and video, which can then be used by an autoregressive prior (like PixelCNN or WaveNet) to generate high-quality, coherent samples.

Review Snapshot

Explore ratings

4.6
★★★★★
5 ratings
5 star
60%
4 star
40%
3 star
0%
2 star
0%
1 star
0%

Recommendation

100%

recommend this content.

Review this content

Share your opinion to help other learners triage faster.

Write a review

Invite a reviewer

Invite someone by email to share an invited review for Neural Discrete Representation Learning.

Author Inquiries

Public questions about this content. Attendemia will route your question to the author. Vote on the most important ones. No guarantee of response.
Post an inquiry
Sort by: Most helpful