Dissecting the Variational Autoencoder's ELBO
Variational Autoencoders (VAEs) are powerful generative models that optimize a lower bound on the data log-likelihood, known as the Evidence Lower Bound (ELBO). The ELBO for a single data point is given by:
where is the approximate posterior (encoder), is the likelihood (decoder), and is the prior over the latent space (often a standard normal distribution).
- Reconstruction Term: Identify and explain the first term, . What is its intuitive role in the VAE? Why is it an expectation? How is it typically approximated in practice (e.g., using Monte Carlo sampling)?
- KL Divergence Term: Identify and explain the second term, . What does the KL divergence measure? What is the purpose of this term in the ELBO, and what prior is commonly chosen?
- Trade-off: Explain the inherent trade-off between these two terms during VAE training. What happens if one term dominates the other?
- Reparameterization Trick: Briefly describe why the reparameterization trick is necessary for training VAEs (specifically for backpropagating through the sampling step). You don't need to derive it, just explain its purpose.
- Verification: Conceptual understanding is key here. Verify your understanding by explaining how changing the weight of the KL divergence term (a common practice) would affect the generated samples and the latent space structure.