DL-09 - Generative models Flashcards
DL-09 - Generative models
What are the types of generative models mentioned in the lecture slides? (2)
- Variational autoencoders
- Generative adversarial networks
DL-09 - Generative models
What is VAE short for?
Variational autoencoder
DL-09 - Generative models
What is GAN short for?
Generative Adversial Networks
DL-09 - Generative models
What is a loss function you might use in an autoencoder
MSE
DL-09 - Generative models
What is a key consideration when building an autoencoder related to the latent space?
The dimensionality of the latent space.
DL-09 - Generative models
The dimensionality of the latent space impacts what in an autoencoder?
The quality of the results, with smaller latent spaces typically leading to poorer outcomes.
DL-09 - Generative models
What are the two main approaches for implementing decoders in autoencoders? (2)
- Traditional techniques like k-NN (Average value of nearest pixels) or bilinear interpolation
- Transposed convolution
DL-09 - Generative models
Autoencoders typically use _______ for implementing decoders. (1)
Transposed Convolution
DL-09 - Generative models
What is a transposed convolution?
Upscaling of an image
DL-09 - Generative models
How do you perform a transposed convolution?
1) Multiply the input by the kernel.
2) Position the intermediate value inside a larger tensor.
3) Sum up all the values.
(See image)
DL-09 - Generative models
What is the main difference between AEs and VAEs?
- AEs are deterministic, same output
- VAEs are probabilistic (random), can generate new output
DL-09 - Generative models
What is the architecture for a VAE?
(See image)
DL-09 - Generative models
What model is in the image? (See image)
A variational autoencoder.
DL-09 - Generative models
Which format does VAE use to describe each latent attribute?
VAE uses probability distributions.
DL-09 - Generative models
How does VAE generate a vector for the decoder model?
VAE randomly samples from each latent state distribution.
DL-09 - Generative models
What does an ideal autoencoder learn?
Descriptive attributes of input data in a compressed representation.
E.g. from a face:
- Smile
- Gender
- Beard
…
DL-09 - Generative models
Describe visually how VAEs represent latent attributes.
(See image)
DL-09 - Generative models
What is a latent attribute?
A latent attribute is a hidden descriptive feature of the data, such as gender, emotion, or skin tone in facial images.
DL-09 - Generative models
Describe how we generate new data with a VAE.
- Sample latent attributes.
- Send to decoder.
- Decoder generates output.
(See image)
DL-09 - Generative models
What should happen for values that are close in latent space?
They should produce very similar reconstructions.
DL-09 - Generative models
How can we achieve an interpretation of what a VAE network is learning?
By perturbating one latent variable while keeping all other variables fixed.
DL-09 - Generative models
What is an approach to encourage independence of latent features in VAEs?
Applying independent component analysis (ICA) to the encoder output.
DL-09 - Generative models
Why do we want features to be uncorrelated?
To learn the richest and most compact representation possible.
DL-09 - Generative models
If your representations are rich and compact, what feature do we need in the latent space?
We want features to be uncorrelated.
DL-09 - Generative models
What terms does the VAE loss consist of?
Reconstruction loss + KL divergence term
(See image)
DL-09 - Generative models
What is KL divergence short for?
Kullback-Leibler divergence
DL-09 - Generative models
What is KL divergence?
KL divergence is a measure of how one probability distribution differs from another.
DL-09 - Generative models
How do you compute KL divergence for VAE loss? (See image)
By setting a fixed prior distribution 𝑝(𝑧) for 𝑞(𝑧|𝑥^hat) based on some initial hypothesis or guess, and model learns 𝑝(𝑧|𝑥) using this prior.
(See image)
DL-09 - Generative models
What is the general formula for KL divergence?
(See image)