Lecture 13 - Unsupervised Learning Flashcards

1
Q

What is unsupervised learning?

A

Unsupervised learning involves discovering patterns within data without labeled examples, focusing on describing variability in the data and estimating distributions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is generative modeling in the context of unsupervised learning?

A

Generative modeling aims to approximate high-dimensional distributions of data, allowing for the generation of new data samples that resemble the original dataset.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Explain unsupervised representation learning.

A

Unsupervised representation learning involves learning useful weights and representations from unlabeled data, which can be transferred to other tasks with minimal labeled examples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the main idea behind self-supervised learning?

A

Self-supervised learning creates surrogate tasks that do not require labels, enabling the model to learn representations by generating labels from the data itself and optimizing a pretext task.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Describe the concept of an autoencoder.

A

An autoencoder is a neural network that learns to encode input data into a lower-dimensional representation and then decodes it back to the original data, minimizing reconstruction error.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is a variational autoencoder (VAE)?

A

A VAE is a type of autoencoder that uses variational inference to encode input data into a probabilistic latent space and decode it, allowing for both data reconstruction and generation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Explain the role of the Evidence Lower Bound (ELBO) in VAEs.

A

ELBO is a variational bound used to approximate the likelihood of the data, allowing the optimization of the VAE by maximizing this bound instead of the intractable true likelihood.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is a Generative Adversarial Network (GAN)?

A

A GAN consists of two neural networks, a generator and a discriminator, that compete in a minimax game, with the generator creating realistic data samples to fool the discriminator, which tries to distinguish between real and fake samples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Describe the concept of contrastive learning.

A

Contrastive learning involves training models to distinguish between similar and dissimilar pairs of data, encouraging the model to learn representations that bring similar samples closer and push dissimilar samples apart.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the purpose of a latent variable model?

A

Latent variable models assume that data lies in a lower-dimensional subspace and build a mapping between the high-dimensional data space and this latent space, facilitating tasks like data generation and reconstruction.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Explain the main idea behind flows in unsupervised learning.

A

Flows use invertible transformations to map between data space and latent space, allowing for exact likelihood computation and sampling without the need for approximation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the challenge with high-dimensional data in unsupervised learning?

A

High-dimensional data requires more samples to effectively learn and model the underlying distribution, making it computationally challenging and necessitating advanced techniques to handle the dimensionality.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Write the formula for the reconstruction loss in an autoencoder.

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Provide the formula for the variational lower bound (ELBO) in VAEs.

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the formula for the contrastive loss?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How does a VAE differ from a traditional autoencoder?

A

A VAE encodes input data into a probabilistic latent space using variational inference, allowing for both data reconstruction and generation, while a traditional autoencoder uses a deterministic latent space.

17
Q

Explain the role of the generator in a GAN.

A

The generator in a GAN creates realistic data samples from random noise, aiming to fool the discriminator into classifying the generated samples as real.

18
Q

What are the advantages of using self-supervised learning?

A

Self-supervised learning allows models to learn useful representations from unlabeled data, reducing the dependency on large labeled datasets and enabling transfer learning to other tasks.

19
Q

Describe the concept of a surrogate loss in self-supervised learning.

A

A surrogate loss is a loss function created from the data itself, allowing the model to learn meaningful representations by solving a pretext task without requiring labeled data.

20
Q

What is the significance of the latent space in VAEs and GANs?

A

The latent space represents a lower-dimensional subspace where the data distribution is modeled, facilitating tasks like data generation, interpolation, and reconstruction.

21
Q

How do flows differ from VAEs in unsupervised learning?

A

Flows use invertible transformations to map between data space and latent space, allowing for exact likelihood computation and sampling without approximation, while VAEs rely on variational inference.

22
Q

What is the importance of contrastive learning in unsupervised representation learning?

A

Contrastive learning helps models learn robust representations by distinguishing between similar and dissimilar pairs, improving the quality of learned features for downstream tasks.

23
Q

How does the jigsaw puzzle pretext task work in self-supervised learning?

A

The jigsaw puzzle pretext task involves shuffling patches of an image and training a model to predict the correct arrangement, encouraging the model to learn spatial relationships and features.

24
Q

Explain the concept of predicting image rotations in self-supervised learning.

A

Predicting image rotations involves training a model to identify the rotation angle of an image, helping the model learn orientation-invariant features and improving representation learning.

25
Q

What is the main challenge with kernel density estimation (KDE) in high-dimensional data?

A

KDE requires a large number of samples to effectively fill the high-dimensional space, making it computationally infeasible for high-dimensional data like images.

26
Q

Describe the main idea behind hierarchical VAEs.

A

Hierarchical VAEs use multiple layers in the latent space to capture complex data distributions, improving the quality and realism of generated samples compared to standard VAEs.

27
Q

How does the contrastive loss function help in unsupervised learning?

A

The contrastive loss function encourages the model to learn representations that bring similar samples closer and push dissimilar samples apart, improving the quality of learned features.