lecture 4 Flashcards

1
Q

What is the focus of probabilistic models in machine learning?

A

Using probability theory to select models for given data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is conditional probability?

A

p(X | Y) is the probability of X given Y.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is Bayes’ theorem?

A

p(X | Y) = p(Y | X) * p(X) / p(Y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is a useful probability rule?

A

p(X, Y) = p(X | Y) * p(Y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is (Conditional) Independence?

A

When knowing Y provides no information about X.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is expectation in probability?

A

The weighted average of all possible values of a random variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are common probability distributions in ML?

A

Bernoulli, Categorical, and Normal distributions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What does p(Data | θ) represent?

A

The likelihood of observing data given parameters θ.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the goal in probabilistic learning?

A

To infer the parameters θ given only the observed data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the frequentist view of probability?

A

The true model exists but is unknown, and we estimate its parameters.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is maximum likelihood estimation (MLE)?

A

Finding parameters that maximize the likelihood of observed data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the likelihood function L(θ)?

A

p(X | θ), the probability of observed data given parameters.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the MLE (maximum likelihood estimation) formula?

A

θ̂ = argmax_θ p(X | θ)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is a simple analogy for probabilistic learning?

A

A ‘machine’ generates data, and we infer its parameters.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What does the model space represent?

A

All possible models that could explain the observed data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What does maximum likelihood estimation (MLE) assume?

A

That the best model is the one that makes the observed data most likely.

17
Q

What is a practical example of MLE?

A

Estimating if a coin is fair or biased based on observed flips.

18
Q

What is the difference between frequentist and Bayesian inference?

A

Frequentists estimate a fixed parameter, while Bayesians use probability distributions over parameters.

19
Q

What is the Bayesian approach to learning?

A

Updating beliefs about parameters using observed data and prior knowledge.

20
Q

What is a prior in Bayesian inference?

A

The initial belief about a parameter before observing data.

21
Q

What is a posterior distribution?

A

The updated belief about a parameter after observing data.

22
Q

What is the likelihood in Bayesian inference?

A

The probability of observed data given a parameter.

23
Q

What is the Bayes rule formula for updating beliefs?

A

Posterior ∝ Likelihood × Prior

24
Q

What is a conjugate prior?

A

A prior that, when updated, results in a posterior with the same functional form.

25
Q

What is the benefit of using conjugate priors?

A

They simplify Bayesian inference calculations.

26
Q

What is MAP estimation?

A

Maximum a posteriori estimation, which finds the most probable parameter given both data and prior.

27
Q

What is the difference between MAP and MLE?

A

MAP considers prior knowledge, while MLE relies only on data.

28
Q

What is the purpose of probabilistic modeling in ML?

A

To incorporate uncertainty and make robust predictions.

29
Q

What is the relationship between likelihood and probability?

A

Likelihood measures how well parameters explain data, while probability represents chance.

30
Q

What is the key takeaway from probabilistic models?

A

They provide a framework for handling uncertainty in machine learning.