lecture 1 - normal distribution Flashcards

1
Q

What is precision (β) in the context of normal distributions?

A
  • inverse of the variance
  • 1/(σ^2)
  • represents how concentrated/sharp the distribution is
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

How does precision change as variance increases or decreases?

A
  • When variance (σ^2) increases, precision (β) decreases.
  • As variance approaches infinity, precision approaches zero.
  • When variance decreases, precision increases, resulting in a sharper peak.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Why is precision (β) used instead of variance (σ^2) in some cases?

A
  • For brevity and convenience
  • In Bayesian statistics and machine learning, expressing inverse variance directly simplifies equations.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How are precision and variance related?

A
  • Precision is the reciprocal of variance
  • σ^2 = 1/β
  • β = 1/(σ^2)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does the Probability Density Function (PDF) of a univariate normal distribution tell us?

A
  • how likely a given value of x is, centered at the mean μ and spreading out based on the variance σ^2.
  • The height of the distribution curve at a given value of x
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What does (x−μ)^2 represent in the normal distribution?

A
  • The squared distance from the mean.
  • Values further from the mean have lower probabilities, creating a tapering effect on both sides of the curve.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the expected value of a univariate normal distribution?

A
  • The expected value of x (first moment)
  • E[x] = μ
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the expected value of x^2 in a univariate normal distribution?

A
  • The expected value of x^2 (second moment)
  • E[x^2] = μ^2 + σ^2
  • So it’s slightly higher than μ^2 as it accounts for the variance as well
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How is variance derived from the moments of the normal distribution?

A
  • Variance is the second moment minus the square of the first moment
  • var(x) = E[x^2] - (E[x])^2 = σ^2
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are the properties of the normal distribution PDF?

A
  1. The PDF is always positive (N(x|μ,σ^2) > 0)
  2. the total area under the curve is 1
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is a multivariate normal distribution?

A
  • A multivariate normal distribution is a generalization of the univariate normal distribution to higher dimensions.
  • It describes the distribution of a vector of variables.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What are the key parameters of a multivariate normal distribution?

A
  1. μ: A D-dimensional mean vector.
  2. Σ: A D×D covariance matrix (symmetric and positive definite).
  3. D: The dimensionality of the data.
  4. ∣Σ∣: The determinant of the covariance matrix.
  5. x: a vector of variables
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What does (x−μ)^T Σ^−1 (x−μ) represent in the multivariate normal distribution?

A
  • It represents the “distance” of the vector x from the mean μ, scaled by the spread and orientation of the distribution.
  • This term measures how likely it is to observe x given the distribution.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What happens when the covariance matrix of a multivariate distribution is an identity matrix?

A
  • When the covariance matrix is an identity matrix, the variables are independent and not correlated.
  • The distribution is spherical, with equal spread in all directions.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What changes about the multivariate distribution when covariances are introduced?

A
  • When covariances are introduced, relationships between variables emerge.
  • The distribution becomes elliptical, indicating correlations between variables.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What are the properties of marginal and conditional distributions for jointly Gaussian variables?

A
  • If two sets of variables are jointly Gaussian, the conditional distribution of one set given the other is also Gaussian.
  • Similarly, the marginal distribution of either set is Gaussian.
17
Q

What is the precision matrix (Λ)?

A
  • The precision matrix (Λ) is the inverse of the covariance matrix (Σ): Λ=Σ^−1
  • It represents the concentration of the multivariate normal distribution.
18
Q

What are the properties of the precision matrix (Λ)?

A
  1. Symmetric: The precision matrix is always symmetric.
  2. Inverse of the covariance matrix: It provides information about how tightly the variables are distributed around the mean.
19
Q

How are the covariance matrix and precision matrix related?

A

The covariance matrix describes the spread of the data, while the precision matrix (its inverse) describes how concentrated the distribution is around the mean.

20
Q

What is the Mahalanobis distance (Δ) in the Gaussian distribution?

A
  • The Mahalanobis distance (Δ) measures the distance from a point x to the mean
    μ, taking into account the spread and orientation of the data described by the covariance matrix Σ: Δ^2
    = (x−μ) ^T Σ^−1 (x−μ)
  • It reduces to the Euclidean distance when Σ is the identity matrix.
21
Q

What is “completing the square” in the context of Gaussian distributions?

A

Completing the square is a method used to rewrite a quadratic form in the exponent of a Gaussian distribution to identify the corresponding mean (μ) and covariance (Σ).

22
Q

What is the general form of the exponent in a multivariate Gaussian distribution?

A

(-1/2) (x−μ)^T Σ−1 (x−μ)

23
Q

How can the mean and covariance be derived by completing the square?

A
  • The second-order term in x corresponds to Σ^−1, allowing the covariance matrix to be identified as
    Σ (inverting the matrix)
  • The coefficient of the linear term in x corresponds to
    Σ^−1μ, from which the mean μ can be obtained.
24
Q

What does the term “const” refer to in the completed square expression?

A

Terms that are independent of x, which do not affect the form of the Gaussian distribution but may influence normalization factors.

25
Q

Why is completing the square useful in machine learning?

A

Completing the square is useful for deriving the parameters of Gaussian distributions in probabilistic models, such as in Bayesian inference and maximum likelihood estimation.