L2 - The Normal and Associated Distributions Flashcards
If X is a continuous random variable, What is P(X=1)?
- An accurate (but unhelpful) answer to this question is P(X=1)=0. In fact 𝑃(𝑋=𝑎)=0 at any given a within the interval. - Why? - as it is just one point we cannot calculate the area - it is an infinitesimally small point with such as infinitesimally small area which we say is 0
∫_-∞^∞f(x)=1 (total probability)
What is a continuous random variable?
- For example, we might wish to measure the temperature in a particular location over a period of time.
- Alternatively, we might wish to measure the distance between the place of residence and the place of work for an individual.
- In both these cases the random variable is more
naturally thought of as lying somewhere on a continuum of possible values rather than taking one of a discrete number of possibilities. - When the number of outcomes for a discrete distribution is large
then we can often approximate it by a continuous distribution.
What criteria must you satisfy to become a probability density function?
PDF example –> p(a ≤ X ≤ b) = ∫_a^b f(x) dx
- f(x) ≥ 0
- ∫_-∞^∞f(x)dx = 1
What is the z transformation or standardising a normal distribution?
- It is possible to transform any normal distribution into the standard normal distribution (mean = 0, standard deviation = 1) as follows:
- X~N(μ,σ^2)
- Z=(X-μ)/σ ~ N(0,1)
- This is useful because we can take data from different sources onto the same scale and only have to tabulate the standard normal distribution to be able to look up critical values and/or p-values for test statistics.
What is the function normal distribution?
- f(x) = (1/σsqrt(2π)) * exp[-((x-μ)^2)/(2σ^2))]
What is a useful deature of the normal distribution?
- is that linear combinations of normally distributed random variables will themselves follow a normal distribution
- For example, let X{2} ~N(μ{1},σ{1}^2) and X{2}~N(μ{2},σ{2}^2) be independent normal random variables
- If a and b are constants then a linear combination of the variables using a and b as weights has the following normal distribution:
- aX{1} + bX{2} ~ N(aμ{1}+bμ{2}, a^2σ{1}^2 + b^2σ{2}^2)
- The normal distribution is unique in having this property and therefore, if we can assume normality, this is very useful in deriving the distribution of random variables which are functions of other random variables
What are moments of distribution?
- are the expectations of integer powers of the random
variable in question - For example, if X is a random variable, then its first three moments are E(X), E(X^2) and E(X^3) .
- These are the raw moments of the distribution
What are central moments?
- the central moments which are the expectations of the deviation of the random variable from its mean (or first moment).
- Thus the second central moment of the random variable X can be written as E(X -E(X))^2= σ^2 which is the variance of x.
What are some examples of higher order moments?
Higher order moments are often scaled by the standard deviation to obtain measures such as:
- skewness –> E(X - E(X))^3/σ^3
- kurotosis –> E(X - E(X))^4/σ^4
- These measures are useful in characterising the shape of a distribution and are often referred to as the moments of the distribution even though, strictly speaking, they are transformations of the raw moments.
What is the mean function for continuous distribution?
μ= E(X)= ∫_-∞^∞f(x)x dx
What is the variance function for continuous distribution?
- σ^2= E(X-E(X))^2= E(X^2- E(X)^2) = ∫_-∞^∞(x-μ)^2 *f(x) dx
How do you find out the function of higher order moments for continous distribution?
- can be calulcuated by integrating a function of the form
- ∫_a^bf(x) (x-E(x))^k *f(x) dx (where a and b are the minimum and maximum possible values) and scaling by σ^k
What is the Chi-squared distribution?
- Is used to test the ‘goodness of fit’ a theoretical model is to a observed one
- so looking at the variance of the residual error (residual sum of squares) or a regression model for the actual data
- Also looking at the probability you could get those errors while holding some variables constant.
- The Chi-squared is derived or sampled from normal distribution with the formula - if a variable is Z is distributed normally it is said that Z^2 will have a Chi-squared distribution :
- χ = Σ_j=1^k (Z_j^2)
- also calculated:
χ = Σ(O-E)^2/E - Z is the risdual of a model
- the random variable defined by this is said to follow a chi-squared distribution with k degrees of freedom.
- When looking up on a table the P-value is the probability of it be larger than that value
- used when one variable depends on another when hypothesis test??? e.g. amount of women and men in a sample depends on their ages?
What does the Chi-squared distribution look like on a graph?
- if degrees of freedom (k) = 0, 1 or 2 –> chi squared no longer has a PDF which takes the value 0 at x=0 –> instead the value of the PDG tends to infinity as x(chi) tends to zero –> look like 1/x graph for k=1 and is downwards sloping for k=2
- if k > 2 –> The PDF takes the value 0 for x=0, reaches a single peak for some value of x >0 and declines asymptotically to 0 as x becomes large. - it is positively skewed.
- you get a negative gradient line at the start because as you are squaring all values of a standard normal distribution you are getting rid of all negative values, and at lower degrees of freedom this creates the negative gradient curve
What is the mean and variance for Chi-squared distribution?
mean –> k
variance –> 2k