Glossary Flashcards
Random variable
A variable whose value depends on the outcome of a random experiment
Support (of a RV)
Set of possible values a RV can take
Probability mass/density function
Each point on the line represents the probability of getting that value
PMF = discrete, e.g. how many heads in 10 tosses of a coin
PDF = continuous, e.g. weight of females in California aged 18-25
fx(x) = P(X=x)
Cumulative distribution function
Integrate the PDF to get the CDF
This is cumulative, so the slope is always upward-sloping, and at each value of x, the corresponding value shows the probability of getting up to that point
Fx(x) = P(X≤x)
Variance
Distance from the mean

Statistic
A single measure of some attribute of a sample, calculated by applying a function to the set of data.
The function itself is independent of the sample’s distribution; that is, the function can be stated before realization of the data.
The mean of a population is not a statistic (no RV), but the mean of a sample is (sample variables chosen randomly)
Uniform distribution
Continuous RV between a and b - same probability of getting every point between a and b
Mean = (a+b)/2
Bernoulli distribution
Either take on the value of “p” or “1-p”
Support {0,1}
Population mean = p
Population variance = p(1-p)
Sample mean = X¯
Sample variance = n/n-1 X¯ ( 1 - X¯ )
Binomial distribution
Probability of something happening over a period of time - X is the number of successes in n independent Bernouilli trials
X~B(n,p)
E(X) = np
CLT: binomial is discrete, but as you do more trials, it starts to look continuous: as n→∞, B ~ N ( μ , σ2/n )
Normal distribution (skewness and kurtosis)
Averages more common than extremes, bell-shaped diagram
Skewness = measure of symmetry Kurtosis = fatness in tails
*z = X-μ/σ *or z = X-μ^/SE(μ^) with samples
Then use the tables: Φ(z) = P(X ≤ z)
Jensen’s inequality
If g(.) is concave, then E[g(X)] < g(E[X])
E.g. mean of the log < log of the mean
Marginal probability
Marginal probability = the probability of an event occurring, p(A)
The probability that A=1, regardless of the value of B
Joint probability
Joint probability = p(A and B), the probability of event A and event B occurring
So, the joint distribution is the set of probabilities of possible pairs of values
Conditional probability
p(A|B), the probability of event A occurring, given that event B occurs
Conditional mean
Mean of the conditional distribution → E[Y|X=x]
Law of iterated expectations
The mean of the conditional means is the mean → E[Y] = E[E[Y|X]]
E.g., suppose we are interested in average IQ generally, but we have measures of average IQ by gender. We could figure out the quantity of interest by weighting average IQ by the relative proportions of men and women
Covariance
Do X and Y vary together?
σ(X,Y) = E[(X - E[X])*(Y - E[X])] = E[XY] - E[X]E[Y]
Correlation
Measures only linear relationship - may be 0 is the relationship is perfect yet non-linear
ρX,Y = cov(X,Y) / σXσY
(σ are standard deviations)
E[aX + bY] =
Var[aX + bY] =
= aE[X] + bE[Y]
= a2Var[X] + b2Var[Y] + 2abCov(X,Y)
Random sample
When a random experiment is repeated n times, we obtain n independent identically distributed (IID) random variables
Drawing one person makes another no more likely, drawn from the same population
Sample mean
The mean of a subset of the population (and a random variable)
xbar = 1/n * Σxi
The larger the sample, the smaller the variance of the sample mean.
The sample mean is an unbiased estimator of the population mean (µ)
Law of large numbers
If Ynare IID, the sample mean converges in probability to the population mean as the sample size grows
As n → ∞, X¯ - μ → 0
Central limit theorem
If Yn are IID with mean µ and variance σ2 and n is large, then the sample mean (Ybar) is approximately normally distributed, with mean µ and variance σ2/n
Sample variance
s2 = 1/n-1 ∑(Xi - X¯)2
We have lost one degree of freedom: if you have n-1 numbers and the mean, you can work out the last number. The last number is not independent - we only have n-1 independent observations
The sample variance is an unbiased estimator of population variance - PROOF





