Week 1 and Week 2 Flashcards
What is probability theory good for?
Learn about a large group:
large group = population
> Too large to look at everyone
• Look at small subgroups
> describe sample = descriptive statistics
• Infer properties of a population from a sample
> Inferential statistics
How to choose a smart sample?
Choose a random sample
How do we denote sample and realized sample?
Sample set: Ω
realized sample: ω_0
Denote Expectation of X
E[X]
How do we calculate variance? And why?
measures how accurately x is predicted by E[X]. Variance of x= var(x)=(x-E[X])^2
If we calculate Var(x) and it’s LARGE, is E[X] a good prediction of x(ω_0)?
No, E[X] is not a good prediction of x(ω_0). Only if variance is small.
How do we calculate standard deviation?
sd(x)=sqrt(var(x))
Optimality of E[X]
E[X]=arg min E[(x-a)^2]
What does it mean if:
Cov(Y_1,Y_2)<0
Cov(Y_1,Y_2)>0
Cov(Y_1,Y_2)=0
Cov(Y_1,Y_2)<0 = negative relationship, negatively correlated
Cov(Y_1,Y_2)>0= positive relationship, positively correlated
Cov(Y_1,Y_2)=0 = uncorrelated
Does a random sample give a representative sample?
Yes
What does k and K denote?
k = observed X’s
K = all of the X’s
“we cannot predict the value of U by observing regressors” is denoted how?
E[U|X1,…,Xk]=0
Assumption OLS-2 (exogeneity)
The linear regression model satisfies: E[U|X1,…,Xk]=0
The regressors are exogenous.
If E[U|X1,…,Xk]=0 holds, what does that say about the covariance?
cov(U,Xj)=0
- each regressor uncorrelated with unobserved component
- find j with cov (U, Xj) ≠ 0 and exogeneity fails
Assumption OLS-3
(Full rank, informal statement): The best linear prediction of Y is unique.
This assumption is often called the “no perfect collinearity assumption”