Midterm Flashcards

1
Q

Independence

A

If Pr(A & B) = Pr(A)*Pr(B) OR Pr(A | B) = Pr(A)
- Determine if variables are correlated
- Avoid biased estimates
- Investigate autocorrelation, which could lead to inefficiency and incorrect standard errors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Conditional Expectation

A
  • Calculating conditional Variance
  • Law of iterated expectations
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Law of Iterated Expectations

A

E[E[g(x,y)|x1, x2]|x1] = E[g(x,y)|x1]
- Essential for deriving unbiased estimators
- How information affects predictions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Mean Independence

A

E[Y|X]=E[Y]
- E[u|X]=0
- Helps to justify assumptions in instrumental variable regression

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Coefficient of Determination (R^2)

A

= SST / SSE
The proportion of the variance in the dependent variable that is captured by the model.
- Predictive Power
- Comparing models

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Leverage Point

A

Observation whose explanatory variables have the potential to exert an unusually strong effect on the fitted model.
- Can disproportionately influence the results of a regression model.
- Model stability
- Important to detect to ensure robust and reliable regression analysis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Frisch-Waugh Theorem

A

Simplifies a MLR model by allowing it to focus on the impact of a single variable, holding others constant.

Beta_1 = (X1’ M2 X1)^-1 X1’ M2 y

  • Useful when dealing with panel data sets and individual effects.
  • How do individual variables contribute to the regression model?
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Positive Definite

A

If x’Ax >0 for all x != 0
- Ensures OLS estimates have a unique and stable solution
- Guarantees the Var-Cov matrix is invertible
- Ensures MLE reach a unique minimum

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Nonnegative Definite

A

If x’Ax >= 0 for all x != 0
- Var-Cov matrix must be non negative to make meaningful statistical inferences

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Square Root Matrix

A

R is a square root matrix if A = RR’
- Used in Monte Carlo
- Essential for Generalized Least Squares

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Level Curves

A

Sets of points where a function takes the same value.
c in R{x: x’Ax = c}
- Visualizing constraints and trade-offs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Normal Distribution

A

f(y) = 1 / [2 sqrt(2pi o^2)] exp(- (y - mu)^2/(2o^2))
- Central Limit Theorem
- Standard statistical tests
- Symmetry and known properties make it easy to deal with

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Multivariate Normal Distribution

A

If it can be expressed as a non-stochastic linear transformation of a vector of independent standard normal random variables.
- Modeling relationships between multiple dependent variables (time series)
- Enables factor analysis and multivariate hypothesis testing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Student’s t-Distribution

A

t(m) = z / sqrt(chi^2(m) / m)
- As sample size increases, it approaches the normal distribution
- Useful for small and large datasets
- Central to testing the significance of individual coefficients in regression analysis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Snedecor’s F-Distribution

A

If the ratio can be expressed as the ratio of two independent chi^2 random variables each divided by its degrees of freedom.
- Comparing variances across groups
- Testing the significance of a regression model
- Test restrictions on multiple regression coefficients
- Testing if nested models provide a significantly worse fit than a more complex model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Neyman-Pearson Lemma

A

The critical region should not contain any outcomes that are relatively more likely under the null than under the alternative.
- Likelihood Ratio Tests
- Balances the tradeoff between Type I and Type II errors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Unbiased Test

A

The probability of rejection under each point in the null is always less than the probability of rejection under each point in the alternative.
- Provides a fair test that does not favour the null or alternative hypothesis.

16
Q

Restricted Least Squares Estimator

A

Incorporates constraints into the regression model.
- Testing hypotheses about relationships between variables
- Improves the efficiency of estimates

17
Q

Consistent Estimator

A

If the estimator converges in probability to the true value of the parameter.
- Essential for reliable long-term inference
- With sufficient data, the estimator will yield results close to the actual population parameter.

18
Q

Law of Large Numbers

A

Sample averages converge to their population means.
- Provides the foundation for the consistency of estimators.

19
Q

Asymptotics

A

The behaviour of estimators and test statistics as the sample size approaches infinity.

20
Q

Independent and Identically Distributed

A

Random variables that are independent of each other and drawn from the same probability distribution.
- OLS, MLE, LLN
- Simplifies assumptions -> Makes it easier to derive consistent and unbiased estimators.

21
Q

Continuous Mapping Theorem

A

plim g(x) = g(plim x)
- Allows the application of continuous functions to converging sequences of random variables while preserving convergence

22
Q

Convergence in Distribution

A

Let z be a sequence of random variables distributed by Fn. If Fn -> F, then we say z converges in distribution to F
- Analyzing the behaviour of r.v.’s as the sample size increases.
- Ensures the distribution of an estimator or test statistic approaches a limiting distribution
- Fundamental for Central Limit Theorem.

23
Q

Wald Test

A

Testing the significance of one or more coefficients in a regression model.
- Widely used in Generalized Linear Models and MLE

24
Q

Maximum Likelihood Estimator

A

The value of the parameter vector that maximizes the Likelihood Ratio Test.
- Produces efficient and consistent estimators.
- Central to Generalized Linear Models
- Incorporates non-linear relationships

25
Q

Efficient Score Test (the Lagrange Multiplier)

A

Tells us how the estimation changes if we relax the constraint.
- Useful when the unrestricted model is complex and difficult to estimate
- Effective for detecting small deviations from the null hypothesis

26
Q

Information Criteria

A

Balances model fit with model complexity
- Choosing the best model
- Time series

27
Q

Akaike (AIC)

A

Penalizes the inclusion of additional parameters
- Prevents overfitting
- Time series

28
Q

Bayes/Schwartz BIC

A

Imposes a stronger penalty for the number of parameters than AIC
- More conservative than AIC

29
Q

Rolling Binned Means Estimator

A
  • Smooths time series data by calculating averages over sliding windows
  • Reduces noise
  • Mitigates the effects of outliers
30
Q

Kernel

A

Weigh observations based on their distance from a target point.
- Bias-variance trade off

31
Q

Curse of Dimensionality

A

The exponential increase in complexity and data requirements as the number of variables in a model grows

32
Q

Ridge Regression

A

Adds a penalty proportional to X’X which helps to shrink the coefficients towards zero without making them zero.
- Used to address multicollinearity
- Balances bias and variance
- Prevents overfitting

33
Q

Best Subset Selection

A

Identifies the best combination of predictors from a larger set of potential explanatory variables.
- Related to Information Criteria

34
Q

LASSO

A

Some coefficients are set to zero, while others are shrunk towards the origin.
- Improves model prediction and interpretability
- Balances model complexity and predictive accuracy

35
Q

Training and Validation Set

A

Sample is randomly divided into two parts: A training set and a validation set. Fit the model with the training set and use the validation set to compute MSE.
- Tests a model on unseen data.

36
Q

Leave one out Cross Validation

A

Train the model on all but one observation. Test it on the remaining observation. Repeat for each observation. Average the results.
- Unbiased estimate of model performance.
- Useful for small datasets.

37
Q

K-Fold Cross Validation

A

Divide into K groups. Treat each fold as the validation sample and average the MSE across the K groups.
- Works well for small and large datasets.
- More robust performance estimate

38
Q

Sieve Estimation

A

A linear combination of a family of functions that can be used to approximate m(x) arbitrarily well.
- Uses a sequence of simpler models instead of a fixed functional form.
- Avoids overfitting
- As the number of observations grows, the sieve converges to the true model.

39
Q

Neural Nets

A

Machine Learning Model
- Used for modeling complex non-linear relationships
- Flexibility and High Predictive Power

40
Q

Panel Data Fixed Effects

A

Controls for unobserved heterogeneity when this heterogeneity is constant over time and correlated with independent variables.
- Controls for time invariant factors