Econometrics Flashcards

1
Q

Acceptance region

A

The set of values of a test statistic for which the null hypothesis is accepted (is not rejected).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Adjusted R2( )

A

A modified version of R2 that does not necessarily increase when a new regressor is added to the regression.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

ADL(p,q)

A

See autoregressive distributed lag model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

AIC

A

See information criterion.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Akaike information criterion

A

See information criterion.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Alternative hypothesis

A

The hypothesis that is assumed to be true if the null hypothesis is false. The alternative hypothesis is often denoted H1.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

AR(p)

A

See autoregression.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

ARCH

A

See autoregressive conditional heteroskedasticity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Asymptotic distribution

A

The approximate sampling distribution of a random variable computed using a large sample. For example, the asymptotic distribution of the sample average is normal.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Asymptotic normal distribution

A

A normal distribution that approximates the sampling distribution of a statistic computed using a large sample.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Attrition

A

The loss of subjects from a study after assignment to the treatment or control group.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Augmented Dickey-Fuller (ADF) test

A

A regressionbased test for a unit root in an AR(p) model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Autocorrelation

A

The correlation between a time series variable and its lagged value.The jth autocorrelation of Y is the correlation between Yt and Yt2j.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Autocovariance

A

The covariance between a time series variable and its lagged value.The jth autocovariance of Y is the covariance between Yt and Yt2j.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Autoregression

A

A linear regression model that relates a time series variable to its past (that is, lagged) values. An autoregression with p lagged values as regressors is denoted AR(p).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Autoregressive conditional heteroskedasticity (ARCH)

A

A time series model of conditional heteroskedasticity. R2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

Autoregressive distributed lag model

A

A linear regression model in which the time series variable Yt is expressed as a function of lags of Yt and of another variable, Xt.The model is denoted ADL(p,q), where p denotes the number of lags of Yt and q denotes the number of lags of Xt.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Average causal effect

A

The population average of the individual causal effects in a heterogeneous population. Also called the average treatment effect.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Balanced panel

A

A panel data set with no missing observations, that is, in which the variables are observed for each entity and each time period.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

Base specification

A

A baseline or benchmark regression specification that includes a set of regressors chosen using a combination of expert judgment, economic theory, and knowledge of how the data were collected.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Bayes information criterion

A

See information criterion.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Bernoulli distribution

A

The probability distribution of a Bernoulli random variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Bernoulli random variable

A

A random variable that takes on two values, 0 and 1.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Best linear unbiased estimator

A

An estimator that has the smallest variance of any estimator that is a linear function of the sample values Y and is unbiased. Under the Gauss-Markov conditions, the OLS estimator is the best linear unbiased estimator of the regression coefficients conditional on the values of the regressors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

Bias

A

The expected value of the difference between an estimator and the parameter that it is estimating. If is an estimator of mY, then the bias of is E( )2 mY.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

BIC

A

See information criterion.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

Binary variable

A

A variable that is either 0 or 1.A binary variable is used to indicate a binary outcome. For example,X is a binary (or indicator, or dummy) variable for a person’s gender if X 5 1 if the person is female and X 5 0 if the person is male. mˆY mˆY mˆY

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

Bivariate normal distribution

A

A generalization of the normal distribution to describe the joint distribution of two random variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

BLUE

A

See best linear unbiased estimator.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

Break date

A

The date of a discrete change in population time series regression coefficient(s).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

Causal effect

A

The expected effect of a given intervention or treatment as measured in an ideal randomized controlled experiment.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

Central limit theorem

A

A result in mathematical statistics that says that, under general conditions, the sampling distribution of the standardized sample average is well approximated by a standard normal distribution when the sample size is large.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

Chi-squared distribution

A

The distribution of the sum of m squared independent standard normal random variables.The parameter m is called the degrees of the freedom of the chi-squared distribution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

Chow test

A

A test for a break in a time series regression at a known break date.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

Coefficient of determination

A

See R2.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

Cointegration

A

When two or more time series variables share a common stochastic trend.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

Common trend

A

A trend shared by two or more time series.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
38
Q

Conditional distribution

A

The probability distribution of one random variable given that another random variable takes on a particular value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
39
Q

Conditional expectation

A

The expected value of one random value given that another random variable takes on a particular value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
40
Q

Conditional heteroskedasticity

A

The variance, usually of an error term, depends on other variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
41
Q

Conditional mean

A

The mean of a conditional distribution; see conditional expectation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
42
Q

Conditional mean independence

A

The conditional expectation of the regression error ui, given the regressors, depends on some but not all of the regressors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
43
Q

Conditional variance

A

The variance of a conditional distribution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
44
Q

Confidence interval (or confidence set)

A

An interval (or set) that contains the true value of a population parameter with a prespecified probability when computed over repeated samples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
45
Q

Confidence level

A

The prespecified probability that a confidence interval (or set) contains the true value of the parameter.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
46
Q

Consistency

A

Means that an estimator is consistent. See consistent estimator.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
47
Q

Consistent estimator

A

An estimator that converges in probability to the parameter that it is estimating.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
48
Q

Constant regressor

A

The regressor associated with the regression intercept; this regressor is always equal to 1.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
49
Q

Constant term

A

The regression intercept.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
50
Q

Continuous random variable

A

A random variable that can take on a continuum of values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
51
Q

Control group

A

The group that does not receive the treatment or intervention in an experiment.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
52
Q

Control variable

A

Another term for a regressor; more specifically, a regressor that controls for one of the factors that determine the dependent variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
53
Q

Convergence in distribution

A

When a sequence of distributions converges to a limit; a precise definition is given in Section 17.2.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
54
Q

Convergence in probability

A

When a sequence of random variables converges to a specific value; for example, when the sample average becomes close to the population mean as the sample size increases; see Key Concept 2.6 and Section 17.2.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
55
Q

Correlation

A

A unit-free measure of the extent to which two random variables move, or vary, together.The correlation (or correlation coefficient) between X and Y is sXY/sXsY and is denoted corr(X,Y).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
56
Q

Correlation coefficient

A

See correlation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
57
Q

Covariance

A

A measure of the extent to which two random variables move together.The covariance between X and Y is the expected value E[(X 2 mX)(Y 2 mY)], and is denoted by cov(X,Y) or by sXY.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
58
Q

Covariance matrix

A

A matrix composed of the variances and covariances of a vector of random variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
59
Q

Critical value

A

The value of a test statistic for which the test just rejects the null hypothesis at the given significance level.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
60
Q

Cross-sectional data

A

Data collected for different entities in a single time period.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
61
Q

Cubic regression model

A

A nonlinear regression function that includes X, X2, and X3 as regressors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
62
Q

Cumulative distribution function (c.d.f.)

A

See cumulative probability distribution.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
63
Q

Cumulative dynamic multiplier

A

The cumulative effect of a unit change in the time series variable X on Y.The h-period cumulative dynamic multiplier is the effect of a unit change in Xt on Yt + Yt+1+ . . . + Yt+h.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
64
Q

Cumulative probability distribution

A

A function showing the probability that a random variable is less than or equal to a given number.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
65
Q

Dependent variable

A

The variable to be explained in a regression or other statistical model; the variable appearing on the left-hand side in a regression.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
66
Q

Deterministic trend

A

A persistent long-term movement of a variable over time that can be represented as a nonrandom function of time.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
67
Q

Dickey-Fuller test

A

A method for testing for a unit root in a first order autoregression [AR(1)].

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
68
Q

Differences estimator

A

An estimator of the causal effect constructed as the difference in the sample average outcomes between the treatment and control groups.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
69
Q

Differences-in-differences estimator

A

The average change in Y for those in the treatment group, minus the average change in Y for those in the control group.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
70
Q

Discrete random variable

A

A random variable that takes on discrete values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
71
Q

Distributed lag model

A

A regression model in which the regressors are current and lagged values of X.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
72
Q

Dummy variable

A

See binary variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
73
Q

Dummy variable trap

A

A problem caused by including a full set of binary variables in a regression together with a constant regressor (intercept), leading to perfect multicollinearity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
74
Q

Dynamic causal effect

A

The causal effect of one variable on current and future values of another variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
75
Q

Dynamic multiplier

A

The h-period dynamic multiplier is the effect of a unit change in the time series variable Xt on Yt+h.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
76
Q

Endogenous variable

A

A variable that is correlated with the error term.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
77
Q

Error term

A

The difference between Y and the population regression function, denoted by u in this textbook.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
78
Q

Errors-in-variables bias

A

The bias in an estimator of a regression coefficient that arises from measurement errors in the regressors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
79
Q

Estimate

A

The numerical value of an estimator computed from data in a specific sample.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
80
Q

Estimator

A

A function of a sample of data to be drawn randomly from a population. An estimator is a procedure for using sample data to compute an educated guess of the value of a population parameter, such as the population mean.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
81
Q

Exact distribution

A

The exact probability distribution of a random variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
82
Q

Exact identification

A

When the number of instrumental variables equals the number of endogenous regressors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
83
Q

Exogenous variable

A

A variable that is uncorrelated with the regression error term.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
84
Q

Expected value

A

The long-run average value of a random variable over many repeated trials or occurrences. It is the probability-weighted average of all possible values that the random variable can take on.The expected value of Y is denoted E(Y) and is also called the expectation of Y.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
85
Q

Experimental data

A

Data obtained from an experiment designed to evaluate a treatment or policy or to investigate a causal effect.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
86
Q

Experimental effect

A

When experimental subjects change their behavior because they are part of an experiment.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
87
Q

Explained sum of squares (ESS)

A

The sum of squared deviations of the predicted values of Yi, ,from their average; see Equation (4.14).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
88
Q

Explanatory variable

A

See regressor.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
89
Q

External validity

A

Inferences and conclusions from a statistical study are externally valid if they can be generalized from the population and the setting studied to other populations and settings.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
90
Q

F-statistic

A

A statistic used to a test joint hypothesis concerning more than one of the regression coefficients.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
91
Q

Fm,n distribution

A

The distribution of a ratio of independent random variables, where the numerator is a chi-squared random variable with m degrees of freedom, divided by m, and the denominator is a chi-squared random variable with n degrees of freedom divided by n.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
92
Q

Fm,∞ distribution

A

The distribution of a random variable with a chi-squared distribution with m degrees of freedom, divided by m.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
93
Q

Feasible GLS

A

A version of the generalized least squares (GLS) estimator that uses an estimator of the conditional variance of the regression errors and covariance between the regression errors at different observations.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
94
Q

Feasible WLS

A

A version of the weighted least squares (WLS) estimator that uses an estimator of the conditional variance of the regression errors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
95
Q

First difference

A

The first difference of a time series variable Yt is Yt 2 Yt21, denoted DYt.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
96
Q

First-stage regression

A

The regression of an included endogenous variable on the included exogenous variables, if any, and the instrumental variable(s) in two stage least squares.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
97
Q

Fitted values

A

See predicted values.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
98
Q

Fixed effects

A

Binary variables indicating the entity or time period in a panel data regression.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
99
Q

Fixed effects regression model

A

A panel data regression that includes entity fixed effects. ˆYi

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
100
Q

Forecast error

A

The difference between the value of the variable that actually occurs and its forecasted value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
101
Q

Forecast interval

A

An interval that contains the future value of a time series variable with a prespecified probability.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
102
Q

Functional form misspecification

A

When the form of the estimated regression function does not match the form of the population regression function; for example, when a linear specification is used but the true population regression function is quadratic.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
103
Q

GARCH

A

See generalized autoregressive conditional heteroskedasticity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
104
Q

Gauss-Markov theorem

A

Mathematical result stating that, under certain conditions, the OLS estimator is the best linear unbiased estimator of the regression coefficients conditional on the values of the regressors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
105
Q

Generalized autoregressive conditional heteroskedasticity

A

A time series model for conditional heteroskedasticity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
106
Q

Generalized least squares (GLS)

A

A generalization of OLS that is appropriate when the regression errors have a known form of heteroskedasticity (in which case GLS is also referred to as weighted least squares, WLS) or a known form of serial correlation.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
107
Q

Generalized method of moments

A

A method for estimating parameters by fitting sample moments to population moments that are functions of the unknown parameters. Instrumental variables estimators are an important special case.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
108
Q

GMM

A

See generalized method of moments.

109
Q

Granger causality test

A

A procedure for testing whether current and lagged values of one time series help predict future values of another time series.

110
Q

HAC standard errors

A

See heteroskedasticity- and autocorrelation-consistent (HAC) standard errors.

111
Q

Hawthorne effect

A

See experimental effect.

112
Q

Heteroskedasticity

A

The situation in which the variance of the regression error term ui, conditional on the regressors, is not constant. Heteroskedasticity- and autocorrelation-consistent

113
Q

(HAC) standard errors

A

Standard errors for OLS estimators that are consistent whether or not the regression errors are heteroskedastic and autocorrelated.

114
Q

Heteroskedasticity-robust standard error

A

Standard errors for the OLS estimator that are appropriate whether the error term is homoskedastic or heteroskedastic.

115
Q

Heteroskedasticity-robust t-statistic

A

A t-statistic constructed using a heteroskedasticity-robust standard error.

116
Q

Homoskedasticity

A

The variance of the error term ui, conditional on the regressors, is constant.

117
Q

Homoskedasticity-only F statistic

A

A form of the Fstatistic that is valid only when the regression errors are homoskedastic.

118
Q

Homoskedasticity-only standard errors

A

Standard errors for the OLS estimator that are appropriate only when the error term is homoskedastic.

119
Q

Hypothesis test

A

A procedure for using sample evidence to help determine if a specific hypothesis about a population is true or false.

120
Q

i.i.d.

A

Independently and indentically distributed.

121
Q

Identically distributed

A

When two or more random variables have the same distribution.

122
Q

Impact effect

A

The contemporaneous, or immediate, effect of a unit change in the time series variable Xt on Yt.

123
Q

Imperfect multicollinearity

A

The condition in which two or more regressors are highly correlated.

124
Q

Included endogenous variables

A

Regressors that are correlated with the error term (usually in the context of instrumental variable regression).

125
Q

Included exogenous variables

A

Regressors that are uncorrelated with the error term (usually in the context of instrumental variable regression).

126
Q

Independence

A

When knowing the value of one random variable provides no information about the value of another random variable.Two random variables are independent if their joint distribution is the product of their marginal distributions.

127
Q

Indicator variable

A

See binary variable.

128
Q

Information criterion

A

A statistic used to estimate the number of lagged variables to include in an autoregression or a distributed lag model. Leading examples are the Akaike information criterion (AIC) and the Bayes information criterion (BIC).

129
Q

Instrument

A

See instrumental variable.

130
Q

Instrumental variable

A

A variable that is correlated with an endogenous regressor (instrument relevance) and is uncorrelated with the regression error (instrument exogeneity).

131
Q

Instrumental variables (IV) regression

A

A way to obtain a consistent estimator of the unknown coefficients of the population regression function when the regressor,X, is correlated with the error term, u.

132
Q

Interaction term

A

A regressor that is formed as the product of two other regressors, such as X1i 3 X2i.

133
Q

Intercept

A

The value of b0 in the linear regression model.

134
Q

Internal validity

A

When inferences about causal effects in a statistical study are valid for the population being studied.

135
Q

J-statistic

A

Astatistic for testing overidentifying restrictions in instrumental variables regression.

136
Q

Joint hypothesis

A

A hypothesis consisting of two or more individual hypotheses, that is, involving more than one restriction on the parameters of a model.

137
Q

Joint probability distribution

A

The probability distribution determining the probabilities of outcomes involving two or more random variables.

138
Q

Kurtosis

A

A measure of how much mass is contained in the tails of a probability distribution.

139
Q

Lags

A

The value of a time series variable in a previous time period.The jth lag of Yt is Yt2j.

140
Q

Law of iterated expectations

A

A result in probability theory that says that the expected value of Y is the expected value of its conditional expectation given X, that is, E(Y) 5 E[E(Y X)].

141
Q

Law of large numbers

A

According to this result from probability theory, under general conditions the sample average will be close to the population mean with very high probability when the sample size is large.

142
Q

Least squares assumptions

A

The assumptions for the linear regression model listed in Key Concept 4.3 (single variable regression) and Key Concept 6.4 (multiple regression model).

143
Q

Least squares estimator

A

An estimator formed by minimizing the sum of squared residuals.

144
Q

Limited dependent variable

A

A dependent variable that can take on only a limited set of values. For example, the variable might be a 021 binary variable or arise from one of the models described in Appendix 11.3.

145
Q

Linear-log model

A

A nonlinear regression function in which the dependent variable is Y and the independent variable is ln(X).

146
Q

Linear probability model

A

A regression model in which Y is a binary variable.

147
Q

Linear regression function

A

A regression function with a constant slope.

148
Q

Local average treatment effect

A

A weighted average treatment effect estimated, for example, by TSLS.

149
Q

Log-linear model

A

A nonlinear regression function in which the dependent variable is ln(Y) and the independent variable is X.

150
Q

Log-log model

A

A nonlinear regression function in which the dependent variable is ln(Y) and the independent variable is ln(X). @

151
Q

Logarithm

A

A mathematical function defined for a positive argument; its slope is always positive but tends to zero.The natural logarithm is the inverse of the exponential function, that is, X 5 ln(eX).

152
Q

Logit regression

A

A nonlinear regression model for a binary dependent variable in which the population regression function is modeled using the cumulative logistic distribution function.

153
Q

Long-run cumulative dynamic multiplier

A

The cumulative long-run effect on the time series variable Y of a change in X.

154
Q

Longitudinal data

A

See panel data.

155
Q

Marginal probability distribution

A

Another name for the probability distribution of a random variable Y, which distinguishes the distribution of Y alone (the marginal distribution) from the joint distribution of Y and another random variable.

156
Q

Maximum likelihood estimator (MLE)

A

An estimator of unknown parameters that is obtained by maximizing the likelihood function; see Appendix 11.2.

157
Q

Mean

A

The expected value of a random variable.The mean of Y is denoted mY.

158
Q

Moments of a distribution

A

The expected value of a random variable raised to different powers.The rth moment of the random variable Y is E(Yr).

159
Q

Multicollinearity

A

See perfect multicollinearity and imperfect multicollinearity.

160
Q

Multiple regression model

A

An extension of the single variable regression model that allows Y to depend on k regressors.

161
Q

Natural experiment

A

See quasi-experiment.

162
Q

Natural logarithm

A

See logarithm.

163
Q

95% confidence set

A

A confidence set with a 95% confidence level; see confidence interval.

164
Q

Nonlinear least squares

A

The analog of OLS that applies when the regression function is a nonlinear function of the unknown parameters.

165
Q

Nonlinear least squares estimator

A

The estimator obtained by minimizing the sum of squared residuals when the regression function is nonlinear in the parameters.

166
Q

Nonlinear regression function

A

A regression function with a slope that is not constant.

167
Q

Nonstationary

A

When the joint distribution of a time series variable and its lags changes over time.

168
Q

Normal distribution

A

A commonly used bell-shaped distribution of a continuous random variable.

169
Q

Null hypothesis

A

The hypothesis being tested in a hypothesis test, often denoted by H0.

170
Q

Observation number

A

The unique identifier assigned to each entity in a data set.

171
Q

Observational data

A

Data based on observing, or measuring, actual behavior outside an experimental setting. OLS estimator. See ordinary least squares estimator.

172
Q

OLS regression line

A

The regression line with population coefficients replaced by the OLS estimators.

173
Q

OLS residual

A

The difference between Yi and the OLS regression line, denoted by in this textbook.

174
Q

Omitted variables bias

A

The bias in an estimator that arises because a variable that is a determinant of Y and is correlated with a regressor has been omitted from the regression.

175
Q

One-sided alternative hypothesis

A

The parameter of interest is on one side of the value given by the null hypothesis.

176
Q

Order of integration

A

The number of times that a time series variable must be differenced to make it stationary.A time series variable that is integrated of order p must be differenced p times and is denoted I(p).

177
Q

Ordinary least squares estimator

A

The estimator of the regression intercept and slope(s) that minimizes the sum of squared residuals.

178
Q

Outlier

A

An exceptionally large or small value of a random variable.

179
Q

Overidentification

A

When the number of instrumental variables exceeds the number of included endogenous regressors.

180
Q

p-value

A

The probability of drawing a statistic at least as adverse to the null hypothesis as the one actually computed, assuming the null hypothesis is correct. Also called the marginal significance probability, the p-value is the smallest significance level at which the null hypothesis can be rejected.

181
Q

Panel data

A

Data for multiple entities where each entity is observed in two or more time periods.

182
Q

Parameter

A

A constant that determines a characteristic of a probability distribution or population regression function.

183
Q

Partial compliance

A

Occurs when some participants fail to follow the treatment protocol in a randomized experiment.

184
Q

Partial effect

A

The effect on Y of changing one of the regressors, holding the other regressors constant.

185
Q

Perfect multicollinearity

A

Occurs when one of the regressors is an exact linear function of the other regressors.

186
Q

Polynomial regression model

A

A nonlinear regression function that includes X, X2, . . . and Xr as regressors, where r is an integer. uˆi

187
Q

Population

A

The group of entities—such as people, companies, or school districts—being studied.

188
Q

Population coefficients

A

See population intercept and slope.

189
Q

Population intercept and slope

A

The true, or population, values of b0 (the intercept) and b1 (the slope) in a single variable regression. In a multiple regression, there are multiple slope coefficients (b1, b2, . . . , bk), one for each regressor.

190
Q

Population multiple regression model

A

The multiple regression model in Key Concept 6.2.

191
Q

Population regression line

A

In a single variable regression, the population regression line is b0 + b1Xi, and in a multiple regression it is b0 + b1X1i + b2X2i + . . . + bkXki.

192
Q

Power

A

The probability that a test correctly rejects the null hypothesis when the alternative is true.

193
Q

Predicted value

A

The value of Yi that is predicted by the OLS regression line, denoted by in this textbook.

194
Q

Price elasticity

A

The percentage change in the quantity demanded resulting from a 1% increase in price.

195
Q

Probability

A

The proportion of the time that an outcome (or event) will occur in the long run.

196
Q

Probability density function (p.d.f.)

A

For a continuous random variable, the area under the probability density function between any two points is the probability that the random variable falls between those two points.

197
Q

Probability distribution

A

For a discrete random variable, a list of all values that a random variable can take on and the probability associated with each of these values.

198
Q

Probit regression

A

A nonlinear regression model for a binary dependent variable in which the population regression function is modeled using the cumulative standard normal distribution function.

199
Q

Program evaluation

A

The field of study concerned with estimating the effect of a program, policy, or some other intervention or “treatment.”

200
Q

Pseudo out-of-sample forecast

A

A forecast computed over part of the sample using a procedure that is as if these sample data have not yet been realized.

201
Q

Quadratic regression model

A

A nonlinear regression function that includes X and X2 as regressors.

202
Q

Quasi-experiment

A

A circumstance in which randomness is introduced by variations in individual circumstances that make it appear as if the treatment is randomly assigned. ˆYi

203
Q

R2

A

In a regression, the fraction of the sample variance of the dependent variable that is explained by the regressors

204
Q

Random walk

A

A time series process in which the value of the variable equals its value in the previous period, plus an unpredictable error term.

205
Q

Random walk with drift

A

A generalization of the random walk in which the change in the variable has a nonzero mean but is otherwise unpredictable.

206
Q

Randomized controlled experiment

A

An experiment in which participants are randomly assigned to a control group, which receives no treatment, or to a treatment group, which receives a treatment.

207
Q

Regressand

A

See dependent variable.

208
Q

Regression specification

A

A description of a regression that includes the set of regressors and any nonlinear transformation that has been applied.

209
Q

Regressor

A

A variable appearing on the right-hand side of a regression; an independent variable in a regression.

210
Q

Rejection region

A

The set of values of a test statistic for which the test rejects the null hypothesis.

211
Q

Repeated cross-sectional data

A

A collection of crosssectional data sets, where each cross-sectional data set corresponds to a different time period.

212
Q

Restricted regression

A

Aregression in which the coefficients are restricted to satisfy some condition. For example, when computing the homoskedasticityonly F-statistic, this is the regression with coefficients restricted to satisfy the null hypothesis.

213
Q

Root mean squared forecast error

A

The square root of the mean of the squared forecast error.

214
Q

Sample correlation

A

An estimator of the correlation between two random variables.

215
Q

Sample covariance

A

An estimator of the covariance between two random variables.

216
Q

Sample selection bias

A

The bias in an estimator of a regression coefficient that arises when a selection process influences the availability of data and that process is related to the dependent variable.This induces correlation between one or more regressors and the regression error.

217
Q

Sample standard deviation

A

An estimator of the standard deviation of a random variable.

218
Q

Sample variance

A

An estimator of the variance of a random variable.

219
Q

Sampling distribution

A

The distribution of a statistic over all possible samples; the distribution arising from repeatedly evaluating the statistic using a R2 series of randomly drawn samples from the same population.

220
Q

Scatterplot

A

A plot of n observations on Xi and Yi, in which each observation is represented by the point (Xi,Yi).

221
Q

Serial correlation

A

See autocorrelation.

222
Q

Serially uncorrelated

A

A time series variable with all autocorrelations equal to zero.

223
Q

Significance level

A

The prespecified rejection probability of a statistical hypothesis test when the null hypothesis is true.

224
Q

Simple random sampling

A

When entities are chosen randomly from a population using a method that ensures that each entity is equally likely to be chosen.

225
Q

Simultaneous causality bias

A

When, in addition to the causal link of interest from X to Y, there is a causal link from Y to X. Simultaneous causality makes X correlated with the error term in the population regression of interest.

226
Q

Simultaneous equations bias

A

See simultaneous causality bias.

227
Q

Size of a test

A

The probability that a test incorrectly rejects the null hypothesis when the null hypothesis is true.

228
Q

Skewness

A

A measure of the aysmmetry of a probability distribution.

229
Q

Standard deviation

A

The square root of the variance. The standard deviation of the random variable Y, denoted sY, has the units of Y and is a measure of the spread of the distribution of Y around its mean.

230
Q

Standard error of an estimator

A

An estimator of the standard deviation of the estimator.

231
Q

Standard error of the regression (SER)

A

An estimator of the standard deviation of the regression error u.

232
Q

Standard normal distribution

A

The normal distribution with mean equal to 0 and variance equal to 1, denoted N(0, 1).

233
Q

Standardizing a random variable

A

An operation accomplished by subtracting the mean and dividing by the standard deviation, which produces a random variable with a mean of 0 and a standard deviation of 1.The standardized value of Y is (Y 2 mY)/sY.

234
Q

Stationarity

A

When the joint distribution of a time series variable and its lagged values does not change over time.

235
Q

Statistically insignificant

A

The null hypothesis (typically, that a regression coefficient is zero) cannot be rejected at a given significance level.

236
Q

Statistically significant

A

The null hypothesis (typically, that a regression coefficient is zero) is rejected at a given significance level.

237
Q

Stochastic trend

A

A persistent but random long-term movement of a variable over time.

238
Q

Strict exogeneity

A

The requirement that the regression error has a mean of zero conditional on current, future, and past values of the regressor in a distributed lag model.

239
Q

Student t distribution

A

The Student t distribution with m degrees of freedom is the distribution of the ratio of a standard normal random variable, divided by the square root of an independently distributed chi-squared random variable with m degrees of freedom divided by m.As m gets large, the Student t distribution converges to the standard normal distribution.

240
Q

Sum of squared residuals (SSR)

A

The sum of the squared OLS residuals.

241
Q

t-distribution

A

See Student t distribution.

242
Q

t-ratio

A

See t-statistic.

243
Q

t-statistic

A

A statistic used for hypothesis testing. See Key Concept 5.1.

244
Q

Test for a difference in means

A

A procedure for testing whether two populations have the same mean.

245
Q

Time effects

A

Binary variables indicating the time period in a panel data regression.

246
Q

Time and entity fixed effects regression model

A

A panel data regression that includes both entity fixed effects and time fixed effects.

247
Q

Time fixed effects

A

See time effects.

248
Q

Time series data

A

Data for the same entity for multiple time periods.

249
Q

Total sum of squares (TSS)

A

The sum of squared deviations of Yi, from its average, .

250
Q

Treatment effect

A

The causal effect in an experiment or a quasi-experiment; see causal effect.

251
Q

Treatment group

A

The group that receives the treatment or intervention in an experiment.

252
Q

TSLS

A

See two stage least squares. Y

253
Q

Two-sided alternative hypothesis

A

When, under the alternative hypothesis, the parameter of interest is not equal to the value given by the null hypothesis.

254
Q

Two stage least squares

A

An instrumental variable estimator, described in Key Concept 12.2.

255
Q

Type I error

A

In hypothesis testing, the error made when the null hypothesis is true but is rejected.

256
Q

Type II error

A

In hypothesis testing, the error made when the null hypothesis is false but is not rejected.

257
Q

Unbalanced panel

A

A panel data set in which some data are missing.

258
Q

Unbiased estimator

A

An estimator with a bias that is equal to zero.

259
Q

Uncorrelated

A

Two random variables are uncorrelated if their correlation is zero.

260
Q

Underidentification

A

When the number of instrumental variables is less than the number of endogenous regressors.

261
Q

Unit root

A

Refers to an autoregression with a largest root equal to 1.

262
Q

Unrestricted regression

A

When computing the homoskedasticity-only F-statistic, this is the regression that applies under the alternative hypothesis, so the coefficients are not restricted to satisfy the null hypothesis.

263
Q

VAR

A

See vector autoregression.

264
Q

Variance

A

The expected value of the squared difference between a random variable and its mean; the variance of Y is denoted .

265
Q

Vector autoregression

A

A model of k time series variables consisting of k equations, one for each variable, in which the regressors in all equations are lagged values of all the variables.

266
Q

Volatility clustering

A

When a time series variable exhibits some clustered periods of high variance and other clustered periods of low variance.

267
Q

Weak instruments

A

Instrumental variables that have a low correlation with the endogenous regressor(s).

268
Q

Weighted least squares (WLS)

A

An alternative to OLS that can be used when the regression error is heteroskedastic and the form of the heteroskedasticity is known or can be estimated.