Econometrics Flashcards
Acceptance region
The set of values of a test statistic for which the null hypothesis is accepted (is not rejected).
Adjusted R2( )
A modified version of R2 that does not necessarily increase when a new regressor is added to the regression.
ADL(p,q)
See autoregressive distributed lag model.
AIC
See information criterion.
Akaike information criterion
See information criterion.
Alternative hypothesis
The hypothesis that is assumed to be true if the null hypothesis is false. The alternative hypothesis is often denoted H1.
AR(p)
See autoregression.
ARCH
See autoregressive conditional heteroskedasticity.
Asymptotic distribution
The approximate sampling distribution of a random variable computed using a large sample. For example, the asymptotic distribution of the sample average is normal.
Asymptotic normal distribution
A normal distribution that approximates the sampling distribution of a statistic computed using a large sample.
Attrition
The loss of subjects from a study after assignment to the treatment or control group.
Augmented Dickey-Fuller (ADF) test
A regressionbased test for a unit root in an AR(p) model.
Autocorrelation
The correlation between a time series variable and its lagged value.The jth autocorrelation of Y is the correlation between Yt and Yt2j.
Autocovariance
The covariance between a time series variable and its lagged value.The jth autocovariance of Y is the covariance between Yt and Yt2j.
Autoregression
A linear regression model that relates a time series variable to its past (that is, lagged) values. An autoregression with p lagged values as regressors is denoted AR(p).
Autoregressive conditional heteroskedasticity (ARCH)
A time series model of conditional heteroskedasticity. R2
Autoregressive distributed lag model
A linear regression model in which the time series variable Yt is expressed as a function of lags of Yt and of another variable, Xt.The model is denoted ADL(p,q), where p denotes the number of lags of Yt and q denotes the number of lags of Xt.
Average causal effect
The population average of the individual causal effects in a heterogeneous population. Also called the average treatment effect.
Balanced panel
A panel data set with no missing observations, that is, in which the variables are observed for each entity and each time period.
Base specification
A baseline or benchmark regression specification that includes a set of regressors chosen using a combination of expert judgment, economic theory, and knowledge of how the data were collected.
Bayes information criterion
See information criterion.
Bernoulli distribution
The probability distribution of a Bernoulli random variable.
Bernoulli random variable
A random variable that takes on two values, 0 and 1.
Best linear unbiased estimator
An estimator that has the smallest variance of any estimator that is a linear function of the sample values Y and is unbiased. Under the Gauss-Markov conditions, the OLS estimator is the best linear unbiased estimator of the regression coefficients conditional on the values of the regressors.
Bias
The expected value of the difference between an estimator and the parameter that it is estimating. If is an estimator of mY, then the bias of is E( )2 mY.
BIC
See information criterion.
Binary variable
A variable that is either 0 or 1.A binary variable is used to indicate a binary outcome. For example,X is a binary (or indicator, or dummy) variable for a person’s gender if X 5 1 if the person is female and X 5 0 if the person is male. mˆY mˆY mˆY
Bivariate normal distribution
A generalization of the normal distribution to describe the joint distribution of two random variables.
BLUE
See best linear unbiased estimator.
Break date
The date of a discrete change in population time series regression coefficient(s).
Causal effect
The expected effect of a given intervention or treatment as measured in an ideal randomized controlled experiment.
Central limit theorem
A result in mathematical statistics that says that, under general conditions, the sampling distribution of the standardized sample average is well approximated by a standard normal distribution when the sample size is large.
Chi-squared distribution
The distribution of the sum of m squared independent standard normal random variables.The parameter m is called the degrees of the freedom of the chi-squared distribution.
Chow test
A test for a break in a time series regression at a known break date.
Coefficient of determination
See R2.
Cointegration
When two or more time series variables share a common stochastic trend.
Common trend
A trend shared by two or more time series.
Conditional distribution
The probability distribution of one random variable given that another random variable takes on a particular value.
Conditional expectation
The expected value of one random value given that another random variable takes on a particular value.
Conditional heteroskedasticity
The variance, usually of an error term, depends on other variables.
Conditional mean
The mean of a conditional distribution; see conditional expectation.
Conditional mean independence
The conditional expectation of the regression error ui, given the regressors, depends on some but not all of the regressors.
Conditional variance
The variance of a conditional distribution.
Confidence interval (or confidence set)
An interval (or set) that contains the true value of a population parameter with a prespecified probability when computed over repeated samples.
Confidence level
The prespecified probability that a confidence interval (or set) contains the true value of the parameter.
Consistency
Means that an estimator is consistent. See consistent estimator.
Consistent estimator
An estimator that converges in probability to the parameter that it is estimating.
Constant regressor
The regressor associated with the regression intercept; this regressor is always equal to 1.
Constant term
The regression intercept.
Continuous random variable
A random variable that can take on a continuum of values.
Control group
The group that does not receive the treatment or intervention in an experiment.
Control variable
Another term for a regressor; more specifically, a regressor that controls for one of the factors that determine the dependent variable.
Convergence in distribution
When a sequence of distributions converges to a limit; a precise definition is given in Section 17.2.
Convergence in probability
When a sequence of random variables converges to a specific value; for example, when the sample average becomes close to the population mean as the sample size increases; see Key Concept 2.6 and Section 17.2.
Correlation
A unit-free measure of the extent to which two random variables move, or vary, together.The correlation (or correlation coefficient) between X and Y is sXY/sXsY and is denoted corr(X,Y).
Correlation coefficient
See correlation.
Covariance
A measure of the extent to which two random variables move together.The covariance between X and Y is the expected value E[(X 2 mX)(Y 2 mY)], and is denoted by cov(X,Y) or by sXY.
Covariance matrix
A matrix composed of the variances and covariances of a vector of random variables.
Critical value
The value of a test statistic for which the test just rejects the null hypothesis at the given significance level.
Cross-sectional data
Data collected for different entities in a single time period.
Cubic regression model
A nonlinear regression function that includes X, X2, and X3 as regressors.
Cumulative distribution function (c.d.f.)
See cumulative probability distribution.
Cumulative dynamic multiplier
The cumulative effect of a unit change in the time series variable X on Y.The h-period cumulative dynamic multiplier is the effect of a unit change in Xt on Yt + Yt+1+ . . . + Yt+h.
Cumulative probability distribution
A function showing the probability that a random variable is less than or equal to a given number.
Dependent variable
The variable to be explained in a regression or other statistical model; the variable appearing on the left-hand side in a regression.
Deterministic trend
A persistent long-term movement of a variable over time that can be represented as a nonrandom function of time.
Dickey-Fuller test
A method for testing for a unit root in a first order autoregression [AR(1)].
Differences estimator
An estimator of the causal effect constructed as the difference in the sample average outcomes between the treatment and control groups.
Differences-in-differences estimator
The average change in Y for those in the treatment group, minus the average change in Y for those in the control group.
Discrete random variable
A random variable that takes on discrete values.
Distributed lag model
A regression model in which the regressors are current and lagged values of X.
Dummy variable
See binary variable.
Dummy variable trap
A problem caused by including a full set of binary variables in a regression together with a constant regressor (intercept), leading to perfect multicollinearity.
Dynamic causal effect
The causal effect of one variable on current and future values of another variable.
Dynamic multiplier
The h-period dynamic multiplier is the effect of a unit change in the time series variable Xt on Yt+h.
Endogenous variable
A variable that is correlated with the error term.
Error term
The difference between Y and the population regression function, denoted by u in this textbook.
Errors-in-variables bias
The bias in an estimator of a regression coefficient that arises from measurement errors in the regressors.
Estimate
The numerical value of an estimator computed from data in a specific sample.
Estimator
A function of a sample of data to be drawn randomly from a population. An estimator is a procedure for using sample data to compute an educated guess of the value of a population parameter, such as the population mean.
Exact distribution
The exact probability distribution of a random variable.
Exact identification
When the number of instrumental variables equals the number of endogenous regressors.
Exogenous variable
A variable that is uncorrelated with the regression error term.
Expected value
The long-run average value of a random variable over many repeated trials or occurrences. It is the probability-weighted average of all possible values that the random variable can take on.The expected value of Y is denoted E(Y) and is also called the expectation of Y.
Experimental data
Data obtained from an experiment designed to evaluate a treatment or policy or to investigate a causal effect.
Experimental effect
When experimental subjects change their behavior because they are part of an experiment.
Explained sum of squares (ESS)
The sum of squared deviations of the predicted values of Yi, ,from their average; see Equation (4.14).
Explanatory variable
See regressor.
External validity
Inferences and conclusions from a statistical study are externally valid if they can be generalized from the population and the setting studied to other populations and settings.
F-statistic
A statistic used to a test joint hypothesis concerning more than one of the regression coefficients.
Fm,n distribution
The distribution of a ratio of independent random variables, where the numerator is a chi-squared random variable with m degrees of freedom, divided by m, and the denominator is a chi-squared random variable with n degrees of freedom divided by n.
Fm,∞ distribution
The distribution of a random variable with a chi-squared distribution with m degrees of freedom, divided by m.
Feasible GLS
A version of the generalized least squares (GLS) estimator that uses an estimator of the conditional variance of the regression errors and covariance between the regression errors at different observations.
Feasible WLS
A version of the weighted least squares (WLS) estimator that uses an estimator of the conditional variance of the regression errors.
First difference
The first difference of a time series variable Yt is Yt 2 Yt21, denoted DYt.
First-stage regression
The regression of an included endogenous variable on the included exogenous variables, if any, and the instrumental variable(s) in two stage least squares.
Fitted values
See predicted values.
Fixed effects
Binary variables indicating the entity or time period in a panel data regression.
Fixed effects regression model
A panel data regression that includes entity fixed effects. ˆYi
Forecast error
The difference between the value of the variable that actually occurs and its forecasted value.
Forecast interval
An interval that contains the future value of a time series variable with a prespecified probability.
Functional form misspecification
When the form of the estimated regression function does not match the form of the population regression function; for example, when a linear specification is used but the true population regression function is quadratic.
GARCH
See generalized autoregressive conditional heteroskedasticity.
Gauss-Markov theorem
Mathematical result stating that, under certain conditions, the OLS estimator is the best linear unbiased estimator of the regression coefficients conditional on the values of the regressors.
Generalized autoregressive conditional heteroskedasticity
A time series model for conditional heteroskedasticity.
Generalized least squares (GLS)
A generalization of OLS that is appropriate when the regression errors have a known form of heteroskedasticity (in which case GLS is also referred to as weighted least squares, WLS) or a known form of serial correlation.
Generalized method of moments
A method for estimating parameters by fitting sample moments to population moments that are functions of the unknown parameters. Instrumental variables estimators are an important special case.
GMM
See generalized method of moments.
Granger causality test
A procedure for testing whether current and lagged values of one time series help predict future values of another time series.
HAC standard errors
See heteroskedasticity- and autocorrelation-consistent (HAC) standard errors.
Hawthorne effect
See experimental effect.
Heteroskedasticity
The situation in which the variance of the regression error term ui, conditional on the regressors, is not constant. Heteroskedasticity- and autocorrelation-consistent
(HAC) standard errors
Standard errors for OLS estimators that are consistent whether or not the regression errors are heteroskedastic and autocorrelated.
Heteroskedasticity-robust standard error
Standard errors for the OLS estimator that are appropriate whether the error term is homoskedastic or heteroskedastic.
Heteroskedasticity-robust t-statistic
A t-statistic constructed using a heteroskedasticity-robust standard error.
Homoskedasticity
The variance of the error term ui, conditional on the regressors, is constant.
Homoskedasticity-only F statistic
A form of the Fstatistic that is valid only when the regression errors are homoskedastic.
Homoskedasticity-only standard errors
Standard errors for the OLS estimator that are appropriate only when the error term is homoskedastic.
Hypothesis test
A procedure for using sample evidence to help determine if a specific hypothesis about a population is true or false.
i.i.d.
Independently and indentically distributed.
Identically distributed
When two or more random variables have the same distribution.
Impact effect
The contemporaneous, or immediate, effect of a unit change in the time series variable Xt on Yt.
Imperfect multicollinearity
The condition in which two or more regressors are highly correlated.
Included endogenous variables
Regressors that are correlated with the error term (usually in the context of instrumental variable regression).
Included exogenous variables
Regressors that are uncorrelated with the error term (usually in the context of instrumental variable regression).
Independence
When knowing the value of one random variable provides no information about the value of another random variable.Two random variables are independent if their joint distribution is the product of their marginal distributions.
Indicator variable
See binary variable.
Information criterion
A statistic used to estimate the number of lagged variables to include in an autoregression or a distributed lag model. Leading examples are the Akaike information criterion (AIC) and the Bayes information criterion (BIC).
Instrument
See instrumental variable.
Instrumental variable
A variable that is correlated with an endogenous regressor (instrument relevance) and is uncorrelated with the regression error (instrument exogeneity).
Instrumental variables (IV) regression
A way to obtain a consistent estimator of the unknown coefficients of the population regression function when the regressor,X, is correlated with the error term, u.
Interaction term
A regressor that is formed as the product of two other regressors, such as X1i 3 X2i.
Intercept
The value of b0 in the linear regression model.
Internal validity
When inferences about causal effects in a statistical study are valid for the population being studied.
J-statistic
Astatistic for testing overidentifying restrictions in instrumental variables regression.
Joint hypothesis
A hypothesis consisting of two or more individual hypotheses, that is, involving more than one restriction on the parameters of a model.
Joint probability distribution
The probability distribution determining the probabilities of outcomes involving two or more random variables.
Kurtosis
A measure of how much mass is contained in the tails of a probability distribution.
Lags
The value of a time series variable in a previous time period.The jth lag of Yt is Yt2j.
Law of iterated expectations
A result in probability theory that says that the expected value of Y is the expected value of its conditional expectation given X, that is, E(Y) 5 E[E(Y X)].
Law of large numbers
According to this result from probability theory, under general conditions the sample average will be close to the population mean with very high probability when the sample size is large.
Least squares assumptions
The assumptions for the linear regression model listed in Key Concept 4.3 (single variable regression) and Key Concept 6.4 (multiple regression model).
Least squares estimator
An estimator formed by minimizing the sum of squared residuals.
Limited dependent variable
A dependent variable that can take on only a limited set of values. For example, the variable might be a 021 binary variable or arise from one of the models described in Appendix 11.3.
Linear-log model
A nonlinear regression function in which the dependent variable is Y and the independent variable is ln(X).
Linear probability model
A regression model in which Y is a binary variable.
Linear regression function
A regression function with a constant slope.
Local average treatment effect
A weighted average treatment effect estimated, for example, by TSLS.
Log-linear model
A nonlinear regression function in which the dependent variable is ln(Y) and the independent variable is X.
Log-log model
A nonlinear regression function in which the dependent variable is ln(Y) and the independent variable is ln(X). @
Logarithm
A mathematical function defined for a positive argument; its slope is always positive but tends to zero.The natural logarithm is the inverse of the exponential function, that is, X 5 ln(eX).
Logit regression
A nonlinear regression model for a binary dependent variable in which the population regression function is modeled using the cumulative logistic distribution function.
Long-run cumulative dynamic multiplier
The cumulative long-run effect on the time series variable Y of a change in X.
Longitudinal data
See panel data.
Marginal probability distribution
Another name for the probability distribution of a random variable Y, which distinguishes the distribution of Y alone (the marginal distribution) from the joint distribution of Y and another random variable.
Maximum likelihood estimator (MLE)
An estimator of unknown parameters that is obtained by maximizing the likelihood function; see Appendix 11.2.
Mean
The expected value of a random variable.The mean of Y is denoted mY.
Moments of a distribution
The expected value of a random variable raised to different powers.The rth moment of the random variable Y is E(Yr).
Multicollinearity
See perfect multicollinearity and imperfect multicollinearity.
Multiple regression model
An extension of the single variable regression model that allows Y to depend on k regressors.
Natural experiment
See quasi-experiment.
Natural logarithm
See logarithm.
95% confidence set
A confidence set with a 95% confidence level; see confidence interval.
Nonlinear least squares
The analog of OLS that applies when the regression function is a nonlinear function of the unknown parameters.
Nonlinear least squares estimator
The estimator obtained by minimizing the sum of squared residuals when the regression function is nonlinear in the parameters.
Nonlinear regression function
A regression function with a slope that is not constant.
Nonstationary
When the joint distribution of a time series variable and its lags changes over time.
Normal distribution
A commonly used bell-shaped distribution of a continuous random variable.
Null hypothesis
The hypothesis being tested in a hypothesis test, often denoted by H0.
Observation number
The unique identifier assigned to each entity in a data set.
Observational data
Data based on observing, or measuring, actual behavior outside an experimental setting. OLS estimator. See ordinary least squares estimator.
OLS regression line
The regression line with population coefficients replaced by the OLS estimators.
OLS residual
The difference between Yi and the OLS regression line, denoted by in this textbook.
Omitted variables bias
The bias in an estimator that arises because a variable that is a determinant of Y and is correlated with a regressor has been omitted from the regression.
One-sided alternative hypothesis
The parameter of interest is on one side of the value given by the null hypothesis.
Order of integration
The number of times that a time series variable must be differenced to make it stationary.A time series variable that is integrated of order p must be differenced p times and is denoted I(p).
Ordinary least squares estimator
The estimator of the regression intercept and slope(s) that minimizes the sum of squared residuals.
Outlier
An exceptionally large or small value of a random variable.
Overidentification
When the number of instrumental variables exceeds the number of included endogenous regressors.
p-value
The probability of drawing a statistic at least as adverse to the null hypothesis as the one actually computed, assuming the null hypothesis is correct. Also called the marginal significance probability, the p-value is the smallest significance level at which the null hypothesis can be rejected.
Panel data
Data for multiple entities where each entity is observed in two or more time periods.
Parameter
A constant that determines a characteristic of a probability distribution or population regression function.
Partial compliance
Occurs when some participants fail to follow the treatment protocol in a randomized experiment.
Partial effect
The effect on Y of changing one of the regressors, holding the other regressors constant.
Perfect multicollinearity
Occurs when one of the regressors is an exact linear function of the other regressors.
Polynomial regression model
A nonlinear regression function that includes X, X2, . . . and Xr as regressors, where r is an integer. uˆi
Population
The group of entities—such as people, companies, or school districts—being studied.
Population coefficients
See population intercept and slope.
Population intercept and slope
The true, or population, values of b0 (the intercept) and b1 (the slope) in a single variable regression. In a multiple regression, there are multiple slope coefficients (b1, b2, . . . , bk), one for each regressor.
Population multiple regression model
The multiple regression model in Key Concept 6.2.
Population regression line
In a single variable regression, the population regression line is b0 + b1Xi, and in a multiple regression it is b0 + b1X1i + b2X2i + . . . + bkXki.
Power
The probability that a test correctly rejects the null hypothesis when the alternative is true.
Predicted value
The value of Yi that is predicted by the OLS regression line, denoted by in this textbook.
Price elasticity
The percentage change in the quantity demanded resulting from a 1% increase in price.
Probability
The proportion of the time that an outcome (or event) will occur in the long run.
Probability density function (p.d.f.)
For a continuous random variable, the area under the probability density function between any two points is the probability that the random variable falls between those two points.
Probability distribution
For a discrete random variable, a list of all values that a random variable can take on and the probability associated with each of these values.
Probit regression
A nonlinear regression model for a binary dependent variable in which the population regression function is modeled using the cumulative standard normal distribution function.
Program evaluation
The field of study concerned with estimating the effect of a program, policy, or some other intervention or “treatment.”
Pseudo out-of-sample forecast
A forecast computed over part of the sample using a procedure that is as if these sample data have not yet been realized.
Quadratic regression model
A nonlinear regression function that includes X and X2 as regressors.
Quasi-experiment
A circumstance in which randomness is introduced by variations in individual circumstances that make it appear as if the treatment is randomly assigned. ˆYi
R2
In a regression, the fraction of the sample variance of the dependent variable that is explained by the regressors
Random walk
A time series process in which the value of the variable equals its value in the previous period, plus an unpredictable error term.
Random walk with drift
A generalization of the random walk in which the change in the variable has a nonzero mean but is otherwise unpredictable.
Randomized controlled experiment
An experiment in which participants are randomly assigned to a control group, which receives no treatment, or to a treatment group, which receives a treatment.
Regressand
See dependent variable.
Regression specification
A description of a regression that includes the set of regressors and any nonlinear transformation that has been applied.
Regressor
A variable appearing on the right-hand side of a regression; an independent variable in a regression.
Rejection region
The set of values of a test statistic for which the test rejects the null hypothesis.
Repeated cross-sectional data
A collection of crosssectional data sets, where each cross-sectional data set corresponds to a different time period.
Restricted regression
Aregression in which the coefficients are restricted to satisfy some condition. For example, when computing the homoskedasticityonly F-statistic, this is the regression with coefficients restricted to satisfy the null hypothesis.
Root mean squared forecast error
The square root of the mean of the squared forecast error.
Sample correlation
An estimator of the correlation between two random variables.
Sample covariance
An estimator of the covariance between two random variables.
Sample selection bias
The bias in an estimator of a regression coefficient that arises when a selection process influences the availability of data and that process is related to the dependent variable.This induces correlation between one or more regressors and the regression error.
Sample standard deviation
An estimator of the standard deviation of a random variable.
Sample variance
An estimator of the variance of a random variable.
Sampling distribution
The distribution of a statistic over all possible samples; the distribution arising from repeatedly evaluating the statistic using a R2 series of randomly drawn samples from the same population.
Scatterplot
A plot of n observations on Xi and Yi, in which each observation is represented by the point (Xi,Yi).
Serial correlation
See autocorrelation.
Serially uncorrelated
A time series variable with all autocorrelations equal to zero.
Significance level
The prespecified rejection probability of a statistical hypothesis test when the null hypothesis is true.
Simple random sampling
When entities are chosen randomly from a population using a method that ensures that each entity is equally likely to be chosen.
Simultaneous causality bias
When, in addition to the causal link of interest from X to Y, there is a causal link from Y to X. Simultaneous causality makes X correlated with the error term in the population regression of interest.
Simultaneous equations bias
See simultaneous causality bias.
Size of a test
The probability that a test incorrectly rejects the null hypothesis when the null hypothesis is true.
Skewness
A measure of the aysmmetry of a probability distribution.
Standard deviation
The square root of the variance. The standard deviation of the random variable Y, denoted sY, has the units of Y and is a measure of the spread of the distribution of Y around its mean.
Standard error of an estimator
An estimator of the standard deviation of the estimator.
Standard error of the regression (SER)
An estimator of the standard deviation of the regression error u.
Standard normal distribution
The normal distribution with mean equal to 0 and variance equal to 1, denoted N(0, 1).
Standardizing a random variable
An operation accomplished by subtracting the mean and dividing by the standard deviation, which produces a random variable with a mean of 0 and a standard deviation of 1.The standardized value of Y is (Y 2 mY)/sY.
Stationarity
When the joint distribution of a time series variable and its lagged values does not change over time.
Statistically insignificant
The null hypothesis (typically, that a regression coefficient is zero) cannot be rejected at a given significance level.
Statistically significant
The null hypothesis (typically, that a regression coefficient is zero) is rejected at a given significance level.
Stochastic trend
A persistent but random long-term movement of a variable over time.
Strict exogeneity
The requirement that the regression error has a mean of zero conditional on current, future, and past values of the regressor in a distributed lag model.
Student t distribution
The Student t distribution with m degrees of freedom is the distribution of the ratio of a standard normal random variable, divided by the square root of an independently distributed chi-squared random variable with m degrees of freedom divided by m.As m gets large, the Student t distribution converges to the standard normal distribution.
Sum of squared residuals (SSR)
The sum of the squared OLS residuals.
t-distribution
See Student t distribution.
t-ratio
See t-statistic.
t-statistic
A statistic used for hypothesis testing. See Key Concept 5.1.
Test for a difference in means
A procedure for testing whether two populations have the same mean.
Time effects
Binary variables indicating the time period in a panel data regression.
Time and entity fixed effects regression model
A panel data regression that includes both entity fixed effects and time fixed effects.
Time fixed effects
See time effects.
Time series data
Data for the same entity for multiple time periods.
Total sum of squares (TSS)
The sum of squared deviations of Yi, from its average, .
Treatment effect
The causal effect in an experiment or a quasi-experiment; see causal effect.
Treatment group
The group that receives the treatment or intervention in an experiment.
TSLS
See two stage least squares. Y
Two-sided alternative hypothesis
When, under the alternative hypothesis, the parameter of interest is not equal to the value given by the null hypothesis.
Two stage least squares
An instrumental variable estimator, described in Key Concept 12.2.
Type I error
In hypothesis testing, the error made when the null hypothesis is true but is rejected.
Type II error
In hypothesis testing, the error made when the null hypothesis is false but is not rejected.
Unbalanced panel
A panel data set in which some data are missing.
Unbiased estimator
An estimator with a bias that is equal to zero.
Uncorrelated
Two random variables are uncorrelated if their correlation is zero.
Underidentification
When the number of instrumental variables is less than the number of endogenous regressors.
Unit root
Refers to an autoregression with a largest root equal to 1.
Unrestricted regression
When computing the homoskedasticity-only F-statistic, this is the regression that applies under the alternative hypothesis, so the coefficients are not restricted to satisfy the null hypothesis.
VAR
See vector autoregression.
Variance
The expected value of the squared difference between a random variable and its mean; the variance of Y is denoted .
Vector autoregression
A model of k time series variables consisting of k equations, one for each variable, in which the regressors in all equations are lagged values of all the variables.
Volatility clustering
When a time series variable exhibits some clustered periods of high variance and other clustered periods of low variance.
Weak instruments
Instrumental variables that have a low correlation with the endogenous regressor(s).
Weighted least squares (WLS)
An alternative to OLS that can be used when the regression error is heteroskedastic and the form of the heteroskedasticity is known or can be estimated.