F7 Matching and subclassification Flashcards

1
Q

What is Jan’s strategy?

A

Try different types of matching and see whether the result hold. You want as many matches for a treated unit as possible

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is sub-classification and problems with this method?

A

Summing over weighted differences between control and treatment group. Weighting differences in means by strata-specific weights.

Problem: Curse of dimensionality (difficult with large datasets). Becomes an obsolete method.

Solution: Collapse groups (do not imputate data)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is exact matching?

A

Match exactly on all confounders (high demands for data - discrete variables needed).

You quickly run into curse of dimensionality.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

There are two kinds of approximate matching. What are they?

A

Approximate matching is a method, where you minimize the distance between control and treated unit.

Nearest neighbor
Propensity score matching

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is nearest neighbor matching?

A

We minimize the sum of the distance between all confounding variables.

To account for different scaling the normalized Euclidean distance is used. This accounts for greater dispersion on some variables.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is propensity score matching?

A

Confounders are collapse into a single dimension - a propensity score [0,1] indicating your probability of being in the treatment group conditional on confounders.

So PSM reduces the problem of finding a suitable match to one dimension: p(X)=Pr(D|X).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are two important assumptions for matching?

A

Common support: At all values of the propensity score/relevant range, we want units in both the control and treatment group.

Conditional independence assumption: Conditional on X (confounders) then Y^1 and Y^0 are independent from treatment D. Once we factor out confounders.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What are advantages of propensity score matching?

A

No curse of dimensionality.

It’s possible to check the significance level of confounders - whether the matter or not.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does Jan think about matching?

A

It can be succesfuld. It’s superior to multivariate regression as you consult the common support assumption. Especially with exact matching.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is CIA?

A

Conditional independence assumption.

With a cross-sectional dataset we are forced to assume that D is independent of potential outcomes once we condition on covariates

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What does garbage in, garbage out mean?

A

Quality of propensity score/matching is intimately linked to quality of covariates

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is one advantage of matching?

A

It casts light on important concepts such as common support! Normally hidden in regression.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What are criticsm of matching?

A

Common support means you reduce the sample size (variance trade-off).

Propensity score matching: You can have 0,8 for very different reasons.

The larger the distance from match - the more bias (the larger the sample size it converges to zero). You could use the bias-correction estimator.

Unobserved confounders.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly