RDD Flashcards

1
Q

What is the key feature of the sharp RDD design

A

The conditional probability of receiving treatment jumps directly from 0 to 1 at the threshold.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What about RDD with time index?

A

To define a RDD design we need to have a score (running variable), a cut off and a treatment. It is usually a crossestional design since there is no time variation.

There should be no time-index t in a pure RDD. There should only be a unit index i.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How should we think about boundary RDDs?

A

For this to work we need two forcing variables, one for longitude and one for latitude. We then have two dimension, which complicate things. We can’t just look at the shortest distance to the borders. In a “spacing”-design like this, we have a problem with multiple treatments. It should be clear in a RDD what the treatment is! Not a bunch of stuff that changes since we do not know what is causing the different.

In a multi dimensional boundary design, we should have multiple forcing variables (longitude/latitude). We thus also need to control for both of them and all the i interactions between them.

There are some good clear border-discontinuity designs. For example comparing minimum wages between two states. Then it is very clear how far away we are from the border, either measured in a one as a one-dimensional thing or using longitude/latitude which is more flexible. KEY in using this kind of designs is that it should be clear what the border accualy is.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Can we let define control and treatment groups before the cut off in a

A

This is actually the other way around than what we should do. There should be a cut off that defines the treatment status. This is something that should be criticised a lot since this is not a standard thing!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is the key identification thing we should worry about in a RDD?

A

When it comes to a RDD-design, it should be all about smoothness (sorting violates smoothness and so on). Smoothness is the KEY in RDD design. We should not combine this with assumptions like parallel trends etc.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Can we use fixed effects in a RDD?

A

Using fixed effects in RDD is non-standard. This is a warning since. We should not mix different kinds of identifications. It is highly problematic if we have heterogeneous treatment effects. The estimates will then change by the inclusion of fixed effects.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Studying BW-graphs, what should we look out for?

A

Both estimates and CI that is varring.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the most important thing to look at when reading a research rapport?

A

When we should write a report, the most important thing is to look at the design. Does it really fit the empirical framework?

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How about using multiple cut off’s and pulling them together in a RDD?

A

This is a non-standard RDD approach. It would be more OK for each of the topics to estimate a separate RDD. It is NOT OK to pull everything together. Pulling together different cut offs, creates multiple forcing variables.

By normalizing the running variable so it becomes one, we will create bunching in the forcing variable. When we have discreteness, we do not have smoothness. This is maybe not a problem if we have very many values, but if we have few then we will have a valuation of smoothness, we then need to extrapolate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Briefly explain RDD with local-randomization estimation.

A

We have not discussed local-randomization estimation but it is like thinking about the design as a RCT really close to the threshold. Then we estimate the means in the treatment and control group and compare them. This is credible since we are so close to the threshold.

For this to be true, the slope of the forcing variable have to be constant on both sides of the threshold. The lines should be completely flat.

Here there is a big problem with SE’s since we need to use randomized inference.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How should we present BW choice in our RDD paper according to Per?

A

In a RDD we have to many choices. A researcher that can choose a BW will of course choose a BW that confirm his estimates. We should therefore instead let the data speak. We should show all the estimates for different BW’s (including) the optimal in a graph. This will then show if it matters if we are choosing the optimal or not.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is a good way of presenting covariate balance in a RDD paper?

A

Canay and Kamat suggest that we compare the distribution (CDF) of the guys on the right and on the left of the threshold.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is a McCrary test?

A

Testing for bunching in the forcing variable = testing the continuity in the density of the running variable. We test this by comparing the values of an estimated density on the right and the left och the cutoff. See rddensity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is the difference by window and bandwidth in a RDD?

A
  • Window = what we are looking at to see our cutoff etc. We pick window size so it looks good. This is thus totally irellavant for estimating the effect. This is just for visualization.
  • Bandwidth = how many data points we use around the the threshold when estimating the effect. This is not what we show.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What should we do according to Goldsmith-Pinkham if we can’t shrink our BW, that is, if we have a discrete running variable?

A

In this case we do not have that much points just at the cutoff so we will definitely have a bias. We should use another RD command called RDHonest.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What can we do if we have bunching according to Goldsmith-Pinkham?

A

Gerard, Rokkanen and Rothe (2020) propose a partial identification approach to allow for the possibility of bunching. This is check on the robustness of our results. How sensitive are our results to the manipulation?

Their approach hingest on the assumption that people only manipulate their result in one direction. Those manipulators will mask our treatment-effect. This paper then derives “shap-bounds” for those who could be affected by the treatment. We identify the share of the masking individuals. This is a nice thing to also add in addition to the Mcrary test.

17
Q

What is a regression Kink Design?

A

When we cross the threshold, the relationship between the outcome variable and the running variable changes rather then jumps. In for example public finance there are many policies that creates linear shifts in incentives rather than discontinuous jumps.

18
Q

What is the biggest threat to a RDD?

A

The most important threat to any RD design is the possibility that units are able to strategically and precisely change their score to be assigned to their preferred treatment condition (Lee, 2008; McCrary, 2008), which might induce a discontinuous change in their observable and/or unobservable characteristics at or near the cutoff and thus confound causal conclusions.

19
Q

What is a big draw back from the RDD?

A

An important limitation of RD designs is that they have low external validity. In the absence of additional assumptions, it is not possible to learn about treatment effects away from the cutoff—that is, to extrapolate the effect of the treatment.

20
Q

Why is local polynomial estimation better than a global approach?

A

Local polynomial methods are preferable to global polynomial methods because they avoid several of the methodological problems created by the use of global polynomials such as erratic behavior near boundary points (Runge’s phenomenon), counterintuitive weighting, overfitting, and general lack of robustness.

21
Q

What is important when doing placebo cut off-tests?

A

Importantly, in order to avoid treatment effect contamination, this validation method should be implemented for units below and above the cutoff separately.

22
Q

What is the RDD checklist

A
  • Is there actually a jump?
    • [ ] Show/picture the running variable without regression lines and cut off line to show that there is a clear jump.
  • No bunching in the running variable
    • [ ] Show density graph
    • [ ] Run a McCrary density test
  • Potential outcome is continuous
    • [ ] Show RDD graph with bin-averages
    • [ ] Placebo checks at other thresholds
  • Pre-determined covariates (characteristics) are balanced around the threshold
    • [ ] Continuity checks (graphical or formally) for other covariates
  • BW sensitivity checks
    • [ ] Show graphically how estimates and CI changes with different BW. Show explicitly what the optimal is.
  • Polynomial sensitivity checks
    • [ ] How?
  • Presentation sensitivity
    • [ ] Show main graphs with other bin-suggestions and show that there is still a clear jump.
  • What happens if we lose observations or clusters?
  • Is it possible to generalize the LATE?
  • No other policy changes by the threshold?
  • Are the there other behavior effects at the threshold that affects the result? E.g barely passing the threshold might make you worse of.
  • No controls other than the forcing variable?
  • Discussion of the bias/variance trade of around the threshold.
  • Not mixing identifications! No fixed effects in the model etc.
23
Q

What is the best practice in a RDD design?

A

It is now understood that best practices for estimation and inference in RD designs should be based on tuning parameter choices that are objective, principled, and data-driven, thereby removing researchers’ ability to engage in arbitrary specification searching, and thus leading to credible and replicable empirical conclusions.

24
Q

What should we think of regarding power in a RDD design

A

We need many observations around the threshold

25
Q

What about heterogeneity and RDD?

A

Heterogeneity can be a big problem in RDD.

26
Q

Which two ways do Cattaneo et al 2020 argue for when choosing the optimal bin size in a RDD?

A

IMSE-minimizing and mimicking variance.

27
Q

What choices to we have to make when estimating a RDD

A

When estimating the RDD we choose, the bandwidth h , kernel, K and polynomial order p.

28
Q

Explain the bias-variance trade off in RDD.

A

Not capturing the true functional form in the window on both sides of the threshold will create a bias.

Smaller bandwidth → less bias. If we choose a large bandwidth and we try to do a linear approximation to a non-linear function, we will likely be very of. The smaller the the bandwidth, the better approximation we are going to make.

If we zoom in (decreese BW = make $h$ smaller), we will get an better approximation. The more we zoom in at the window, the better approximation we get. However, it is a bias-variance trade off. The more we zoom in, the less bias we get but also more variance. This since the smaller $h$ becomes, the more data is thrown out and thus the greater the variance.

29
Q

How should we rapport estimates and inference in a RDD-paper?

A

It happens to be that the optimal bandwidth for the point estimates is not optimal for how we compute the standard errors. We will have bias in our SE with the point estimate optimal.

Estimates = Conventional
Confidence interval = Robust bias corrected

30
Q

What is a nice to present placebo cutoff checks?

A

A graph, showing the estimates for different cut offs

31
Q

Explain the donut-hole-approach in RDD

A

The so-called “donut hole” approach (Bajari, Hong, Park, and Town, 2011) investigates the sensitivity of the empirical conclusions by removing the few observations closest to the cutoff, and then implementing estimation and inference. Heuristically, the goal is to assess how much influence the few observations closest to the cutoff have on the overall empirical results. The intuition is that if there is endogenous sorting of units across the cutoff, such sorting might occur only among units whose scores are very close to the cutoff, and thus when those observations are excluded the RD treatment effect may change.

32
Q

What degree of polynomial should one use in a RDD setting?

A

The order of the polynomial should always be low, to avoid overfitting and erratic behavior near the cutoff point. The default recommendation is p = 1 (local linear). See @Pei et al (2021) for a data driven approach.

33
Q

What is the optimal BW in RDD?

A

The choice of BW is a matter of optimally solving the bias-variancetrade off. We try to minimise the mean square error of the local polynomial (MSE). A optimal BW will yield a optimal point estimate.