, 10 tweets, 3 min read
I just taught @CdeChaisemartin & D'Haultfoeuille (2018) Fuzzy DID and thought I would share some insights I have about the nature of the problem that Fuzzy DID introduces.

*These insights are in their paper, this is just my take on an explanation

1/n
Consider two groups g, and time periods t.

You can write the expected value of the outcome for a group-year combo as a function of the avg. outcome in the state of the world that units are not treated + the ATT * the share that are treated.

Makes sense.

2/n
Let's consider the standard DID estimator, we can write is as function of these four expectations. Why do we do this? Because we are trying to remove the unobserved determinants of the outcomes (in levels) that are potential sources of bias. I.e., we want to kill Y(0) term.

3/n
Plug in the first equation into the DID we get the following. As desired, all the Y(0) terms cancel out from the Common Trends Assumption. So we can celebrate!

Or can we?

4/n
Our efforts to remove unobserved confounders left us with an odd equation. Our DID gives us a weighted sum of the ATTs specific to each group and time period. This would be fine if we added all the ATTs together, but instead two terms are subtracted!

5/n
As a result, even if the ATTs are positive in each period and group, we can end up with a null or even negative DID estimate! Had different ATTs been negative, it would makes sense for them to cancel out, but it is odd to cancel out based on negative weights alone.

6/n
In sharp case, where only treated group is treated in post period. The equation simplifies and DID estimates the ATT for the treatment group in the post period alone. This is sensible. When units in other groups and periods are treated, then we have this odd weighted sum.

7/n
One solution is to assume that ATT is homogenous for each group and period, but this is likely false in most contexts.

What to do?

They provide Stata code to allow you to estimate these weights to see how prevalent negative weighting is in your context.

8/n
Further, they show less restrictive assumptions that let you still estimate a sensible treatment effect. Finally, they propose "Time-Corrected" estimator that relaxes required assumptions further and isn't plagued by these negative weights. (I can tweet more on this later)

9/n
I love this paper, as well as the TWFE follow-up piece. It is accessible, extremely important (this applies to nearly every panel data design!), and has practical applications in applied work. I recommend a thorough read!

n/n
Missing some Tweet in this thread? You can try to force a refresh.

Enjoying this thread?

Keep Current with Jason Cook

Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

Twitter may remove this content at anytime, convert it as a PDF, save and print for later use!

Try unrolling a thread yourself!

how to unroll video

1) Follow Thread Reader App on Twitter so you can easily mention us!

2) Go to a Twitter thread (series of Tweets by the same owner) and mention us with a keyword "unroll" @threadreaderapp unroll

You can practice here first or read more on our help page!

Follow Us on Twitter!

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just three indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3.00/month or $30.00/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!