If your treatment variable has bunching, this is good news! It may be possible to use this bunching to build a correction for endogeneity. We show how this may be done in our new paper. Follow the thread. #EconTwitter#econometrics 1/
We focus here on bunching at one of the ends of the support, which turns out to be very common. For example, TV watching, enrichment activities, maternal labor supply, smoking amounts, all have bunching at zero, and the list of variables where this happens goes on and on. 2/
Take the amount of time children spend watching TV. You can't watch a negative amount of time. Even if your characteristics would place you as a negative watcher, you will still watch zero, and you will be bunched with all the other extreme people. 3/
So, at the bunching point, there are discontinuities in the confounders. When we compare zero TV per week with a few minutes of TV per week, the treatment difference is irrelevant, but the confounder difference is substantial. 4/
Thus, a discontinuity in the outcome at the bunching point will only reflect the effect of the confounders, separate of the effect of the treatment. This reveals information about the endogeneity bias. We can use this information to build the correction of the endogeneity bias.5/
The main requirement is that the information the confounders give you at the bunching point can be used to infer the endogeneity bias in other places. This is often testable. 6/
The second requirement is that we must be able to use the choice of TV on the positive side to infer what the choice of TV would have been in the negative side if there was no constraint. This is an out-of-sample prediction exercise. Read the next tweet. 7/
We went pretty far studying how to do this: partial identification, using families of distributions with parametric and nonparametric parameters, tail symmetry conditions, and using clustering methods. We also propose several tests of the assumptions. 8/
As for the TV, turns out that watching TV will not affect the child’s cognitive skills on the net, but will decrease non-cognitive skills.
In a linear model the correction is an estimated regressor which gets added to the main regression. Our estimation results allow for very general estimators of that regressor, and you can bootstrap the standard errors. go.usa.gov/xGQjw 10/
If you prefer to learn the method in an applied paper, we apply this technique in this paper: bit.ly/3mFQUIi.
• • •
Missing some Tweet in this thread? You can try to
force a refresh
I’m very interested to hear any thoughts on this paper. There was a version out in 2014, but seems way more complete now. Should choice of polynomial degree in RDD be endogenous? 1/8 …9-a-62cb3a1a-s-sites.googlegroups.com/site/peizhuan/…
In principle, this certainly makes sense. Near the threshold all processes we study are fairly linear. However, as we move out convexities and concavities appear and bias increases. If we did a quadratic instead of linear fit, we could go father away from the threshold. 2/8
Which means that if the degree of polynomial and bandwidth are chosen jointly, we could improve the variance without increasing bias. This is very tempting, especially with packaged software that does it for you. 3/8
Really important work. I worked with the PSID and know that the potential uses go way beyond what is currently being done with those datasets. It’s very hard data to collect well, but extremely useful to understand household decisions.
Here is why time use has potential beyond its current uses: bunching. For almost every time use activity there are plenty of people who do zero of it. This bunching can be used to test endogeneity in models and even to correct for endogeneity.
Which means that household decisions can be studied with time use data and you get some built in protections in dealing with endogeneity. Even better, multivariate bunching at zero for several activities creates pretty powerful endogeneity tests.
Our paper on the RDD with multiple endogenous variables and a single running variable and threshold is ready. I’m proud of this work with Gregorio Caetano and Juan Carlos Escanciano. arxiv.org/abs/2007.00185
We had some versions of this out before, but it took us a while do understand this problem fully. At the essence, it is an impossible problem. We show which assumptions must be done to gain each piece of information. 2/
For practitioners: our identification strategies are really, really easy to apply. 1 line Stata code (we even give the code in the paper). Standard errors and tests can be used directly from the software.