In 2018 I was invited to give a talk at SOCHER in Chile, to give my opinions about using spatial methods for policy analysis. I like the idea of putting in spatial lags of policy variables to measure spillovers. Use fixed effects with panel data, compute fully robust ses.
For the life of me, I couldn't figure out how putting in spatial lags of Y had any value. After preparing a course in July 2020, I was even more negative about this practice. It seems an unnecessary complication developed by theorists.
As far as I can tell, when spatial lags in Y are used, one always computes the effects of own policy changes and neighbor policy changes, anyway, by solving out. This is done much more robustly and much more easily modeling spillovers directly without spatial lags in Y.
Writing y(i) = a + b*w(i)Y + x(i)*c + u(i) makes no sense. It has no ceteris paribus interpretation because there's no interesting counterfactual. It's not at all like time series; that connection is superficial. Spatial lags in policy variables has a potential outcomes interp.
I made these points in my SOCHER talk and I don't think they went over well. But they were very nice to me and they didn't leave me in the Atacama Desert.
Imagine a model like
y(i) = a + b*w(i)Y + x(i)*c + w(i)X*d + u
What's the thought experiment for d? We hold y(j) fixed to see the effect that changing x(j) on y(i)? But x(j) directly affects y(j), so how can we hold y(j) fixed?
• • •
Missing some Tweet in this thread? You can try to
force a refresh
I taught a bit of GMM for cross-sectional data the other day. In the example I used, there was no efficiency gain in using GMM with a heteroskedasticity-robust weighting matrix over 2SLS. I was reminded of the presentation on GMM I gave 20 years ago at ASSA.
The session was organized by the AEA, and papers were published in the 2001 JEL issue "Symposium on Econometric Tools." Many top econometricians gave talks, and I remember hundreds attended. (It was a beautiful audience. The largest ever at ASSA. But ASSA underreported the size.)
In my talk I commented on how, for standard problems -- single equation models estimated with cross-sectional data, and even time series data -- I often found GMM didn't do much, and using 2SLS with appropriately robust standard errors was just as good.
I think frequentists and Bayesians are not yet on the same page, and it has little to do with philosophy. It seems some Bayesians think a proper response to clustering standard errors is to specify an HLM. But in the linear case, HLM leads to GLS, not OLS.
Moreover, a Bayesian would take the HLM structure seriously in all respects: variance and correlation structure and distribution. I'm happy to use an HLM to improve efficiency over pooled estimation, but I would cluster my standard errors, anyway. A Bayesian would not.
There still seems to be a general confusion that fully specifying everything and using a GLS or joint MLE is a costless alternative to pooled methods that use few assumptions. And the Bayesian approach is particular unfair to pooled methods.
In such cases, we have a clear tradeoff between consistency and efficiency.
In models additive in endogenous explanatory variables with constant coefficients, CF reduces to 2SLS or FE2SLS -- which is neat. Of course, the proof uses Frisch-Waugh.
The equivalence between CF and 2SLS implies a simple, robust specification test of the null that the EEVs are actually exogenous. One can use "robust" or Newey-West or "cluster robust" very easily. The usual Hausman test is not robust, and suffers from degeneracies.
If you teach prob/stats to first-year PhD students, and you want to prepare them to really understand regression, go light on measure theory, counting, combinatorics, distributions. Emphasize conditional expectations, linear projections, convergence results.
@metricstotheface.
This means, of course, law of iterated expectations, law of total variance, best MSE properties of CEs and LPs. How to manipulate Op(1) and op(1). Slutsky's theorem. Convergence in distribution. Asymptotic equivalence lemma. And as much matrix algebra as I know.
If you're like me -- and barely understand basic combinatorics -- you'll also be happier. I get the birthday problem and examples of the law of very large numbers -- and that's about it.
When I teach regression with time series I emphasize that even if we use GLS (say, Prais-Winsten), we should make standard errors robust to serial correlation (and heteroskedasticity). Just like with weighted least squares.
I like the phrase "quasi-GLS" to emphasize, in all contexts, we shouldn't take our imposed structure literally. In Stata, it would be nice to allow this:
prais y x1 x2 ... xK, vce(hac nw 4)
vce(robust) is allowed, but it's not enough. The above would be easy to add.
To its credit, Stata does allow
reg y x1 ... xK [aweight = 1/hhat], vce(robust)
to allow our model of heteroskedasticity, as captured by hhat, to be wrong. I've pushed this view in my introductory econometrics book.