Several comments on this paper. First, it's nice to see someone taking the units of measurement issue seriously. But I still see many issues, especially when y >= 0 and we have better alternatives.

1. A search is required over units of measurement.

#metricstotheface
How do a compute a legitimate standard error of, say, an elasticity? I've estimated theta but then I ignore the fact that I estimated it? That's not allowed.

2. As with many transformation models, the premise is there exists a transformation g(.) such that g(y) = xb + u.
u is assumed to be indep of x, at a minimum. Often the distrib is restricted. In 1989 in an IER paper I argued this was a real problem with Box-Cox approaches b/c u >= -xb. If I model E(y|x) directly I need none of that. It's what Poisson regression does.
3. I'm puzzled by the application to Lalonde, which is an experiment with y >= 0. The experimental estimate of the semi-elasticity is .394 (se = .164); we can take this as the target. On the plus side, you conclude it's likely between .31 and .38 (but no stnd error).
In about 2 minutes I had the Poisson estimate that controls for 1975 earnings: .384 (se = .162). I don't much care about the elasticity wrt 1975 earnings, but I get .056 using arcsinh(re75).
So I can use Poisson to essentially recover the experimental value and obtain a valid se in 2 minutes, or I can go through a complicated data mining exercise that doesn't produce a valid standard error -- and implicitly imposes strong assumptions. Why should I do this?

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Jeffrey Wooldridge

Jeffrey Wooldridge Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @jmwooldridge

9 May
A year ago on Facebook, at the request of a former MSU student, I made this post. I used to say in class that econometrics is not so hard if you just master about 10 tools and apply them again and again. I decided I should put up or shut up.

#metricstotheface
I cheated by combining tools that are connected, so there are actually more than 10 ....
1. Law of Iterated Expectations, Law of Total Variance
2. Linearity of Expectations, Variance of a Sum
3. Jensen's Inequality, Chebyshev’s Inequality
4. Linear Projection and Its Properties
5. Weak Law of Large Numbers, Central Limit Theorem
6. Slutksy's Theorem, Continuous Convergence Theorem, Asymptotic Equivalence Lemma
7. Big Op, Little op, and the algebra of them.
Read 5 tweets
6 May
Have we yet figured out when we should include a lagged dependent variable in either time series or panel data models when the goal is to infer causality? (For forecasting the issue is clear.) Recent work on macroeconomics on causal effects is a positive sign.

#metricstotheface
And the answer cannot be, "Include y(t-1) if it is statistically significant." Being clear about potential outcomes and the nature of the causal effects we hope to estimate are crucial. I need to catch up on this literature and I need to think more.
In panel data settings, if our main goal is to distinguish state dependence from heterogeneity, clearly y(t-1) gets included. But what if our interest is in a policy variable? Should we hold fixed y(t-1) and the heterogeneity when measuring the policy effect?
Read 6 tweets
2 May
When I teach the senior seminar to economics students I sometimes take 20 minutes to discuss common grammar mistakes. I was worried it would come off as patronizing (even obnoxious), but I actually got positive comments about it on my teaching evaluations.

#grammartotheface
I set it up as a quiz. Samples:

1. James will present his report to Kayla and I.
2. James will present his report to Kayla and me.
3. James will present his report to Kayla and myself.

1. He should not have went home early today.
2. He should not have gone home early today.
1. I should of taken less cookies.
2. I should’ve taken fewer cookies.
3. I should’ve taken less cookies.

1. She and I are going to visit my parents.
2. Her and I are going to visit my parents.
3. Her and me are going to visit my parents.
Read 4 tweets
30 Apr
Tests that should be retired from empirical work:

Durbin-Watson statistic.
Jarque-Bera test for normality.
Breusch-Pagan test for heteroskedasticity.
B-P test for random effects.
Nonrobust Hausman tests.

I feel uncomfortable using names, but this is #metricstotheface.
D-W test only gives bounds. More importantly, it maintains the classical linear model assumptions.
J-B is an asymptotic test. If we can use asymptotics then normality isn't necessary.
B-P test for heteroskedasticity: maintains normality and constant conditional 4th moment.
B-P test for RE: maintains normality and homoskedasticity but, more importantly, detects any kind of positive serial correlation.
Nonrobust Hausman: maintains unnecessary assumptions under the null that conflict with using robust inference. Has no power to test those assumps.
Read 5 tweets
27 Apr
Historically, economics has fallen into the bad habit of thinking the fancier estimation method is closer to being "right" -- based on one sample of data. We used to think this of GLS vs OLS until we paid careful attention to exogeneity assumptions.

#metricstotheface
We used to think this of 2SLS vs OLS until problems with weak instruments were revealed.

We still seem to think matching methods are somehow superior to regression adjustment if they give somewhat different estimates.
And we now seem to think ML methods are to be preferred over basic methods. Perhaps a reckoning is on its way.
Read 4 tweets
24 Apr
Yesterday I was feeling a bit guilty about not teaching lasso, etc. to the first-year PhD students. I'm feeling less guilty today. How much trouble does one want to go through to control for squares and interactions for a handful of control variables?

#metricstotheface
And then it gets worse if I want my key variable to interact with controls. You can't select the variables in the interactions using lasso. I just looked at an application in an influential paper and a handful of controls, some continuous, were discretized.
Discretizing eliminates the centering problem I mentioned, but in a crude way. So I throw out information by arbitrarily using five age and income categories so I can use pdslasso? No thanks.
Read 8 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!

Follow Us on Twitter!

:(