🧵 on my replication of Moscona & Sastry (2023, QJE).
TL;DR: MS23 proxy 'innovation exposure' with a measure of heat. Using direct innovation measures from the paper’s own data decreases headline estimates of innovation’s mitigatory impact on climate change damage by >99.8%. 1/x
Moscona & Sastry (2023) reach two findings. First, climate change spurs agricultural innovation. Crops with croplands more exposed to extreme heat see increases in variety development and climate change-related patenting. 2/x
academic.oup.com/qje/article/13…
Second, MS23 find that innovation mitigates damage from climate change. They develop a county-level measure of 'innovation exposure' and find that agricultural land in counties with higher levels of 'innovation exposure' is significantly less devalued by extreme heat. 3/x Image
MS23 don’t claim that innovation will prevent all damage from climate change, but they do claim it’ll prevent a lot. They project that innovation prevented 20% of climate change-related agricultural land devaluation since 1960, and will abate 13% of such devaluation by 2100. 4/x
MS23 is one of the 81 papers I replicated for my job market paper, where I assess whether null results from top economics papers are robust to equivalence testing. This new replication focuses on something else. 5/x
Re-reading through MS23, I noticed that their ‘innovation exposure’ measure is actually a leave-one-out (LOO), area-weighted sum of other counties' extreme heat exposure. In fact, in their repository, they call this ‘innovation exposure’ variable 'LOO extreme heat exposure'. 6/x Image
In other words, MS23 use LOO extreme heat exposure as a direct measure of innovation. Their 'innovation exposure' variable is a measure of heat, not innovation. Resultantly, their estimates of the mitigatory impacts of 'innovation exposure' have nothing to do with innovation. 7/x
Unsurprisingly, LOO extreme heat exposure is virtually identical to local extreme heat exposure. These variables move in linear lockstep, and are virtually unit elastic. Regressing LOO extreme heat exposure on local extreme heat exposure gives t-statistics in the 40s and 50s. 8/x Image
MS23 estimate mitigatory impacts by interacting LOO extreme heat exposure with local extreme heat exposure. But this is basically interacting local extreme heat exposure *with itself*. The positive coefficient on that interaction term (ϕ) has nothing to do with innovation. 9/x Image
Extreme heat has nonlinear impacts on agricultural value. Near optimal growing temperatures, marginal increases in extreme heat can kill lots of crops. But if it’s already 40°C in May, an extra degree doesn't lose you that much more yield. 10/x Image
The interaction between local and LOO extreme heat is really a square term in extreme heat, which is positive because a parabola fit to this graph faces upwards. MS23’s estimates of ‘innovation’s mitigatory impact’ just reflect decelerating marginal effects of extreme heat. 11/x Image
You can nearly identically replicate MS23's qualitative conclusions on 'innovation's mitigatory impact' by regressing agricultural land value on a second-order polynomial in local extreme heat exposure. Estimates are smaller, but conclusions nearly perfectly line up. 12/x Image
MS23 try justifying the proxy. They say LOO computation purges the proxy of correlations with local heat, but we've seen that's not true. They run robustness checks on the second-order polynomial, but they're incorrectly specified and don't replicate (more on that later). 13/x
They have a robustness check using leave-*state*-out computation instead of LOO computation, but the relationship between leave-state-out and LOO extreme heat exposures is practically as tight as the relationship between local and LOO extreme heat exposures. 14/x Image
But you may have already noticed that MS23 never needed a proxy to estimate the mitigatory impacts of innovation. They got their first set of findings using *direct* measures of innovation, specifically crop variety development and climate change-related patenting. 15/x
So I use their innovation data to develop direct county-level measures of innovation exposure, specifically variety exposure and patent exposure. I then replicate MS23's main mitigatory impact estimates using these variables instead of LOO extreme heat exposure. 16/x Image
Image
In standardized units, replacing the heat proxy with direct measures of innovation exposure decreases MS23’s estimates of the mitigatory impact of innovation exposure by at least 99.8%. None of these estimates are statistically significantly different from zero. 17/x Image
Image
Maybe innovation is just more endogenous than weather patterns. So I pursue an IV strategy, instrumenting my direct innovation measures with LOO extreme heat exposure. Per MS23’s justifications, LOO extreme heat exposure should be a good exogenous instrument for innovation. 18/x
Thinking about this in an instrumental variables setting makes the proxy's problems intuitive. There's a clear exclusion restriction violation going on here. LOO extreme heat exposure impacts agricultural value not just through innovation, but also through extreme heat. 19/x
The instrument also ends up being pretty weak, with first-stage F-statistics all less than eight. Instrumental variables estimates for mitigatory impacts are somewhat more positive, but none are statistically significantly different from zero. 20/x Image
Image
A few points on MS23's response. First, they claim that I ignore their rationales for the proxy. *A third of my paper* is dedicated to the section titled “Justifications for the Proxy”, which directly addresses MS23's empirical justifications from their paper and reply. 21/x Image
Second, MS23 claim that I exactly replicate their paper's results, that the replicability issues I find are minuscule (rounding errors in deep decimal places), and that my classification of the paper as 'partially replicable' is misleading. None of these claims are true. 22/x Image
The reason I maintain my 'partially replicable' classification is because of replicability problems for an important robustness check in the appendix, one that is supposed to address the exact concern that underpins my entire replication. 23/x
In Appendix Table A18, MS23 report mitigatory impact estimates after controlling for a square term in local extreme heat exposure. This is supposed to rule out my main critique, that MS23's mitigatory impact estimates capture nonlinear impacts of local extreme heat exposure. 24/x
There is no replication code for Appendix Table A18, and my best attempts to reproduce it yield mitigatory impact estimates that are up to 38.3% weaker than those in the published version of the table. 25/x Image
The published Appendix Table A18 implies that if anything, controlling for squared local extreme heat exposure *increases* estimates of innovation's mitigatory impacts. My replication of that table implies that this control actually *decreases* those estimates considerably. 26/x Image
I also find that the models in Appendix Table A18 are misspecified (see Section 3.1.4 of my paper for details). When this is corrected, controlling for squared local extreme heat exposure leaves most estimates of innovation's mitigatory impact not stat. sig. diff. from zero. 27/x Image
Third and finally, MS23 imply that my paper actually reaches the same conclusion as theirs: that innovation has a relatively weak mitigatory impact on agricultural damage from climate change. Our papers do not reach the same conclusion. 28/x Image
As I mentioned, MS23 conclude that market innovation has prevented about 20% of the potential agricultural damages of climate change since 1960. This is obviously a meaningfully large effect, and my main replication reduces this estimate by at least 99.8%. 29/x
MS23’s paper says: market innovation is useful for abating some climate change damage, but don’t trust it to abate *everything*. My paper says: don’t trust it to abate *anything*. The market won’t save us. We need collective action on climate change. 30/31
I will be submitting this replication report to @QJEHarvard, and will keep posted about the progress of that submission. The most recent version of the paper will be updated at . 31/31jack-fitzgerald.github.io/files/MS23_Rep…
@QJEHarvard Also, if you care enough about this issue to read this far, please consider donating to All Faiths Food Bank, a local charity in Sarasota, where Hurricane Milton made landfall.
allfaithsfoodbank.org
@QJEHarvard @threadreaderapp unroll this

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Jack Fitzgerald | @jackfitzgerald.bsky.social

Jack Fitzgerald | @jackfitzgerald.bsky.social Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @FitzgeraldJack_

Dec 20, 2024
A new working paper for holiday reading! @peder_isager and I provide an introduction to three-sided testing, a framework for testing an estimate's practical significance. We offer a tutorial, Shiny app, + commands/code in #Rstats, Jamovi, and #Stata (🔗 below!) 1/9

#EconTwitter Image
Equivalence testing lets us test whether estimates are stat. sig. bounded beneath practically negligible effect size Δ (e.g., pink estimate). But estimates can be both stat. sig. diff. from zero and stat. sig. bounded beneath Δ. 2/9 Pink estimate is statistically significantly bounded between the delta bounds. Blue estimate is statistically significantly bounded above the upper delta bound. The confidence interval of the orange estimate intersects one of the delta bound, but does not cross zero.
Estimates can also be stat. sig. bounded outside of Δ (e.g., blue estimate). What should we conclude about estimates like these blue/orange estimates? Standard equivalence testing frameworks don't give us clear answers. We introduce researchers to a framework that does. 3/9 Pink estimate is statistically significantly bounded between the delta bounds. Blue estimate is statistically significantly bounded above the upper delta bound. The confidence interval of the orange estimate intersects one of the delta bound, but does not cross zero.
Read 10 tweets
Nov 18, 2024
Do real stakes/incentives matter in experiments? Recent studies say they don’t. My new paper shows that these studies’ results — and those of most hypothetical bias experiments — are uninformative when we care about experimental treatment effects. 1/x
🔗: papers.tinbergen.nl/24070.pdfImage
Historically, experimental economists virtually always tied experimental choices to real stakes/payoffs to improve generalizability. That’s changing: many economists now use hypothetical stakes in online experiments + large-scale survey experiments. 2/x
There’s also recently been a wave of new studies showing that certain outcomes don’t stat. sig. differ between real-stakes and hypothetical-stakes experiments. These results are affecting thinking at the highest levels of experimental economics. 3/x Image
Read 21 tweets
Oct 2, 2024
My paper is out in @PNASNews! I replicate a paper on the impact of COVID vaccine mandates on vaccine uptake. Removing a single bad control variable sign-flips several of the paper’s headline results. The reply’s findings are also not robust. 1/x
pnas.org/doi/10.1073/pn…
@PNASNews Rains & Richards (2024) — henceforth RR24 — reach two findings. First, RR24 claim that difference-in-differences estimates show that US state COVID vaccine mandates had imprecise impacts on COVID vaccine uptake. 2/x
pnas.org/doi/10.1073/pn…
@PNASNews Second, RR24 find that states that mandated COVID vaccination statewide now see lower uptake of COVID boosters and both adult + child flu vaccines than states that banned local COVID vaccine mandates. 3/x
Read 25 tweets
Jul 22, 2024
🚨 WP alert! 🚨 I design equivalence tests for running variable (RV) manipulation in regression discontinuity (RDD), show that serious RV manipulation can't be ruled out in lots of published RDD research, and offer the lddtest command in Stata/R. 1/x

🔗: hdl.handle.net/10419/300277Image
Credible RDD estimation relies on the assumption that agents can’t endogenously sort their RVs to opt themselves into or out of treatment. If they can, then RDD estimates are confounded: agents who manipulate RVs are likely different in important ways from agents who don't. 2/x
Such manipulation often causes jumps in RV density at the cutoff, which can either come from genuine distributional distortions or from strategic reporting. E.g., consider the French examples below. 3/x

Read 16 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(