So much of what we hear around RCTs are exciting stories about how evidence is used to inform policy. Which is awesome! I love evidence-informed policy. However, I'm sure many of us have also had experiences that are different, and more challenging.
In the spirit of transparency, wanted to share some different (anonymized) stories about use of evidence. Short 🧵
#1: program has mixed effects (largely null for downstream outcomes). Funding for the implementer concludes, implementer and funder move on. What happens? Brief discussion, draft paper shared (0 replies), almost no policy learning (hopefully research community benefits).
#2: null effects. Findings are rejected. Other researchers engaged as consultants to rebut our findings. Despite ex ante agreement on evaluation design and outcomes, endless discussions are held about how the evaluation was wrong;
was not designed to measure the right effects; could never have in any reality possibly measured the hypothetical very important outcomes the program definitely affected; etc.
#3: intervention has meaningful effects, findings disseminated . Awesome, right? It is, but interestingly a lot of the discussion focuses defensively on outcomes where there wasn’t an impact.
Not sure what the take-away is here. I sympathize: no one likes to hear bad news. That being said, for researchers these discussions can be discouraging. FWIW, these included partners and funders that have an active engagement in research (not evaluation “newbies”).
Another puzzling (related) trend I’ve noticed is the emphasis on qual research as a “insurance policy”: i.e., add a qual substudy so we'll be sure to have some good news. I’m not a qual researcher, but fairly sure that doesn’t make sense.
My fear is that while we hope there are reputational benefits for participating in research for partners, in any particular project those benefits can be uncertain, and the risks of publicizing discouraging program effects (with implications for funding etc.) dominate.
If so, the implications for selection into who participates in evaluations and what is therefore learned are substantial. If it’s going to be easier to share disappointing results, perhaps the reputational benefits of just conducting high-quality research need to grow, a lot.

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Jessica Leight

Jessica Leight Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @leightjessica

25 Jun
As new PhD students start to look forward to their first year, short 🧵 on challenges in collaboration in grad school (and its potentially gendered dimensions).
Many people advise grad students to rely on their classmates: first in coursework, later on projects / as coauthors.
I endorse that advice! But it can also be hard to follow. I attended two grad programs (MPhil and PhD) and had similar experiences in both. There were large, energetic, overlapping-networks problem set groups that formed quickly.
They were mostly dominated by men (unsurprisingly; econ grad programs are mostly dominated by men) and, to describe it neutrally, had a fast-paced style. Always an introvert who was becoming more so, I was uncomfortable and anxious about trying to participate.
Read 13 tweets
24 Jun
Enjoyed the presentation by @elianalaferrara today at World Bank DIME of work joint with Baumgartner, Rosa-Dias, Breza and my awesome coauthor Victor Orozco: evidence around a peer education program targeting early sexual activity teen pregnancy in Brazil.
The authors have a fascinating evaluation comparing a peer educator program with three alternate selection mechanisms for educators (school-driven; selection via peer nomination of popularity; selection by centrality in a formally mapped network) to a control arm.
In general, the peer education program is very effective: ⬆️ knowledge and communication around sexuality, contraceptive use; ⬇️ teen pregnancy. The peer educators chosen by schools (the default method), however, were generally ineffective!
Read 6 tweets
3 Jun
Lately I’ve been thinking more and participating in various conversations about how USAID commissions, uses research. Huge topic! But wanted to do a short🧵on what I’ve learned. 1/n
First and foremost in the hearts of most economists is DIV. DIV is awesome, as many others have pointed out! See this recent blog by @DaveEvansPhD and colleagues 2/n
cgdev.org/blog/case-evid…
But, DIV primarily funds evaluation of pilots and other interventions that are implemented outside of USAID and are not directly related to the work of missions – as summarized above (there are some exceptions). In that sense it is often separate from the main aid portfolio 3/n
Read 15 tweets
1 Jun
After doing a lot of reformatting to meet a journal page limit (which, TBC, I support), I started to wonder - why don't journals impose limits on the referee reports that lead to these long papers? E.g., 1-2 pages; or alternatively, 3-5 (choose N) substantive suggested revisions
Seems like this could help with a lot of problems - long review times, tedious revisions, bloated papers that are hard to read, indigestible appendices, etc. Hard to enforce, but editors could suggest that material beyond the limit would be ignored.
Plus, I suspect many referees would be very happy for explicit guidance that allows for coordination, since (many of us) are concerned about the perception of submitting an un-thorough or low-quality report relative to others. . .
Read 4 tweets
3 May
Happy to see the latest JEP table of contents and took a look at the @bfjo paper on teamwork right away (h/t @jenniferdoleac). Really fascinating and some striking graphics on the rise of teamwork in econ; short 🧵 #EconTwitter
On a subfield note, was very surprised to see development was significantly under the average for team size in the 1980s, though it has now converged up. Anecdotally it seems like much larger teams (5+, 10+) starting to surface in dev, still rare in the profession at large
Also a thoughtful discussion of questions about credit, attribution and equity. Lots of evidence already that some team members (particularly women) receive less credit than others, e.g. Sarsons et al.
journals.uchicago.edu/doi/abs/10.108…
Read 4 tweets
12 Apr
New week, new twitter project! In recent years more and more randomized trials have analyzed interventions targeted at non-cognitive skills (soft skills, life skills, socio-emotional skills) broadly defined in developing countries. 1/n
This is a major interest of mine – I’ve decided to start a running thread quickly summarizing and linking to papers of interest. Please add links to other papers, including your own – I’ll add them. 2/n
Two today. First, Acevedo, Cruces, @paul_gertler, and Martinez analyzed a soft skills and vocational skills training program in the DR. Targeted skills include grit (perseverance, ambition) and social competencies (leadership, conflict resolution, social skills, empathy) 3/n
Read 55 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!

Follow Us on Twitter!

:(