So much of what we hear around RCTs are exciting stories about how evidence is used to inform policy. Which is awesome! I love evidence-informed policy. However, I'm sure many of us have also had experiences that are different, and more challenging.
In the spirit of transparency, wanted to share some different (anonymized) stories about use of evidence. Short 🧵
#1: program has mixed effects (largely null for downstream outcomes). Funding for the implementer concludes, implementer and funder move on. What happens? Brief discussion, draft paper shared (0 replies), almost no policy learning (hopefully research community benefits).
#2: null effects. Findings are rejected. Other researchers engaged as consultants to rebut our findings. Despite ex ante agreement on evaluation design and outcomes, endless discussions are held about how the evaluation was wrong;
was not designed to measure the right effects; could never have in any reality possibly measured the hypothetical very important outcomes the program definitely affected; etc.
#3: intervention has meaningful effects, findings disseminated . Awesome, right? It is, but interestingly a lot of the discussion focuses defensively on outcomes where there wasn’t an impact.
Not sure what the take-away is here. I sympathize: no one likes to hear bad news. That being said, for researchers these discussions can be discouraging. FWIW, these included partners and funders that have an active engagement in research (not evaluation “newbies”).
Another puzzling (related) trend I’ve noticed is the emphasis on qual research as a “insurance policy”: i.e., add a qual substudy so we'll be sure to have some good news. I’m not a qual researcher, but fairly sure that doesn’t make sense.
My fear is that while we hope there are reputational benefits for participating in research for partners, in any particular project those benefits can be uncertain, and the risks of publicizing discouraging program effects (with implications for funding etc.) dominate.
If so, the implications for selection into who participates in evaluations and what is therefore learned are substantial. If it’s going to be easier to share disappointing results, perhaps the reputational benefits of just conducting high-quality research need to grow, a lot.
As new PhD students start to look forward to their first year, short 🧵 on challenges in collaboration in grad school (and its potentially gendered dimensions).
Many people advise grad students to rely on their classmates: first in coursework, later on projects / as coauthors.
I endorse that advice! But it can also be hard to follow. I attended two grad programs (MPhil and PhD) and had similar experiences in both. There were large, energetic, overlapping-networks problem set groups that formed quickly.
They were mostly dominated by men (unsurprisingly; econ grad programs are mostly dominated by men) and, to describe it neutrally, had a fast-paced style. Always an introvert who was becoming more so, I was uncomfortable and anxious about trying to participate.
Enjoyed the presentation by @elianalaferrara today at World Bank DIME of work joint with Baumgartner, Rosa-Dias, Breza and my awesome coauthor Victor Orozco: evidence around a peer education program targeting early sexual activity teen pregnancy in Brazil.
The authors have a fascinating evaluation comparing a peer educator program with three alternate selection mechanisms for educators (school-driven; selection via peer nomination of popularity; selection by centrality in a formally mapped network) to a control arm.
In general, the peer education program is very effective: ⬆️ knowledge and communication around sexuality, contraceptive use; ⬇️ teen pregnancy. The peer educators chosen by schools (the default method), however, were generally ineffective!
Lately I’ve been thinking more and participating in various conversations about how USAID commissions, uses research. Huge topic! But wanted to do a short🧵on what I’ve learned. 1/n
First and foremost in the hearts of most economists is DIV. DIV is awesome, as many others have pointed out! See this recent blog by @DaveEvansPhD and colleagues 2/n cgdev.org/blog/case-evid…
But, DIV primarily funds evaluation of pilots and other interventions that are implemented outside of USAID and are not directly related to the work of missions – as summarized above (there are some exceptions). In that sense it is often separate from the main aid portfolio 3/n
After doing a lot of reformatting to meet a journal page limit (which, TBC, I support), I started to wonder - why don't journals impose limits on the referee reports that lead to these long papers? E.g., 1-2 pages; or alternatively, 3-5 (choose N) substantive suggested revisions
Seems like this could help with a lot of problems - long review times, tedious revisions, bloated papers that are hard to read, indigestible appendices, etc. Hard to enforce, but editors could suggest that material beyond the limit would be ignored.
Plus, I suspect many referees would be very happy for explicit guidance that allows for coordination, since (many of us) are concerned about the perception of submitting an un-thorough or low-quality report relative to others. . .
Happy to see the latest JEP table of contents and took a look at the @bfjo paper on teamwork right away (h/t @jenniferdoleac). Really fascinating and some striking graphics on the rise of teamwork in econ; short 🧵 #EconTwitter
On a subfield note, was very surprised to see development was significantly under the average for team size in the 1980s, though it has now converged up. Anecdotally it seems like much larger teams (5+, 10+) starting to surface in dev, still rare in the profession at large
Also a thoughtful discussion of questions about credit, attribution and equity. Lots of evidence already that some team members (particularly women) receive less credit than others, e.g. Sarsons et al. journals.uchicago.edu/doi/abs/10.108…
New week, new twitter project! In recent years more and more randomized trials have analyzed interventions targeted at non-cognitive skills (soft skills, life skills, socio-emotional skills) broadly defined in developing countries. 1/n
This is a major interest of mine – I’ve decided to start a running thread quickly summarizing and linking to papers of interest. Please add links to other papers, including your own – I’ll add them. 2/n
Two today. First, Acevedo, Cruces, @paul_gertler, and Martinez analyzed a soft skills and vocational skills training program in the DR. Targeted skills include grit (perseverance, ambition) and social competencies (leadership, conflict resolution, social skills, empathy) 3/n