Tamay Profile picture
22 Nov, 12 tweets, 4 min read
A few months ago, I wrote an economics dissertation on whether machine learning models are getting harder to find. Here’s a summary of what I found:
Some background. @ChadJonesEcon, @johnvanreenen and others wrote an awesome article that found that ideas are getting harder to find: in semiconductors, agricultural production and medicine, research productivity has been declining steadily.
In my dissertation, I explored to how this story holds up for machine learning. I used a dataset on the top performing ML models on 93 machine learning benchmarks—mostly related to computer vision and NLP—and data on research input derived from data on publications.
Standard R&D-based growth models imply a tight relationship between technological progress and research effort. I exploited this to estimate a ‘knowledge production function’—commonly found in macro models—that describes how tech evolves with stock of existing knowledge.
I found that the marginal returns of researchers are rapidly declining. There is what’s called a “standing on toes” effect: researcher productivity declines as the field grows. Because ML has recently grown very quickly, this makes better ML models much harder to find.
On the other hand, I find that progress now makes progress in the future easier. This is called a “standing-on-the-shoulders” effect (innovations today are bootstrapped by previous progress).
A “standing-on-the-shoulders” effect in ML is on the whole not that surprising: it seems that finding one approach to solving one task can often be repurposed to solve other, related tasks (e.g. transformers, attention, etc.)
There are thus two conflicting effects: Adding researchers today results in reduced productivity of other researchers (a “standing-on-toes” effect). OTOH, additional researchers can make future researchers more productive by enabling them to ‘stand on their shoulders’.
It turns out that the “standing-on-toes” effect dominates. I estimate that overall research productivity declined by between 4% to 26% (depending on which sub-field and which model).
This is in line with (if not slightly higher than) @ChadJonesEcon, @johnvanreenen et al. Part of this drop in productivity is due to it becoming harder to beat SOTA performance on benchmarks that are close to ideal performance (e.g. nearing 100% accuracy), but most of it isn’t.
I compute counterfactual technology paths, to see what would have happened if getting closer to ideal performance did not negatively affect researcher productivity, and still mostly find yearly % declines of productivity in the double digits.
References. @ChadJonesEcon, @johnvanreenen's awesome paper: aeaweb.org/articles?id=10…

My dissertation can be accessed here: tamaybesiroglu.com/s/AreModels.pdf

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Tamay

Tamay Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!

Follow Us on Twitter!