andy jones Profile picture
engineering & research at @AnthropicAI. DC, SF

Apr 8, 2021, 11 tweets

🚨 I've a paper out today: Scaling Scaling Laws with Board Games! 🚨

arxiv.org/abs/2104.03113

Principle result is that by studying a sequence of small problems in ML, I could predict the outcome of experiments on orders-of-magnitude larger problems 🤯

I worked on Hex. Hex is a board game, with all the strategic depth of Go but also a much simpler rule set. Crucially, Hex on small boards is easy, and Hex on big boards is hard!

I wrote a fast, all-GPU version of AlphaZero, and used it to train ~200 different neural nets across a bunch of board sizes. Plotted together, the best-performing nets at each level of compute form a steady trend, the *compute frontier*

All of these frontiers can be explained by a simple family of curves, which say

* If you've 2x the compute of your opponent, you've a 2/3 chance of winning

* Adding +1 to the board size makes perfect play 7x harder

And this holds across ~6 orders of magnitude!

This similarity across scales means that curves fit on small, cheap board sizes are excellent predictors of the compute frontiers at bigger board sizes.

In fact, the error appears to decay exponentially as you add more small board sizes:

So what's the takeaway from all this? There's no generality here yet, but it's a proof of concept, a proof that you can study the small and make claims about the large.

If you're a resource-bound researcher and you want to study big models, I think the scaling laws paradigm is something to dive into.

Bonus plot: while doing all the above, I was prodded into looking at how train-time and test-time compute trade off. Quite strikingly, at every level of performance you can knock off 15x test compute, add 10x train compute, and keep the same performance!

And a final acknowledgement: I'm officially an independent researcher, but the level of support I got from the ML and games community was incredible.

Only a subset of the folk I'd like to thank hang out on Twitter, but @paulfchristiano @janleike @sharky6000 @JacobHHilton @ClemensWinter all went out of their way to help me out with this, and it's dramatically better for their help.

The other group I need to thank are the folks on the RL Discord. I posted a draft Monday evening, and got a dozen researchers offering me advice on how to improve it. I only got ~half of that advice in in time, but that half made a huge difference.

discord.gg/xhfNqQv

Share this Scrolly Tale with your friends.

A Scrolly Tale is a new way to read Twitter threads with a more visually immersive experience.
Discover more beautiful Scrolly Tales like this.

Keep scrolling