My Authors
Read all threads
Survey of #MachineLearning experimental methods (aka "how do ML folks do their experiments") at #NeurIPS2019 and #ICLR2020, a thread of results:
1. "Did you have any experiments in your paper?"

The future is empirical! If we historically look at NeurIPS papers (not just 2019), the number of theoretical submissions is dwindling and now almost relegated to conferences like UAI, and that's unfortunate.
side note: There was a time when folks used to say, "what experiments? It's a NIPS paper!" (also, I am a dinosaur).
2. "Did you optimize your hyperparameters?"

With compute costs coming down, it is becoming more affordable to run hyperparameter optimization, as long as you stay away from the Sesame Street. It would be interesting to condition this based on where the authors are coming from.
3. "If yes, how did you tune them?"

lol, manual tuning of course! Graduate Student Optimization FTW! Also, it points to the brittleness of results in our community. Carefully (or luckily?) manually-tuned hyperparams on one dataset fails to deliver on your problem. Enjoy!
4. "How many hyperparameters did you optimize?"
who are all these folks tuning 1 hyperparam? :)
5. "How many trials/experiments in total during the optimization? (ie. how many different sets of hyperparameters
were evaluated)"
Pretty sure >100 trials are googlers. 💸💸💸
6. "Did you optimize the hyperparameters of your baselines?"
No, you didn't 😛 The "With optimization" group is l̶y̶i̶n̶g̶. The "Not applicable" group did not do lit. review to even consider a baseline.😇
7. "How many baselines (models, algorithms) did you compare
with?"
3-5 baselines most popular response?🤔 Also, why are the "Not applicable" counts different here from the previous chart?
8. "How many datasets or tasks did you compare on?"
Interestingly, 3-5 is also the most popular answer here .. omne trium perfectum?
9. "How many results did you report for each model? (ex:
for different seeds)"
All the the "1-sample" people have minimum variance unbiased estimators 😄
10. If you answered 1 to the previous question, did you use
the same seed in all your experiments? (If you did not
seed your experiments, you should answer ’no’ to this
question.)
Green people -> 🤷‍♂️🤷🤷‍♀️
Finally thanks to @bouthilx and @GaelVaroquaux for putting this in a report🙏. Read their report here: hal.archives-ouvertes.fr/hal-02447823/d…
Missing some Tweet in this thread? You can try to force a refresh.

Enjoying this thread?

Keep Current with Delip Rao

Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

Twitter may remove this content at anytime, convert it as a PDF, save and print for later use!

Try unrolling a thread yourself!

how to unroll video

1) Follow Thread Reader App on Twitter so you can easily mention us!

2) Go to a Twitter thread (series of Tweets by the same owner) and mention us with a keyword "unroll" @threadreaderapp unroll

You can practice here first or read more on our help page!

Follow Us on Twitter!

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just three indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3.00/month or $30.00/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!