Snippet 1 from the #NeurIPS2020 tutorial: @balajiln What do we mean by uncertainty and out-of-distribution robustness? nips.cc/virtual/2020/t…
Snippet 2: On even simple benchmarks, neural networks are not only poor at generalizing to OOD but also degrade in their uncertainty estimates.
Snippet 3: There are a number of applications where uncertain & robust methods are already being used. It's at the heart of many AI & ML areas.
Snippet 4: @balajiln How do we measure the quality of uncertainty? Here's a toy example for weather forecasting.
Snippet 5: @balajiln Another popular way is to examine proper scoring rules, which assess the fit to the overall data distribution.
Snippet 6: @balajiln How do we measure the quality of robustness? For benchmarks, evaluate models according to a _general_ collection of shifts. This assesses the model's ability to be invariant to many real-world OOD (different magnitudes and OOD types). Don't overfit to one.
Snippet 7: @latentjasper Neural nets with SGD capture data uncertainty well, fitting an overall distribution to the data. But it doesn't capture model uncertainty.
Snippet 8: @latentjasper Probabilistic machine learning in one slide
Snippet 9: @latentjasper Bayesian neural networks with SGD follow a simple recipe. Check out the Approximate Inference symposium to dive even deeper approximateinference.org
Snippet 10: @latentjasper You can explicitly reason about predictive behavior by examining the NN correspondence to Gaussian processes.
Snippet 11: @latentjasper A perhaps more intuitive approach for model uncertainty is ensemble learning. Just build a collection of models and aggregate.
Snippet 12: @latentjasper Bayes and ensembles are not simply special cases of one another, enabling categorical wins for both fields. Important to understand each modality to analyze problems!
Snippet 13: @latentjasper To get Bayesian models to work well requires important considerations, e.g., interpretation, training dynamics, model specification.
Snippet 14: @latentjasper There are a number of simple baselines to try: temperature scaling, MC dropout, deep ensembles, hyperparameter ensembles, bootstrap, SWAG.
Snippet 15: @latentjasper Bootstrap flat out doesn't work for deep nets. Check out a very relevant @NeurIPS2020 workshop: …cant-believe-its-not-better.github.io @AaronSchein
Snippet 16: What's the core vision for uncertainty & robustness? Continually improve our methods as we move further in the age of computing
Snippet 17: Two recent directions advancing this frontier: marginalization, which aggregates multiple network predictions; and priors & inductive biases, which engrain predictive invariances with respect to shift.
Snippet 18: For marginalization, advance the frontier by understanding ensembles as a giant model. How do we better share information across the network?
Snippet 19: Bridge this gap by sharing parameters. Apply low-rank weight perturbations to decorrelate individual ensemble members!
Snippet 20: An simpler approach suggests obtaining strong ensembles may be possible purely through implicitly learned subnetwork paths. Treat the overall ensemble as a single model. But no perturbations required.
Snippet 21: How do we specify the prior? So.... many.. challenges. Thinking about functional behavior is often the most intuitive! But how do you actually enforce such behavior?
Snippet 21: Stop thinking about just probability distributions. Leverage the inductive biases of core DL techniques like data augmentation. They precisely encourage functional behavior!
Snippet 23: Another important invariance to capture: distance awareness!
Snippet 24: Open challenge of scale. This raises both existential questions and huge opportunities for robustness (diverse tasks) and uncertainty (model parallelism).
Snippet 25: Open challenge of understanding. We're on the cusp of breakthroughs bridging the gap from theory to practice with progress in analyses like generalization theory of deep nets.
Snippet 26: Open challenge of benchmarks. Announcing Uncertainty Baselines lead by
@zacharynado
to easily build upon well-tuned methods! Joint effort with
@OATML_Oxford github.com/google/uncerta…
Snippet 27: Open challenge of benchmarks. Announcing Robustness Metrics lead by @MarioLucic_ & Josip Djolonga to comprehensively study the robustness of DL models. Check it out to evaluate your trained models. github.com/google-researc…

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Dustin Tran

Dustin Tran Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @dustinvtran

6 Dec
Tomorrow @latentjasper @balajiln and I present a #NeurIPS2020 tutorial on "Practical Uncertainty Estimation and Out-of-Distribution Robustness in Deep Learning". Whether you're new to the area or an expert, there is critically useful info! 8-10:30a PT nips.cc/virtual/2020/t…
The talk is split into three sections: 1. Why Uncertainty & Robustness; 2. Foundations; and 3. Recent Advances.
Tutorials do _not_ need registration to attend!
Read 4 tweets
7 Sep
How I spent this weekend: upgrading my battlestation.
Cable management is a huge quality of life improvement.
KVM switch also helps for managing personal and work computers.
Read 4 tweets
13 Dec 19
Check out BatchEnsemble: Efficient Ensembling with Rank 1 Perturbations at the #NeurIPS2019 Bayesian DL workshop. Better accuracies and uncertainty than dropout and competitive with ensembles across a wide range of tasks. 1/- Image
It’s a drop in replacement for individual layers, like dropout, batchnorm, and variational layers and is available with baselines and Bayesian layers at github.com/google/edward2. 2/-
Unlike ensembles, it’s trained end to end under a single loss function (NLL) and computation can be parallelized across ensemble members in GPU/TPUs. BatchEnsemble is like a new parameterization for neural nets. 3/-
Read 6 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!

Follow Us on Twitter!