Rohan Taori Profile picture
research @AnthropicAI | phd from @StanfordAILab🌲| proud @Cal bear 🐻 | taught w @BerkeleyML

Sep 15, 2022, 6 tweets

🎉 The last few weeks have seen the release of #StableDiffusion, #OPT, and other large models.

⚠️ But should we be concerned about an irreversible influx of AI content on the internet?

⚙️ Will this make it harder to collect clean training data for future AI models?

🧵👇 1/6

(thread based on recent work arxiv.org/pdf/2209.03942…)

Q: So what’s the root issue?

A: Biases in AI models will be represented in their outputs, which become *training data* for future models! (if we’re not careful).

These feedback cycles have the potential to get nasty.

2/6

A concrete example -

Generating from a language model with beam search is known to be repetitive/disfluent.

Under feedback (where a model is re-trained on its outputs), this problem very quickly magnifies by 2-3x!

Nucleus sampling, OTOH, is surprisingly stable.

3/6

In fact, we find that the more a model behaves like a sampler, the more stable it is.

We can connect samplers to necessary conditions for stability, which imply bounded bias amplification *even in the limit of infinite time feedback*.

This result has wider consequences -

4/6

Take image classifiers. Classifiers pick the argmax vote (like beam search), so perhaps they behave poorly.

In fact, the opposite is true bc classifiers actually kind of behave like samplers, a finding by @PreetumNakkiran & @whybansal called “Distributional Generalization”.

5/6

Lots more discussion & experiments in paper with @tatsu_hashimoto - arxiv.org/abs/2209.03942

More on:
1) when sampling-like behavior appears naturally,
2) what this means for bias amplification on the internet, and
3) how to induce stability in otherwise unstable systems.

6/6

Share this Scrolly Tale with your friends.

A Scrolly Tale is a new way to read Twitter threads with a more visually immersive experience.
Discover more beautiful Scrolly Tales like this.

Keep scrolling