Alex Ratner Profile picture
Apr 2 10 tweets 4 min read Twitter logo Read on Twitter
1/ Prediction: Everyone will soon be using foundation models (FMs) like GPT-4.

However, they'll be using FMs trained on their own data & workloads:

"GPT-You", not GPT-X

Tl/dr:
- Closed APIs aren't defensible
- The durable moat is data
- The last mile generates the real value
2/ *Closed APIs aren't defensible*

- Recent examples like @StanfordCRFM Alpaca tinyurl.com/yc78bnct shows that cloning closed API-based FMs like ChatGPT can be done for a few $100 on top of small OSS base FMs (e.g. here, fine-tuning LLaMa 7B via exs from the ChatGPT API).
3/
- Since Alpaca dozens of others have cloned this cloning procedure (e.g. Dolly github.com/databrickslabs…)
- Will these withstand legal scrutiny? Enough potential legal issues w/ original FM training on web data may muddle things... but either way, the key point still stands...
4/ *The durable moat is data*
- Recent progress has shown that *data* is the secret sauce and real differentiator for FMs (e.g. ChatGPT is just GPT-3 fine-tuned with human feedback)
- Training on open web data can only get you so far for complex, enterprise-specific tasks.
5/
- Examples of FMs trained on proprietary, domain-specific data like BloombergGPT arxiv.org/abs/2303.17564 show the way forward: enterprises (and people!) using the durable moat of their own private data to build powerful, domain-specific FMs.
6/
- Recent OSS progress shows that FM model architectures are commoditizing (and standardizing)
- This means proprietary data will soon be the only durable moat.
- This data will be the edge that determines AI success.
- However: developing this data takes effort...
7/ *The last mile generates the real value*
- Getting real, complex AI use cases to production-level accuracy takes significant data labeling & development!
- See arxiv.org/abs/2302.10724, opensamizdat.com/posts/chatgpt_… - ChatGPT loses to specialized fine-tuned models 75%+ of the time!
8/
- Fine-tuning significantly out-performs ZSL/prompt approaches (e.g. see arxiv.org/pdf/2012.15723…)
- Even the OpenAI docs recommend min. 100 labeled examples/class for fine-tuning (for a 100-way classifier = 10K+ examples!), which empirical data shows is often a significant min
9/
- However, this data development is not just a chore- it's the source of a powerful flywheel.
- The more you fine-tune, the more powerful your FM becomes for your data & workloads, and the more value accrues!
- The "base" FM will matter less and less- as long as you own it.
10/ Tl/dr: the future will be "GPT-You", not GPT-X
- Closed APIs aren't defensible
- The durable moat is data
- The last mile generates the real value

Stay tuned for more on what we're building @SnorkelAI to support developing FMs on *your* data, for *your* tasks...

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Alex Ratner

Alex Ratner Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @ajratner

Dec 18, 2022
1/ 2023 AI prediction: the gap between generative and predictive AI will widen.

Despite product & business model innovation in generative AI, real-world ROI will remain concentrated around predictive AI- leading to frustrated expectations.

This gap will all come down to data...
2/ First, basic definitions:

- Generative (ie. LLMs / foundation models): Goal is to output a data point (e.g. an image)

- Predictive (or "discriminative"): Goal is to label a data point (e.g. predict whether an image contains offensive content).
3/ A natural response would be: isn't generating data fundamentally "harder" than just labeling it? And formally the answer is (roughly) yes.

However, the widening efficacy gap comes down to how each is used in the real world.
Read 8 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us on Twitter!

:(