sridhar Profile picture
Dec 21, 2022 16 tweets 7 min read Read on X
1/ #ChatGPT is closing out 2022 with a bang, but what’s next? 💥

@OpenAI’s #GPT4 is set to be the first big #AI thing in 2023.

So here are some bold, optimistic, yet sensible predictions from me, @vivek7ue and @rajhans_samdani ... 👀
2/ Biggest model size for GPT-4 will be 1T parameters. Up 6x.

Not 100T parameters like some AI hypers are claiming ().2/ Biggest model size for GPT-4 will be 1T parameters. Up 6x Image
3/ The reason is simple: instruction fine tuning achieves same quality with 100x smaller models.
arxiv.org/pdf/2203.02155… Image
4/ As such, the pre-trained GPT-4 model will appear to be a modest improvement over Chincilla, PALM and U-PALM on HELM and BigBench.

The raw stats on GPT-4 will look underwhelming at first glance and incremental relative to GPT-3.
5/ The hidden secret of #LLMs? How much training data you have matters as much as model size.

GPT-4 will use 10T tokens. Up 33x, and putting them on the Chinchilla scaling curve.
6/ Biggest user facing change? Longer context windows.

We expect 16384 tokens (⬆️ from 4096).
7/ Biggest pre-training modeling change? A loss function that looks like UL2 (arxiv.org/pdf/2205.05131…).
8/ Put together, at least 800x more compute for the pre-trained model.

And that will mean it’s better. 🙌 🙌
9/ Lots of the pre-training secret sauce will be in what goes in. 🤫

We expect:
➡️ A lot more dialog data (from @Twitter, @Reddit and elsewhere)
➡️ Proprietary signals from @bing's index and maybe even Bing clicks
10/ The instruction-following models will continue to be state of the art relative to everyone else (see the HELM comparisons at arxiv.org/abs/2211.09110)
11/ They will:

👉 Incorporate RLHF/PPO (like GPT3.5)
👉 Use proprietary prompt-following training data from the OpenAI playground (that other research groups can't access)
12/ PPO preference training will re-use some of the tricks @AnthropicAI is using to be more helpful and harmless in their constitutional training paradigm
13/ InstructGPT-3 used about 20B tokens during PPO = 6% of total GPT-3 compute.

Since instruction fine tuning is a lot more compute optimal, we expect a lot more compute to be spent in the supervised fine-tuning and PPO phases.
14/ GPT-4 will be fine-tuned on all the feedback data from ChatGPT, and that will be the key to a significant improvement.

With a million prompts a day from ChatGPT, we expect compute used in PPO to go up a lot in GPT-4.
15/ And finally, like with ChatGPT, OpenAI will NOT publish details about GPT4 as a paper, leaving the world guessing what's in there.

This will start a trend where all the big foundation model companies will stop publishing details of their models.

OpenAI will be Open no more.
16/ This will leave a BIG opportunity for open model efforts from the likes of @AiEleuther, @huggingface, Big Science’s BLOOM, @togethercompute, and @carperai to step up their game.

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with sridhar

sridhar Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @RamaswmySridhar

Feb 8, 2023
The web needs real, authentic content.

You’ve honed your craft as a publisher, producing high quality material.

And a lot of your hard-earned traffic comes from page ranking well in Google search.

But things are about to get a whole lot worse for you… 🧵
Even in the past, a lot of your traffic from Google was drive-by.

Maybe people stayed for exactly one view.

You’ve looked for ways to increase stickiness.

But you haven’t really built search & recommendation engines on your site. They’re hard to custom build for your content.
Here come the LLMs and AI answer bots aggregating and summarizing YOUR content in search results.

You’ve seen how ChatGPT gives fluent answers and what that previews.

With answer-bots, you aren’t getting attribution nor traffic.

Does this sound like “fair use” to you?
Read 13 tweets
Feb 7, 2023
Lots of AI-focused investments being announced w/ BIG $$$ at play.

Many rounds are being led by cloud vendors like Azure and GCP.

What's going on?

These clouds have 50% utilization on their AI compute.

AI is a HUGE opportunity for cloud providers

What @vivek7ue + I see🧵
LLM startups are aiming to secure access to guaranteed compute at effective prices.

Why?

There’s a big squeeze on (a) cost of compute and (b) availability.

Say you’re an LLM startup. Having compute is a competitive advantage over your peers.
Compute power to run an LLM can be high! (esp if you need to play the "AGI big boy" game...)

And availability for compute access is a first order issue.

Unless you are willing to commit large amounts of $$, you are NOT prioritized for availability for compute.
Read 10 tweets
Feb 5, 2023
⚠️Publishers, beware! ⚠️

Your content is going to get sucked into AI bots pulling YOUR work into their large language models.

Your referral traffic will go down!

If Google clones @Neeva AI (which we expect), then publishers are at risk

@vivek7ue & I have solutions 🧵
As search engines become answer engines, referral traffic will drop!

It’s happened before: Google featured snippets caused this on 10-20% of queries in the past.

This is dangerous for pubs unless they take control of their relationship w/their search overlords. 3 ways to this:
1️⃣ Don’t get crawled – Tough to do. Unless publishers band together, unilateral disarmament is hard for a single publisher to achieve against a monopoly.
2️⃣ Ask for payment for LLM inclusion – again, tough against a monopoly. (@Neeva is committed to pub revenue share)

Or…
Read 14 tweets
Feb 4, 2023
@VeronicaSaron from my team said. "Dude, I feel old. How to I talk to genZ?!"

(NOOOOO...she did not say that. I am, um, paraphrasing...)

Me: "Can totally do a tool for you. I hear chatGPT is good at programming."
easy enough, right?! does not work. 403 blah blah.
back to @Neeva top result is this: medium.com/codingthesmart…
Read 4 tweets
Feb 4, 2023
Want to know how our team at @neeva seamlessly combined LLMs with our search stack?

And how we reduced time to first byte from 8 seconds to less than 1.5 seconds?

Read on below…🧵
To get answers on NeevaAI, there are 4 phases:
1/ Retrieve and rank pages from our 4 billion index (soon to be 6b)

2/ Compute extractive and abstractive summaries for the top pages

3/ Send back a search result page w/a skeleton NeevaAI answer

4/ Stream the AI answer back
Step 1 can be done in about 500ms but step 2 took >7 seconds.

That is an unacceptable delay for any search engine!

Our team had to get into model brain surgery!

While deep learning has traditionally favored model symmetry we find this can be troublesome for efficient inference
Read 11 tweets
Jan 26, 2023
1/ Ten blue links headed to a museum near you!!

@Neeva is applying cutting edge AI to definitely change up the search experience!

Not only are we providing real time, cited AI, because we are making the whole search experience a breeze!

🧵
2/ First, a quick backstory on #NeevaAI.

We were the first company to launch cited summaries backed by actual data. With our real-time AI search and web crawling capabilities, we provide authoritative answers that you can trust.

No more hallucinations!
3/ And now, we're taking our experience to the next level with the launch of Citation cards.

Simply click on a sentence in our summaries and see a card with additional details, allowing you to fact-check AI in a user-first experience.
Read 10 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(