).2/ Biggest model size for GPT-4 will be 1T parameters. Up 6x
3/ The reason is simple: instruction fine tuning achieves same quality with 100x smaller models. arxiv.org/pdf/2203.02155…
4/ As such, the pre-trained GPT-4 model will appear to be a modest improvement over Chincilla, PALM and U-PALM on HELM and BigBench.
The raw stats on GPT-4 will look underwhelming at first glance and incremental relative to GPT-3.
5/ The hidden secret of #LLMs? How much training data you have matters as much as model size.
GPT-4 will use 10T tokens. Up 33x, and putting them on the Chinchilla scaling curve.
6/ Biggest user facing change? Longer context windows.
We expect 16384 tokens (⬆️ from 4096).
7/ Biggest pre-training modeling change? A loss function that looks like UL2 (arxiv.org/pdf/2205.05131…).
8/ Put together, at least 800x more compute for the pre-trained model.
And that will mean it’s better. 🙌 🙌
9/ Lots of the pre-training secret sauce will be in what goes in. 🤫
We expect:
➡️ A lot more dialog data (from @Twitter, @Reddit and elsewhere)
➡️ Proprietary signals from @bing's index and maybe even Bing clicks
10/ The instruction-following models will continue to be state of the art relative to everyone else (see the HELM comparisons at arxiv.org/abs/2211.09110)
11/ They will:
👉 Incorporate RLHF/PPO (like GPT3.5)
👉 Use proprietary prompt-following training data from the OpenAI playground (that other research groups can't access)
12/ PPO preference training will re-use some of the tricks @AnthropicAI is using to be more helpful and harmless in their constitutional training paradigm
As search engines become answer engines, referral traffic will drop!
It’s happened before: Google featured snippets caused this on 10-20% of queries in the past.
This is dangerous for pubs unless they take control of their relationship w/their search overlords. 3 ways to this:
1️⃣ Don’t get crawled – Tough to do. Unless publishers band together, unilateral disarmament is hard for a single publisher to achieve against a monopoly.
2️⃣ Ask for payment for LLM inclusion – again, tough against a monopoly. (@Neeva is committed to pub revenue share)
We were the first company to launch cited summaries backed by actual data. With our real-time AI search and web crawling capabilities, we provide authoritative answers that you can trust.
No more hallucinations!
3/ And now, we're taking our experience to the next level with the launch of Citation cards.
Simply click on a sentence in our summaries and see a card with additional details, allowing you to fact-check AI in a user-first experience.