I was amazed that readers of the last thread pledged >$500, without us even asking. I am considering accepting and using that money on an editor, so I don't make mistakes like this:
This move puts OpenAI's price at a fraction of their most direct competitors' prices, just as competitors are ramping up.
Driving costs so low so quickly will make things very difficult for non-hyperscalers. How are they ever supposed to recoup all that fixed training cost??
To compare foundation model prices across providers, I created this document. Feel free to make a copy and plug in your own usage patterns.
And indeed, I've noticed in my initial testing that it cannot follow some rather complex/intricate instructions that text-davinci-003 can follow successfully. Fwiw, Claude from @AnthropicAI is the only other model I've seen that can reliably follow these instructions.
Pro tip: when evaluating new models, it helps to have a set of go-to prompts that you know well. I can get a pretty good sense for relative strengths and weaknesses in a few interactions this way.
How did this 90% price drop happen?
Simple really – "Through a series of system-wide optimizations, we’ve achieved 90% cost reduction for ChatGPT since December; we’re now passing through those savings to API users."
Since December? Surely the work was in progress earlier.
Though, I could imagine that all the inference data they've collected via ChatGPT might be useful for powering a pruning procedures, which could bring parameter size down significantly. So maybe?
Related: OpenAI apparently has all the data it needs, thank you very much.
Perhaps having so much data that you no longer need to use retain or use customer data in model training is the real moat?
For the global poor, this is a huge win. Chat interactions that consist of multiple rounds of multiple paragraphs will now cost just a fraction of a cent. For a nickel / day, people can have all the AI chat they need.
The educational possibilities alone are incredible, and indeed OpenAI highlights "global learning platform" Quizlet, already with 60 million students, and now launching "a fully-adaptive AI tutor" today. For ~$1 / day, you can sponsor a full classroom!
If you're inspired by the idea of Universal Basic Intelligence and think you can run a program that delivers more value per dollar than @GiveDirectly, I'd love to hear about it!
Meanwhile, in the rich world, AI was already cheap and will now be super-abundant
But one thing that is missing at this price level: fine-tuning! For the turbo version, you'll have to make do with old-fashioned prompt engineering. The default voice of ChatGPT will be everywhere
Every product is going to be talking to us, generating multiple versions, proactively offering variations, suggesting next actions. Executed poorly it will be overwhelming, executed well it will feel effortless.
Have to shout out my own company @waymark as a great example here. We had the vision for a "Watch first" video creation experience years ago, and we now combine a bunch of different AIs to make it possible.
And while it wasn't the biggest news OpenAI made this week, they did revamp their website to highlight customer stories, and we were genuinely honored to be featured!
AI auto-debugging is amazing, but often looks even clumsier than human debugging. Lot of calls, but … that matters a lot less now
It follows that AI bot traffic is going to absolutely explode. I have a vision of @LangChainAI lobsters, just smart enough to slowly stumble around the web until they accomplish their goals. Image credit to @playground_ai
upshot: OpenAI revenue might not drop as much as you'd think
Anecdote from @CogRev_Podcast: @Suhail told us that when @playgroundAI made image generation twice as fast, they immediately usage jump 2X. Something like that might happen here.
Embedding databases like @trychroma win too. Cheaper tokens make it worthwhile to embed everything and figure it out context management later / on the fly.
To give you a sense for the volume of usage that OpenAI is expecting… they also announced dedicated instances that "can make economic sense for developers running beyond ~450M tokens per day."
That's a lot!
As I said the other day, longer context windows and especially "robust fine-tuning" seem to be OpenAI's core price discrimination strategy, and I think it's very likely to work.
While essay writers in India pay a tenth of a penny for homework help, corporate customers will pay 1-2 orders of magnitude more for the ability to really dial in the capabilities they want.
And I'll bet some will pay 3 orders of magnitude more, simply because their workload doesn't take full advantage of their dedicated compute. Few companies run evenly 24/7.
What might OpenAI do with that compute when it's not in use? Safe to say they won't be mining bitcoin.
For open source? This takes some air out of open source model projects – when a quality hosted version that runs fast and reliability is so cheap, who needs the comparative hassle of open source? – but open source will remain strong. It's driven by more than money.
And importantly, the open source community can do things OpenAI can't
Best AI podcast of 2023 so far is "No Priors" interview with @EMostaque – his vision for a cambrian explosion of small models, each customized to its niche, is beautiful and inspiring
OpenAI's leaked Foundry pricing says a lot – if you know how to read it – about GPT4, The Great Implementation, a move from Generative to Productive AI, OpenAI's safety & growth strategies, and the future of work.
Another AI-obsessive megathread on what to expect in 2023 🧵
Disclaimer: I'm an OpenAI customer, but this analysis is based purely on public info
I asked our business contact if they could talk about Foundry, and got a polite "no comment"
As this is outside analysis, I'm sure I'll get some details wrong, and trust you'll let me know 🙏
If you prefer to read this on a substack, sign up for mine right here.
In the future I might publish something there and not here, and you wouldn't want to miss that. :)
Introducing "text-2-commercial" – the unique text-2-video experience we're building @Waymark
Watch our CEO @aperskystern make an original, creative, compelling marketing video for a small business in <1 minute.
I'll explain how it works in the thread
With Waymark, users only have to do two things to get *watchable* videos:
(1) identify their business by name and (optional) location, and
(2) tell us, in their own words, about the video they want to create – or just let the AI come up with something :)
Waymark searches the web for business content, then uses AI to build a business profile – including "about" content, logo, images, & color palette – all in <30 seconds
No disrespect to Ezra here – he's not an AI expert, and he asked some great questions. And I think Gary deserves credit for flagging a number of important issues & likely problems that will come with wide-scale AI deployment – I agree that society is not ready for what's coming!
Nevertheless, there are so many inaccuracies in this interview that I expect it will confuse the audience more than it informs, so while trying to be earnest and polite, I will endeavor to correct the record.