I'd like to answer questions about our work with the DoW and our thinking over the past few days. Please AMA.
@natseckatrina who leads some of our national security work is going to jump in to answer some of your questions
@boazbaraktcs is also going to help out with answers!
Three general things from this AMA:
1. There is more open debate than I thought ther ewould be, at least in this part of Twitter, about whether we should prefer a democratically elected government or unelected private companies to have more power. I guess this is something people disagree on, but…I don’t. This seems like an important area for more discussion.
2. I think the is a question behind a lot of the questions but I haven’t seen quite articulated: What happens if the government tries to nationalize OpenAI or other AI efforts? I obviously don’t know; I have thought about it of course (it has seemed to me for a long time it might be be better if building AGI were a government project) but it doesn’t seem super likely on the current trajectory. That said, I do think a close partnership between governments and the companies building this technology is super important.
3. People take their safety (in the national security sense) more for granted than I realized, which I think is a good thing on balance but I don’t think shows enough respect to the tremendous work it takes for that to happen.
Also, I am on the whole very grateful for the level of reasonable and good-faith engagement here. It was not what I expected.
I have to go do something for awhile but can answer more questions tonight.
• • •
Missing some Tweet in this thread? You can try to
force a refresh
We have raised a $110 billion round of funding from Amazon, NVIDIA, and SoftBank.
We are grateful for the support from our partners, and have a lot of work to do to bring you the tools you deserve.
We are excited to partner with Amazon to bring a new generation of products to market, especially around new enterprise products like the stateful runtime environment. We are also very excited to make great use of Tranium.
We continue to have a great relationship with Microsoft. Our stateless API will remain exclusive to Azure, and we will build out much more capacity with them.
first, GPT-5 in an integrated model, meaning no more model switcher and it decides when it needs to think harder or not.
it is very smart, intuitive, and fast.
it is available to everyone, including the free tier, w/reasoning!
evals aren't the most important thing--the most important thing is how useful we think the model will be--but it does well on evals. for example, a new high on SWE-bench and many other metrics.
it is by far our most reliable and factual model ever.
rolling out today for free, plus, pro, and team users. next week to enterprise and edu.
making this available in the free tier is a big deal to us; PhD-level intelligence for everyone!
it can go use the internet, do complex research and reasoning, and give you back a report.
it is really good, and can do tasks that would take hours/days and cost hundreds of dollars.
people will post lots of great examples, but here is a fun one:
i am in japan right now and looking for an old NSX. i spent hours searching unsuccessfully for the perfect one. i was about to give up and deep research just...found it.
it is very compute-intensive and slow, but it's the first ai system that can do such a wide variety of complex, valuable tasks.
going live in our pro tier now, with 100 queries per month.
plus, team, and enterprise will come soon, and then free tier.
here is o1, a series of our most capable and aligned models yet:
o1 is still flawed, still limited, and it still seems more impressive on first use than it does after you spend more time with it. openai.com/index/learning…
but also, it is the beginning of a new paradigm: AI that can do general-purpose complex reasoning.
o1-preview and o1-mini are available today (ramping over some number of hours) in ChatGPT for plus and team users and our API for tier 5 users.
screenshot of eval results in the tweet above and more in the blog post, but worth especially noting:
a fine-tuned version of o1 scored at the 49th percentile in the IOI under competition conditions! and got gold with 10k submissions per problem.