Say hello to GPT-4o, our new flagship model which can reason across audio, vision, and text in real time:
Text and image input rolling out today in API and ChatGPT with voice and video in the coming weeks. openai.com/index/hello-gp…
Two GPT-4os interacting and singing
Realtime translation with GPT-4o
Lullabies and whispers with GPT-4o
Happy birthday with GPT-4o
@BeMyEyes with GPT-4o
Dad jokes with GPT-4o
Meeting AI with GPT-4o
Sarcasm with GPT-4o
Math problems with GPT-4o and @khanacademy
Point and learn Spanish with GPT-4o
Rock, Paper, Scissors with GPT-4o
Harmonizing with two GPT-4os
Interview prep with GPT-4o
Fast counting with GPT-4o
Dog meets GPT-4o
Live demo of GPT-4o realtime conversational speech
Live demo of GPT-4o voice variation
Live demo of GPT-4o vision
Live demo of coding assistance and desktop app
Live audience request for GPT-4o realtime translation
Live audience request for GPT-4o vision capabilities
All users will start to get access to GPT-4o today. In coming weeks we’ll begin rolling out the new voice and vision capabilities we demo’d today to ChatGPT Plus.
• • •
Missing some Tweet in this thread? You can try to
force a refresh
Our new GPT-4 Turbo is now available to paid ChatGPT users. We’ve improved capabilities in writing, math, logical reasoning, and coding.
Source: github.com/openai/simple-…
For example, when writing with ChatGPT, responses will be more direct, less verbose, and use more conversational language.
We continue to invest in making our models better and look forward to seeing what you do. If you haven’t tried it yet, GPT-4 Turbo is available in ChatGPT Plus, Team, Enterprise, and the API.
Sora can create videos of up to 60 seconds featuring highly detailed scenes, complex camera motion, and multiple characters with vibrant emotions.
Prompt: “Beautiful, snowy Tokyo city is bustling. The camera moves through the bustling city street, following several people enjoying the beautiful snowy weather and shopping at nearby stalls. Gorgeous sakura petals are flying through the wind along with snowflakes.”openai.com/sora
We’ll be taking several important safety steps ahead of making Sora available in OpenAI’s products.
We are working with red teamers — domain experts in areas like misinformation, hateful content, and bias — who are adversarially testing the model.
Prompt: “Several giant wooly mammoths approach treading through a snowy meadow, their long wooly fur lightly blows in the wind as they walk, snow covered trees and dramatic snow capped mountains in the distance, mid afternoon light with wispy clouds and a sun high in the distance creates a warm glow, the low camera view is stunning capturing the large furry mammal with beautiful photography, depth of field.”
We are systemizing our safety thinking with our Preparedness Framework, a living document (currently in beta) which details the technical and operational investments we are adopting to guide the safety of our frontier model development. openai.com/safety/prepare…
Our Preparedness Team will drive technical work, pushing the limits of our cutting edge models to run evaluations and closely monitor risks, including during training runs. Results will be synthesized in scorecards that track model risk.
Our new safety baselines and governance process will turn these technical findings into safety decisions for model development and deployment. This involves establishing a cross-functional Safety Advisory Group to make safety recommendations.
In the future, humans will need to supervise AI systems much smarter than them.
We study an analogy: small models supervising large models.
Read the Superalignment team's first paper showing progress on a new approach, weak-to-strong generalization: openai.com/research/weak-…
Large pretrained models have excellent raw capabilities—but can we elicit these fully with only weak supervision?
GPT-4 supervised by ~GPT-2 recovers performance close to GPT-3.5 supervised by humans—generalizing to solve even hard problems where the weak supervisor failed!
Naive weak supervision isn't enough—current techniques, like RLHF, won't be sufficient for future superhuman models.
But we also show that it's feasible to drastically improve weak-to-strong generalization—making iterative empirical progress on a core challenge of superalignment
ChatGPT can now browse the internet to provide you with current and authoritative information, complete with direct links to sources. It is no longer limited to data before September 2021.
Since the original launch of browsing in May, we received useful feedback. Updates include following robots.txt and identifying user agents so sites can control how ChatGPT interacts with them.
Browsing is particularly useful for tasks that require up-to-date information, such as helping you with technical research, trying to choose a bike, or planning a vacation.
ChatGPT can now see, hear, and speak. Rolling out over next two weeks, Plus users will be able to have voice conversations with ChatGPT (iOS & Android) and to include images in conversations (all platforms). openai.com/blog/chatgpt-c…
Use your voice to engage in a back-and-forth conversation with ChatGPT. Speak with it on the go, request a bedtime story, or settle a dinner table debate.
Sound on 🔊
Show ChatGPT one or more images. Troubleshoot why your grill won’t start, explore the contents of your fridge to plan a meal, or analyze a complex graph for work-related data.