It's OpenAI's new AI agent that autonomously takes action across the web on your behalf.
The 9 most impressive use cases I’ve tried (videos sped up):
1. Ordering dinner ingredients based on a picture and a recipe
2. Planning a weekend trip based on hidden gems off Reddit, my budget and interests
Notice how at 0:06, ChatGPT Operator was blocked from Reddit but then decided to just do a Bing search with "Reddit" at the end
Very impressive decision-making
3. Crypto investment research based on tokens that are actually worth looking into
Notice how ChatGPT Operator got hit with a "Are you human" CAPTCHA, then pinged me to take control to confirm
Wild workaround
4. Booking a one-way flight from Zurich to Vienna using the Booking integration
This one required a bit of back and forth, with ChatGPT Operator pinging me and asking for my flight preference and having me take control of entering payment details
5. Scheduling an appointment with my barber after looking at my Google Calendar schedule/availability
Note that in this demo, ChatGPT Operator pinged me that I needed to sign in to Google to check my calendar
I tried a second time, and my login was saved session-to-session
6. Researching a good birthday gift for my mom based on what she likes
Similar to the Reddit block, ChatGPT Operator couldn't access NYTimes, so it pivoted and found another site.
Really neat.
Also cool to see it compare and find the best price across the web for me, too
7. Booking a one-time house cleaner for my home through the Thumbtack integration based on my budget
ChatGPT Operator came back to me with four highly rated options within my price range
8. Finding the best/cheapest health insurance coverage in Switzerland
This was interesting since most prices are not publicly available and are gated behind a meeting
ChatGPT Operator did what it could, and presented me with a good blog for me to read further
9. Finding a top-rated dog walker in Vancouver BC
This is no easy task, so I wanted to test how well ChatGPT Operator could handle it
To my surprise, I got 3 really solid options at the end
Overall, I was very impressed by the research preview of Operator.
I loved that it can do tasks for me as I do other work, and simply ping me when it needs me to "take over"
I also really enjoyed the saved tasks tab, and adding Custom Instructions for specific websites.
But it's important to note that Operator is still a research preview and is improving.
I found that:
-Quite a few sites were blocked after they detected the AI
-There's a limited set of partner integrations
-It's true purpose is to take actions across the web (more below)
Operator *operates* within ChatGPT, but it's a completely different tool.
Its output lengths are small, and its true purpose is to take actions across the web (typing, clicking, scrolling).
Meaning it's not like ChatGPT, which can produce essays and write long code
With every new tool, comes a new way of using it optimally.
E.g. with GPT-4, CoT prompting produced the best results, but prompting o1 best is completely different.
The exact same thing is happening here with Operator, and I'm 100% just scratching the surface with these tests.
The future of tech work is here. And personally, I'm incredibly excited about it.
Agents can do the boring work, so I can spend more time doing what I love.
I'll be publicly sharing all the ways I automate my work with agents, so follow me @rowancheung for more.
Lastly, big thanks to @OpenAI for granting me early access. I had a ton of fun early testing Operator.
If you want to support my work, like/retweet the first tweet of this thread to share with friends:
That's a wrap for day 2 of the world's largest consumer tech event, CES 2025.
The top 10 tech and gadget reveals from day 2:
1. A stretchable Micro LED display that turns 2D into 3D by Samsung
2. A multitasking household robot that does everything from vacuuming, organization, air purification, monitoring pets, and even delivering you food while you sit on the couch by SwitchBot
3. An immerse location-based entertainment concept that allows players to use flashlights and guns in an LED environment by Sony
Google just released Veo 2, a new state-of-the-art AI video model.
In testing, Veo beat OpenAI Sora in BOTH quality and prompt adherence.
The video compilation below is 100% created by AI (more details in thread):
Veo can generate 8 second videos in up to 4K resolution (720p at launch).
The model also features:
— Better understanding of physics for more natural movement, lighting, etc.
— Enhanced clarity and sharpness of outputs
— Reduced hallucinated objects and details
The model also excels at a variety of cinematic styles, with better camera control for more creative storytelling.
I've been an early tester + had in-person demos for most of Google’s AI projects announced today.
I found several practical use cases that will benefit everyday people.
12 use cases of Project Astra/Deep Research Agents/Project Mariner (beyond the hype):
Project Astra: Google's AI agent that can 'see the world' using your phone camera
Use cases that stood out to me:
> Summarizing a book page in seconds and chatting with it for follow-ups on complex topics (professor-in-your-pocket)
> Identifying a rash: just seasonal hives or something more serious?
> Real-time translation of languages, sign language, and books (worked great for Japanese writing → English summary).
> Locating landscapes in photos and estimating their distance using the Google Maps integration.
> Remembering cookbook recipes and recommending wine pairings based on the recipe and budget.
> Summarizing thousands of Amazon/Airbnb reviews in seconds using mobile screen sharing, with highlights of any negative feedback.
Deep Research Agent: Google’s new research assistant that create's full reports on any topic and links back to the relevant sources.
Use cases that stood out to me:
> Coming up with interview questions based on what people are curious about across the internet.
> Conducting market research on stocks (e.g., "Why did Google stock go up today?").
> Creating a full Christmas gift plan for my mom (based on current trends and her preferences highlighted in the prompt)
> Creating an analysis and report of my health/fitness and how I can improve based on my Whoop data.
AI NEWS: Meta just unexpectedly dropped Llama 3.3—a 70B model that's ~25x cheaper than GPT-4o.
Plus, Google released a new Gemini model, OpenAI reinforcement finetuning, xAI's Grok is available for free, Copilot Vision, ElevenLabs GenFM, and more.
Here's what you need to know:
Meta just dropped Llama 3.3 — a 70B open model that offers similar performance to Llama 3.1 405B, but significantly faster and cheaper.
It's also ~25x cheaper than GPT-4o.
Text only for now, and available to download at llama .com/llama-downloads
Google launched a new model, gemini-exp-1206 on Gemini's 1 year birthday (today!)
It tops the Chatbot Arena rankings in ALL domains.
It also looks like it's available for everyone in AI Studio for free!