Ethan Mollick Profile picture
Professor @Wharton studying AI, innovation & startups. Democratizing education using tech Book: https://t.co/CSmipbJ2jV Substack: https://t.co/UIBhxu4bgq
56 subscribers
Aug 10 4 tweets 4 min read
“Claude Opus 4.1, De-Carcinize the Great Gatsby” (That was the only prompt)

Pretty clever, actually. Image
Image
Image
Image
# The Great Gatsby: De-Carcinized

## Chapter 1: Direct Approach

Nick Carraway moved to West Egg and immediately walked straight to his neighbor Jay Gatsby’s house.

“Hello,” Nick said. “I’m your new neighbor. I notice you throw parties.”

“Yes,” Gatsby replied, opening the door fully instead of peering through a crack. “I throw them hoping Daisy Buchanan will attend. She lives across the bay. I love her.”

“She’s my cousin,” Nick said. “She’s married to Tom.”

“I know. Would you arrange a meeting?”

“Sure.”

## Chapter 2: Honest Communications

Tom Buchanan stood in his living room, addressing everyone directly. “I’m having an affair with Myrtle Wilson. Daisy, you probably know this. Myrtle, you should know I’ll never leave Daisy for you.”

Myrtle nodded. “I understand but choose to continue anyway.”

Daisy turned to face them. “I’m unhappy but too comfortable to change anything.”

## Chapter 3: The Party Without Pretense

Gatsby’s next party featured a sign at the entrance: “These Parties Are Specifically to Attract Daisy Buchanan’s Attention.”

Guests arrived and immediately stated their intentions:
“I’m here for free champagne.”
“I’m networking for business.”
“I enjoy judging rich people.”

Gatsby stood on a platform with a megaphone: “Thank you all for coming to my Daisy-attraction event. Please enjoy yourselves while I wait hopefully by the window.”

## Chapter 4: Swimming Forward

Gatsby drove Nick directly to the city—no detours through the Valley of Ashes, no sideways routes through East Egg.

“I should tell you my backstory without any embellishment,” Gatsby said. “I was born poor as James Gatz. I made my fortune through questionable means with Meyer Wolfsheim. Everything I’ve built is to win Daisy back.”

“That seems unhealthy,” Nick observed.

“It absolutely is,” Gatsby agreed.

## Chapter 5: The Reunion

At Nick’s house, Gatsby walked straight to Daisy when she arrived.

“I’ve loved you for five years,” he said. “I’ve built my entire life around the hope of resuming our relationship.”

“That’s concerning,” Daisy replied. “But I’m flattered and bored in my marriage, so let’s proceed.”

They looked at each other directly, no dancing around the truth.

“This will end badly,” they said in unison.

“Want to see my house?” Gatsby asked.

“Yes, though material possessions won’t fix our fundamental incompatibility.”

## Chapter 6: Clear Waters

Tom attended Gatsby’s party and immediately announced: “I distrust you and will investigate your background.”

“Fair enough,” Gatsby replied. “I am indeed hiding things.”

Daisy pulled Gatsby aside. “These parties make me uncomfortable. You’re not the same person I loved five years ago.”

“You’re right,” Gatsby admitted. “But I’m unable to accept that.”

## Chapter 7: Direct Confrontation

In the hotel room, everyone laid their cards on the table—literally, they brought actual cards with their feelings written on them.

Tom’s card: “Gatsby is a criminal.”
Gatsby’s card: “Daisy loves me, not you.”
Daisy’s card: “I love security and comfort more than either of you.”
Nick’s card: “I’m uncomfortable and want to leave.”
Jordan’s card: “This is entertaining.”

“Well,” Tom said, “since Daisy has clarified her priorities, and I offer more security, I win.”

Gatsby nodded sadly. “Your logic is sound.”

## Chapter 8: Forward Motion

Driving home, Daisy didn’t swerve or panic. She drove straight and responsibly, arriving home without incident.

“That was anticlimactic,” Gatsby said.

Later, Wilson approached Gatsby at his pool.

“Did you have an affair with my wife?” Wilson asked directly.

“No, that was Tom Buchanan. He just told you it was me to deflect blame.”

“Oh. My mistake.” Wilson left to find Tom.
Aug 2 4 tweets 2 min read
This is, in fact, what early acceleration of science looks like with jagged AI.

Writing abstracts takes a lot of time, many scientists are bad writers and clearly written abstracts are important. As long as the authors check over the results, this is a good use case for AI. I wrote about some of the early impact of AI on science last year, including for writing. oneusefulthing.org/p/four-singula…
Jul 27 5 tweets 2 min read
Kinda amazing: the mystery model "summit" with the prompt "create something I can paste into p5js that will startle me with its cleverness in creating something that invokes the control panel of a starship in the distant future" & "make it better"

2,351 lines of code. First time Code: editor.p5js.org/emollick/sketc…

Play it: editor.p5js.org/emollick/full/…
Jul 20 4 tweets 1 min read
Don't leave AI to the STEM folks.

They are often far worse at getting AI to do stuff than those with a liberal arts or social science bent. LLMs are built from the vast corpus human expression, and knowing the history & obscure corners of human works lets you do far more with AI These are systems that respond to human writing and (often) techniques that apply to human psychology.

Everyone now has a machine that makes words, images, video, sound where the limit is often your own ability to imagine something new (or invoke old ideas others do not know).
Jul 6 4 tweets 1 min read
X (and other social media sites) make our 1990s optimism about the Information Age seem silly.

Even with all of the world's information a click away (& a free AI that can help explain that information in a personalized way), half-mangled anecdotes with no source win every time. It really is not what most people who was working on building the early web in the late 1990s were expecting. Universal access to information was going to transform everything, creating widespread learning and bridging divides.

It really is shocking how much that didn't happen.
Jun 5 4 tweets 1 min read
So, OpenAI Deep Research can connect directly to Dropbox, Sharepoint, etc.

Early experiments only, but it feels like what every "talk to our documents" RAG system has been aiming for, but with o3 smarts and easy use. I haven't done robust testing yet, but very impressive so far. I think it is going to be a shock to the market, since "talk to our documents" is one of the most popular implementations of AI in large organizations, and this version seems to work quite well and costs very little.
May 20 4 tweets 2 min read
Very big impact: The final version of a randomized, controlled World Bank study finds using a GPT-4 tutor with teacher guidance in a six week after school progam in Nigeria had "more than twice the effect of some of the most effective interventions in education" at very low costs Image
Image
Image
Image
Paper: documents.worldbank.org/en/publication…
Apr 24 5 tweets 2 min read
Microsoft keeps launching Copilot tools that seem interesting but which I can't ever seem to locate. Can't find them in my institution's enterprise account, nor my personal account, nor the many Copilot apps or copilots to apps or Agents for copilots

Each has their own UIs. 🤷‍♂️ Image For a while in 2023, Microsoft, with its GPT-4-powered Bing, was the absolute leader in making LLMs accessible and easy to use.
Mar 24 4 tweets 2 min read
As someone who has spent a lot of time thinking and building in AI education, and sees huge potential, I have been shown this headline a lot

I am sure Alpha School is doing interesting things, but there is no deployed AI tutor yet that drives up test scores like this implies. Image I am not doubting their test results, but I would want to learn more about the role AI is playing, and what they mean by AI tutor, before attributing their success to AI as opposed to the other dials they are turning.

AI tutoring is still being understood
Mar 15 4 tweets 1 min read
I suspect that a lot of "AI training" in companies and schools has become obsolete in the last few months

As models get larger, the prompting tricks that used to be useful are no longer good; reasoners don't play well with Chain-of-Thought; hallucination rates have dropped, etc. I think caution is warranted when teaching prompting approaches for individual use or if training is trying to define clear lines about tasks where AI is bad/good. Those areas are changing very rapidly.
Mar 8 5 tweets 2 min read
“GPT-4.5, Give me a secret history ala Borges. Tie together the steel at Scapa Flow, the return of Napoleon from exile, betamax versus VHS, and the fact that Kafka wanted his manuscripts burned. There should be deep meanings and connections”

“Make it better” a few times… Image It should have integrated the scuttling of the High Seas Fleet better but it knocked the Betamax thing out of the park
Mar 4 4 tweets 2 min read
🚨Our Generative AI Lab at Wharton is releasing its first Prompt Engineering Report, empirically testing prompting approaches. This time we find:
1) Prompting “tricks” like saying “please” do not help consistently or predictably
2) How you measure against benchmarks matters a lot Image
Image
Using social science methodologies for measuring prompting results helped give us some useful insights, I think. Here’s the report, the first of hopefully many to come. papers.ssrn.com/sol3/papers.cf…
Feb 25 4 tweets 1 min read
The lack of benchmarks for writing, telling stories, persuasion, creativity, emotional intelligence, perceived empathy, and doing office work are...

(1) holding back AI advances, (2) hiding big differences between models & (3) obscuring how good these models are for real work If you want to influence the future, now is the time to release a really good benchmark.
Feb 16 4 tweets 1 min read
The significance of Grok 3, outside of X drama, is that it is the first full model release that we definitely know is at least an order of magnitude larger than GPT-4 class models in training compute, so it will help us understand whether 1st scaling law (pre-training) holds up. It is possible that Gemini 2.0 Pro is a RonnaFLOP* model, but we are only seeing the Pro version, not the full ultra.

* AI trained on 10^27 FLOPs of compute, an order of magnitude more than then GPT-4 level (I have been calling them Gen3 models because it is easier)
Feb 11 5 tweets 2 min read
There is a lot of important stuff in this new paper by Anthropic that shows how people are actually using Claude.
1) The tasks that people are asking AI to do are some of the highest-value (& often intellectually challenging)
2) Adoption is uneven, but many fields already high Image
Image
Image
Paper: assets.anthropic.com/m/2e23255f1e84…
Feb 10 4 tweets 2 min read
Thoughts on this post:
1) It echoes what we have been hearing from multiple labs about the confidence of scaling up to AGI quickly
2) There is no clear vision of what that world looks like
3) The labs are placing the burden on policymakers to decide what to do with what they make Image
Image
I wish more AI lab leaders would spell out a vision for the world, one that is clear about what they think life will actually be like for humans living in a world of AGI

Faster science & productivity, good - but what is the experience of a day in the life in the world they want?
Jan 21 4 tweets 1 min read
$500B committed towards AGI, still no articulated vision of what a world with AGI looks like for most people. Even the huge essay by the CEO of Anthropic doesn't paint a vivid picture

For those convinced they are making AGI soon - what does daily life look like 5-10 years later? Lets leave aside the risk of catastrophe for now.

Assume we get an aligned AGI that supercharges science and we have a healthier, more advanced, safer world. What does that actually mean for most people, what does their life look like in the future? (Hint: UBI is not an answer)
Jan 15 6 tweets 2 min read
New randomized, controlled trial of students using GPT-4 as a tutor in Nigeria. 6 weeks of after-school AI tutoring = 2 years of typical learning gains, outperforming 80% of other educational interventions.

And it helped all students, especially girls who were initially behind Image
Image
No working paper yet, but the results and experiment are written up here. They used Microsoft Copilot and teachers provided guidance and initial prompts: blogs.worldbank.org/en/education/F…
Jan 13 7 tweets 2 min read
Most of the discussions around AI and energy use refer to an older 2020 estimate of GPT-3 energy consumption, but a more recent paper directly measures energy use of Llama 65B as 3-4 joules per decoded token.

So an hour of streaming Netflix is equivalent to 70-90,000 65B tokens. Image
Image
Paper: arxiv.org/pdf/2310.03003
Dec 17, 2024 10 tweets 4 min read
Veo 2 prompt: "a distant shot zooms in to reveal a knight wearing a golden helmet, he begins to charge on his zebra, lowering his lance, charging towards a clockwork octopus" (this is one of the initial 4 videos it made) "an woman with short black hair assembles an impossibly complicated device, close up on her face, she is sweating"

The consistency of small details is really impressive, the fact that the shaft of the screw turns at the same speed and direction, hair and sweat, tattoos...
Dec 8, 2024 5 tweets 2 min read
I thought puzzle design beyond current AI, so this is the first time I have seen a model actually create a solvable and interesting puzzle.

"o1, create a d&d puzzle that is non-trivial, but solvable. it should not involve elements or statues, be creative."

(Solution in thread) Image Here are the hints it gave and the solution.

(There is one unnecessary step in the puzzle, by the way, but a zero shot coherent word puzzle is pretty neat) Image
Image