Ruben Hassid Profile picture
Aug 25 14 tweets 5 min read Read on X
For the first time, Google has measured how much energy AI really uses in production.

Spoiler: the gap vs. all previous estimates is huge... 🧵 Image
Despite AI transforming healthcare, education, and research, we've been flying blind on its environmental footprint.

Every estimate was based on lab benchmarks, not real-world production systems serving billions of users.

Google decided to measure what actually happens. Image
The results from measuring Gemini in production:

• 0.24 watt-hours per text prompt
• Equivalent to watching TV for 9 seconds
• 5 drops of water consumed
• 0.03 grams of CO2 emissions

Substantially lower than public estimates. Image
Even if you add the water used in nearby power plants to generate electricity, the total rises only to 2.7 mL per prompt.

That’s 0.0001% of the average American’s daily water use.

Substantially lower than public estimates. Image
Over just 12 months, while AI quality improved dramatically, Gemini's environmental impact plummeted:

• Energy per prompt: down 33x
• Carbon footprint: down 44x

This is efficiency innovation at scale. Image
Why were all previous estimates wrong?

They measured isolated chips running perfect benchmarks.

Google measured the full reality: idle machines for reliability, cooling systems, power distribution, CPU overhead... everything needed for global AI service. Image
The measurement gap is enormous:

• Lab benchmark approach: 0.10 Wh per prompt
• Full production reality: 0.24 Wh per prompt

That 2.4x difference reveals the infrastructure complexity everyone else ignored. Image
→ The efficiency breakthroughs come from Google's full-stack AI approach.

Transformer architecture (invented at Google) delivers 10-100x efficiency over previous language models.

Mixture-of-Experts activates only needed model parts, cutting computation by 10-100x. Image
→ Google builds custom chips specifically for AI instead of using regular computer chips.

Their latest AI chips are 30x more energy-efficient than their first generation.

Like designing a race car for Formula 1 instead of using a family sedan. Image
→ Smart software tricks that save massive energy

Small AI models write rough drafts, big models just check the work.

Models automatically move between chips based on demand, so nothing sits idle. Image
→ Infrastructure efficiency that matters at billions of prompts

Google's data centers operate at 1.09 PUE, only 9% energy overhead beyond actual computing.

Advanced cooling balances energy, water, and emissions with science-backed watershed assessments. Image
This measurement revolution matters because AI adoption is exploding.

Without understanding real environmental impact, we can't make responsible decisions about technology that could reshape the entire global economy. Image
Google is open-sourcing this measurement methodology to create industry standards.

As AI unlocks trillions in economic value, we need to ensure we're tracking, and minimizing, its true environmental cost.

You can read the full paper here: cloud.google.com/blog/products/…
If you found this thread valuable:

1. Follow me @RubenHssd for more threads around what's happening around AI and it's implications.

2. RT the first tweet

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Ruben Hassid

Ruben Hassid Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @RubenHssd

Aug 12
Meta just won the world's biggest brain competition by building an AI that can READ YOUR MIND while you watch movies.

1st place out of 263 teams.

This is the most insane paper I've ever read: 🧵

(hint: mind reading is here)
For context, the Algonauts competition challenged teams to build AI that predicts brain activity from videos.

263 teams competed.

Meta crushed it with the biggest 1st-2nd place gap ever.

Let me break down how: Image
TRIBE (TRImodal Brain Encoder) is the first AI trained to predict brain responses across multiple senses simultaneously.

Most brain studies focus on one thing; vision OR hearing OR language.

TRIBE does all three at once, just like your actual brain. Image
Read 17 tweets
Aug 5
China built a computer with 2 billion neurons mimicking a monkey's brain.

If Moore's Law is still valid, we will have human-level brain computers with 86 billion neurons by 2033.

We are closer to duplicating humans.

Thread Image
China's progress is insane:

2020: Darwin Mouse (120 million neurons)
2025: Darwin Monkey (2 billion neurons)
2027: 4 billion neurons
2030: 16 billion neurons
2033: 86 billion neurons ← Human brain level

China went from mouse to monkey in 5 years. Image
What does a human brain computer actually mean?

Every thought, memory, and decision you make could theoretically be replicated in silicon.

We're talking about artificial consciousness that thinks like you do.
Read 11 tweets
Aug 3
NVIDIA just dropped paper exposing a $57 billion AI industry mistake.

While Big Tech keeps pushing expensive LLMs like ChatGPT & Claude...

Small language models handle 70% of AI agent work at 1/30th the cost.

Here's why this changes everything:

(hint: less is more) Image
→ The $57 billion mistake ↓

The AI industry invested massively in centralized LLM infrastructure in 2024.

But the actual market for LLM API services is only $5.6 billion.

That's a 10x gap between investment and revenue no one wants to admit. Image
→ Most companies are betting everything on one operational model that may be fundamentally flawed.

They assume centralized, generalist LLMs will remain the cornerstone without substantial alterations.

The problem? This assumption is about to get very expensive. Image
Read 19 tweets
Jul 30
BREAKING: Scientists just analyzed 740,000 hours of human speech across YouTube and podcasts.

Turns out, ChatGPT is rewiring how humans speak to each other.

Here's what they discovered:

(hint: the first AI to successfully colonize our brains) Image
This shook me up first:

The changes showed up in SPONTANEOUS conversations, not scripts or prepared thoughts.

Random people chatting on podcasts started using ChatGPT's favorite words without realizing it.

The way scientists proved this was ingenious ↓ Image
They fed thousands of human texts to ChatGPT for "editing" and tracked every single change.

ChatGPT uses certain words up to 300x more than humans naturally would.

300 times. Not 3x or 30x, but three hundred.
Read 16 tweets
Jul 23
Years before ChatGPT existed, Sam Altman saw one YC startup that he believed would be bigger than all the others combined.

He put millions of his own money. Became chairman. Later took it public himself.

Today it's up 500% with ZERO revenue.

Here's the wild story of Oklo: Image
Years ago, Sam realized there couldn't be abundant AI without abundant energy.

Now that ChatGPT exists and data centers are exploding from 5% to 12% of US power demand, tech giants will pay ANY price for clean energy.

Sam was right. Image
Traditional nuclear is a disaster.

Take Vogtle in Georgia:
- 15 years to build
- $35 BILLION cost
- Reactor vessel alone: 12,200 tons
- Needs thousands of workers
- Consumes 3L water per kWh produced

We can't build these fast enough for AI.
And that's what Oklo solves Image
Read 13 tweets
Jun 17
BREAKING: Stanford just surveyed 1,500 workers and AI experts about which jobs AI will actually replace and automate.

Turns out, we've been building AI for all the WRONG jobs.

Here's what they discovered:

(hint: the "AI takeover" is happening backwards) Image
→ Workers don't want full automation. They want partnership.

46.1% of tasks got positive automation ratings, but here's the kicker:

Workers prefer "H3 equal partnership" with AI over full replacement. Image
→ Scientists divided AI adoption into 4 zones based on worker desire vs. technical capability.

41% of startup investments are going to the WRONG zones, i.e. areas workers don't want automated or that have low capability. Image
Read 17 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(