Ruben Hassid Profile picture
Aug 25, 2025 14 tweets 5 min read Read on X
For the first time, Google has measured how much energy AI really uses in production.

Spoiler: the gap vs. all previous estimates is huge... 🧵 Image
Despite AI transforming healthcare, education, and research, we've been flying blind on its environmental footprint.

Every estimate was based on lab benchmarks, not real-world production systems serving billions of users.

Google decided to measure what actually happens. Image
The results from measuring Gemini in production:

• 0.24 watt-hours per text prompt
• Equivalent to watching TV for 9 seconds
• 5 drops of water consumed
• 0.03 grams of CO2 emissions

Substantially lower than public estimates. Image
Even if you add the water used in nearby power plants to generate electricity, the total rises only to 2.7 mL per prompt.

That’s 0.0001% of the average American’s daily water use.

Substantially lower than public estimates. Image
Over just 12 months, while AI quality improved dramatically, Gemini's environmental impact plummeted:

• Energy per prompt: down 33x
• Carbon footprint: down 44x

This is efficiency innovation at scale. Image
Why were all previous estimates wrong?

They measured isolated chips running perfect benchmarks.

Google measured the full reality: idle machines for reliability, cooling systems, power distribution, CPU overhead... everything needed for global AI service. Image
The measurement gap is enormous:

• Lab benchmark approach: 0.10 Wh per prompt
• Full production reality: 0.24 Wh per prompt

That 2.4x difference reveals the infrastructure complexity everyone else ignored. Image
→ The efficiency breakthroughs come from Google's full-stack AI approach.

Transformer architecture (invented at Google) delivers 10-100x efficiency over previous language models.

Mixture-of-Experts activates only needed model parts, cutting computation by 10-100x. Image
→ Google builds custom chips specifically for AI instead of using regular computer chips.

Their latest AI chips are 30x more energy-efficient than their first generation.

Like designing a race car for Formula 1 instead of using a family sedan. Image
→ Smart software tricks that save massive energy

Small AI models write rough drafts, big models just check the work.

Models automatically move between chips based on demand, so nothing sits idle. Image
→ Infrastructure efficiency that matters at billions of prompts

Google's data centers operate at 1.09 PUE, only 9% energy overhead beyond actual computing.

Advanced cooling balances energy, water, and emissions with science-backed watershed assessments. Image
This measurement revolution matters because AI adoption is exploding.

Without understanding real environmental impact, we can't make responsible decisions about technology that could reshape the entire global economy. Image
Google is open-sourcing this measurement methodology to create industry standards.

As AI unlocks trillions in economic value, we need to ensure we're tracking, and minimizing, its true environmental cost.

You can read the full paper here: cloud.google.com/blog/products/…
If you found this thread valuable:

1. Follow me @RubenHssd for more threads around what's happening around AI and it's implications.

2. RT the first tweet

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Ruben Hassid

Ruben Hassid Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @rubenhassid

Oct 8, 2025
I bet 99% of people who use ChatGPT don't know how to set it up to make it 10x more useful.

They obsess over prompts, but prompts are only 20% of the equation.

Setup is 80%.

In this thread, I'll show you how to (actually) set up your ChatGPT:
→ First, take a look at how much ChatGPT knows about you.

Then, delete everything.

Go to Settings > Personalization > Manage Memories > Delete All Memories.

Most users have 6+ months of random, contradictory memories that actively hurt performance.
→ The 13-Question Framework.

Use this prompt with GPT-5: "Design my ChatGPT digital twin. Ask 13 items one at a time:

Identity, Role, Audience, Outputs, Personality, Voice, Formatting, Values, Projects, Goals, Preferences, Do-Not List, Privacy.

Target <1300 words total." Image
Read 15 tweets
Sep 21, 2025
A new Yale paper reveals the brutal reality of the AGI economy:

Half the population could stop working tomorrow and GDP wouldn't budge.

Humans become economically meaningless.

The paper suggests we'll keep our jobs but lose something far more important: Image
We lose our economic purpose.

For centuries, human labor drove progress. We built cities, advanced science, created wealth. Work meant you mattered.

In the AGI economy, that connection breaks.

We keep jobs but lose our role as drivers of growth and progress.
The key insight comes from distinguishing "bottleneck" vs "accessory" work: Image
Read 9 tweets
Aug 27, 2025
BREAKING: New Stanford study tracking 25 million US workers finds AI is systematically eliminating entry-level jobs.

Here are 6 disturbing facts from one of the largest AI employment study ever conducted:

(hint: young workers are getting obliterated) Image
Fact 1: Employment for early-career workers (ages 22-25) has declined substantially in occupations most exposed to AI.

Software developers aged 22-25 saw nearly 20% employment decline since late 2022, while older workers in the same occupations continued to grow. Image
Fact 2: Overall employment continues to grow robustly, but employment growth for young workers has been stagnant since late 2022.

In the highest AI-exposed occupations, young workers declined 6% while older workers in those same occupations grew 9%. Image
Read 12 tweets
Aug 12, 2025
Meta just won the world's biggest brain competition by building an AI that can READ YOUR MIND while you watch movies.

1st place out of 263 teams.

This is the most insane paper I've ever read: 🧵

(hint: mind reading is here)
For context, the Algonauts competition challenged teams to build AI that predicts brain activity from videos.

263 teams competed.

Meta crushed it with the biggest 1st-2nd place gap ever.

Let me break down how: Image
TRIBE (TRImodal Brain Encoder) is the first AI trained to predict brain responses across multiple senses simultaneously.

Most brain studies focus on one thing; vision OR hearing OR language.

TRIBE does all three at once, just like your actual brain. Image
Read 17 tweets
Aug 5, 2025
China built a computer with 2 billion neurons mimicking a monkey's brain.

If Moore's Law is still valid, we will have human-level brain computers with 86 billion neurons by 2033.

We are closer to duplicating humans.

Thread Image
China's progress is insane:

2020: Darwin Mouse (120 million neurons)
2025: Darwin Monkey (2 billion neurons)
2027: 4 billion neurons
2030: 16 billion neurons
2033: 86 billion neurons ← Human brain level

China went from mouse to monkey in 5 years. Image
What does a human brain computer actually mean?

Every thought, memory, and decision you make could theoretically be replicated in silicon.

We're talking about artificial consciousness that thinks like you do.
Read 11 tweets
Aug 3, 2025
NVIDIA just dropped paper exposing a $57 billion AI industry mistake.

While Big Tech keeps pushing expensive LLMs like ChatGPT & Claude...

Small language models handle 70% of AI agent work at 1/30th the cost.

Here's why this changes everything:

(hint: less is more) Image
→ The $57 billion mistake ↓

The AI industry invested massively in centralized LLM infrastructure in 2024.

But the actual market for LLM API services is only $5.6 billion.

That's a 10x gap between investment and revenue no one wants to admit. Image
→ Most companies are betting everything on one operational model that may be fundamentally flawed.

They assume centralized, generalist LLMs will remain the cornerstone without substantial alterations.

The problem? This assumption is about to get very expensive. Image
Read 19 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(