Epoch AI Profile picture
Nov 10 11 tweets 3 min read Read on X
AI data center buildouts already rival the Manhattan Project in scale, but there’s little public info about them.

So we spent the last few months reading legal permits, staring at satellite images, and scouring news sources.

Here’s what you need to know. 🧵 Image
AI data centers will be some of the biggest infrastructure projects in history

e.g. OpenAI’s Stargate Abilene will need:

- As much power as Seattle (1 GW)

- >250× the compute of the GPT-4 cluster

- 450 soccer fields of land

- $32B

- Thousands of workers

- 2 years to build
By the end of the year, AI data centers could collectively see >$300 billion in investment, around 1% of US GDP.

That’s bigger than the Apollo Program (0.8%) and Manhattan Project (0.4%) at their peaks.
Only a few countries have enough power to build many >1 GW data centers like Stargate

E.g. 30 GW is ~5% of the US’ power, ~2.5% of China’s, but ~90% of the UK’s

Other countries can build some frontier data centers and grow their power capacity — but they need more time/money
So power is the core determinant of where AI data centers are built.

Other factors like latency matter surprisingly little — it takes >100× longer to generate model responses than transmit data from Texas to Tokyo.

Even serving LLMs from the Moon may not be a big latency issue!
Where does this power come from?

Usually a mix of on-site fossil fuel generation and interconnection to the grid.

E.g. Stargate Abilene will start off with on-site natural gas, then connect to the grid to access Texas’ abundant renewable power.
This power runs IT equipment on “server racks” with a small area of 0.5 m^2. But each rack uses enough power for 100 homes!

This means a huge amount of heat in a small space. So you can’t cool these chips with fans — you need liquid coolants to efficiently soak up the heat. Image
3 implications for AI policy/impacts:

1) AI’s climate impact has been small so far.

For example, AI data centers use 1% of US power vs 8% for lighting and air conditioning (12%).

But if trends continue, this could change in the next decade.
2) AI data centers are growing fast enough to support 5×/year growth in frontier training compute.

So companies probably won’t need to decentralize AI training over the next 2 years.

However, in practice they might choose to anyway, e.g. to soak up excess power in the grid.
3) Thousands of people from multiple parties build >1 GW data centers, and cooling infrastructure can be spotted using satellite imagery.

So unlike the Manhattan Project, AI data center buildouts are hard to keep secret from the rest of the world.

epoch.ai/data/data-cent…
Much more in the post!

The post was written by @ansonwhho, @benmcottier, and @YafahEdelman.epoch.ai/blog/what-you-…

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Epoch AI

Epoch AI Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @EpochAIResearch

Nov 7
The Epoch Capabilities Index is a useful way to measure model capabilities, but what does a score of 150 actually mean?

One way to read our new capability index is by plotting the benchmark performance you expect to see, for a range of ECI scores 🧵 Image
Three important takeaways:

1. Benchmarks vary in overall difficulty, and in slope. Steeper slopes imply a narrower range of difficulties at the question level and mean the benchmark saturates quickly once some progress is made.
2. While a model with a score of 140 is expected to get 45% on SWE-Bench Verified, this is just an expectation. Individual models perform better or worse on specific tasks.

For instance, GPT-5 underperforms in GPQA Diamond but overperforms in VPCT.
Read 5 tweets
Nov 4
Announcing our Frontier Data Centers Hub!

The world is about to see multiple 1 GW+ AI data centers.

We mapped their construction using satellite imagery, permits & public sources — releasing everything for free, including commissioned satellite images.

Highlights in thread! Image
Several data centers will soon demand 1 GW of power, starting early next year:

- Anthropic–Amazon New Carlisle (January)
- xAI Colossus 2 (February)
- Microsoft Fayetteville (March, borderline 1GW)
- Meta Prometheus (May)
- OpenAI Stargate Abilene (July) Image
The largest 2026 facility (xAI Colossus 2) will have the compute equivalent of 1.4M H100 GPUs.

Even larger data centers are coming: Meta Hyperion and Microsoft Fairwater will each have 5M H100e when they reach full capacity in late 2027 to early 2028. Image
Read 8 tweets
Oct 9
We evaluated Gemini 2.5 Deep Think on FrontierMath. There is no API, so we ran it manually. The results: a new record!

We also conducted a more holistic evaluation of its math capabilities. 🧵 Image
Note that this is the publicly available version of Deep Think, not the version that achieved a gold medal-equivalent score on the IMO. Google has described the publicly available Deep Think model as a “variation” of the IMO gold model.
Good performance on FrontierMath requires deep background knowledge and precise execution of computations. Deep Think has made progress but hasn’t yet mastered these skills, still scoring lower on the harder tiers of the benchmark. Image
Read 10 tweets
Oct 3
Sora 2 can solve questions from LLM benchmarks, despite being a video model.

We tested Sora 2 on a small subset of GPQA questions, and it scored 55%, compared to GPT-5’s score of 72%.
GPQA Diamond is a benchmark of challenging multiple-choice science questions, like the attached example. We randomly selected 10 questions from the benchmark, and tried running Sora on them until we generated four videos per question. Image
To evaluate Sora on a test designed for language models, we prefixed prompts requesting a video of a professor showing the answer letter (A–D) on a piece of paper. Videos without an unambiguous letter were counted as incorrect.
Read 5 tweets
Sep 30
Announcing our new AI Companies Data Hub!

We collected key data on frontier AI companies, including revenue run rates, funding, staff, usage rates, and compute spend.

This free resource will help you understand the trajectory and economics of AI.

Highlights in thread! Image
Revenue:

The combined revenue rates of OpenAI and Anthropic have grown around 10x since early 2024.

OpenAI’s annualized revenue reached $13 billion in August 2025, up from $5B at the start of the year.

Anthropic’s revenue has exploded this year, from $1B to $5B by July! Image
Funding:

OpenAI, Anthropic, and xAI have attracted massive investor interest.

OpenAI's last raised at a value of $300 billion, with a $500B valuation under discussion.

And collectively, the frontier AI companies in our data have raised ~$100B in equity and debt funding. Image
Image
Read 7 tweets
Sep 26
Why did OpenAI train GPT-5 with less compute than GPT-4.5?

Due to the higher returns to post-training, they scaled post-training as much as possible on a smaller model

And since post-training started from a much lower base, this meant a decrease in total training FLOP 🧵 Image
The invention of reasoning models made it possible to greatly improve performance by scaling up post-training compute. This improvement is so great that GPT-5 outperforms GPT-4.5 despite having used less training compute overall.
How did OpenAI do this?

They rapidly scaled up post-training, while scaling down pre-training enough for total training compute to decrease.
Read 11 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(