As someone who has reported on AI for 7 years and covered China tech as well, I think the biggest lesson to be drawn from DeepSeek is the huge cracks it illustrates with the current dominant paradigm of AI development. A long thread. 1/
First, what is DeepSeek? A Chinese firm that was able to produce an open-source AI model with roughly 1/50th of the resources of state-of-the-art models yet still beat OpenAI’s o1 on several benchmarks. 2/
Much of the coverage has been focused on US-China tech competition. That misses a bigger story: DeepSeek has demonstrated that scaling up AI models relentlessly, a paradigm OpenAI introduced & champions, is not the only, and far from the best, way to develop AI. 3/
Thus far OpenAI & its peer scaling labs have sought to convince the public & policymakers that scaling is the best way to reach so-called AGI. This has always been more of an argument based in business than in science. 4/
There is empirical evidence that scaling AI models can lead to better performance. For businesses, such an approach lends itself to predictable quarterly planning cycles and offers a clear path for beating competition: Amass more chips. 5/
The problem is there are myriad huge negative externalities of taking this approach - not least of which is that you need to keep building massive data centers, which require the consumption of extraordinary amounts of resources. 6/
Many journalists have written extensively about those externalities.
When these massive data centers come to town, they distort power supplies (per @business’s @Leonardonclt, @naurtorious, @andretartar)… 7/ bloomberg.com/graphics/2024-…
…they take up drinking water - yes, drinking water, because of the water quality required to cool these facilities - which threatens to hike water prices for families… 8/ theatlantic.com/technology/arc…
…they extend the lives of gas & coal plants, worsening air quality (per @washingtonpost’s @evanhalper) - not just because fossil fuels are the fastest way to surge energy supply but because 24/7 facilities can't run solely on intermittent renewables. 9/ washingtonpost.com/business/2024/…
Then there are the global impacts. More coal & gas means more carbon emissions, accelerating our climate crisis. Hence why Google & MSFT saw a ~50% & 30% jump in emissions since 2019 & 2020 respectively (from @darakerr at @npr). 10/ npr.org/2024/07/12/g-s…
OpenAI & co have come up with a funny argument for why to accept these externalities: Yes unfortunately there are trade-offs in the short term, they say, but they're necessary to reach so-called AGI; AGI will then help us solve all these problems! So trade offs = worth it. 11/
This line of thinking has been so effective in capturing the mindshare of powerful decision makers that president Trump just last week announced the Stargate Project, a $500B private investment into data centers & other computing infrastructure for OpenAI. 12/
Should Stargate follow through with $500B, it'll be the largest infrastructure spend in history and significantly accelerate the already startling pace of data center development post-ChatGPT. In turn that will significantly accelerate the aforementioned consequences. 13/
Now on top of the negative externalities on people’s power, water, air, and global climate, we have another: the ceding of more & more control over critical energy & water infrastructure to Silicon Valley. I wrote about this last week in @TheAtlantic. 14/ theatlantic.com/technology/arc…
But DeepSeek shows - in the same moment as the Stargate announcement - that the trade off that OpenAI & co frame as wholly necessary is actually not. 15/
Wait a minute. You mean to say that we don’t need to blanket the earth with data centers and coal & gas plants to maybe arrive at a future where we can wave a magical AGI wand to make all of the consequences of that go away?
Yes. This is a false trade off. Let that sink in. 16/
As I said earlier, scaling was always more about business than science. Scientifically, there’s no law of physics that says AI advancements must come from scaling rather than approaches using the same or fewer resources. Scaling is just an incredibly easy-to-follow formula. 17/
DeepSeek now also undermines the business case for scaling. OpenAI has been burning through staggering sums of cash to keep up its scaling paradigm and has yet to figure out how to balance its checkbooks - and it turns out it didn’t need to spend so much cash. 18/
So it doesn’t matter if you're a company in the US, China, or elsewhere. DeepSeek should be a cue to pivot HARD toward investing in far more efficient methods of AI development. Even if you care nothing about community & climate impacts, it’s just better business. 19/
And if a company does not make this pivot, that should be a major red flag for its capacity to innovate. And I don’t mean product innovation but innovation on how to develop AI models. The base of the stack. 20/
When I worked at an SV startup, we used to call an investment from SoftBank the kiss of death. SoftBank would pump so much cash into startups that it would completely kill that startup’s need to innovate or develop a financially sustainable business. 21/
They could mask all of the shakiness in their foundations with the hilariously large infusions of cash for quite a while — but eventually everything would topple. Necessity is the mother of innovation. Without necessity, the existential drive to innovate disappears. 22/
We’re basically seeing this again. OpenAI with Sam Altman is so good at raising capital (incl. now from Softbank) that it's covering up the company’s technical & commercial weaknesses. See also this excellent @AINowInstitute report from @bcmerchant. 23/ ainowinstitute.org/general/ai-gen…
DeepSeek is the flip side of the same coin: It innovated because of, not in spite of, its constraints. And now that it has upended the assumptions of the dominant AI paradigm, we should reject its costly trade-offs and seek new ways to develop AI without so many harms. 24/
• • •
Missing some Tweet in this thread? You can try to
force a refresh
Some news years in the making: My book EMPIRE OF AI, out May 20, is ready for pre-order at . It tells the inside story of OpenAI as a lens for understanding the moment we’re in: the tech elite's extraordinary seizure of power and its threat to democracy. 1/ empireofai.com
This book is the culmination of my ~7 years of reporting on AI for @techreview, @WSJ, and @TheAtlantic. It is based on 300+ interviews with ~260 people, including 150+ interviews with 90+ current & former OpenAI people, and an extensive trove of correspondence and documents. 2/
The book asks the central question: How do we govern AI? More importantly, *who* should govern it? AI is the most consequential technology of our era. The future of AI is thus inextricably tied to our future, and whether it will get better or worse. 3/
Free speech no longer exists in the US government. For @TheAtlantic I spoke with 12+ federal workers in 6 agencies who said the Trump admin’s actions have led to pervasive self-censorship, even on issues some view as critical to national security. 1/ theatlantic.com/technology/arc…
Transitions of power have always led to changes in priorities, but that is not what the workers say they're witnessing. The executive orders, mass firings & internal comms amount to what some feel can only be described as the administration engineering ideological obedience. 2/
One example: After the inauguration, Secretary Marco Rubio sent an email to the State Department outlining its new priorities. He included the text below, accusing workers of censoring Americans while dictating the new "truth." A State employee described this as “Orwellian.” 3/
For decades, the US government has painstakingly kept American science #1 globally—and every facet of American life has improved because of it. The internet? Flu shot? Ozempic? All grew out of federally-funded research. Now all that's being dismantled. 1/ technologyreview.com/2025/02/21/111…
For @techreview I spoke with 10+ federal workers who occupy, or until recently occupied, scientific and technical positions across various agencies. They explained to me how much behind-the-scenes work the US government does to keep America’s engine of innovation humming. 2/
Government funds account for more than *half* the R&D budget that US universities spend each year—in 2023, that was $60B out of $109B. The gov also does a lot of its own research at national labs like Oak Ridge & Los Alamos and agencies like NOAA and DoD.
To the public, Microsoft uses its reputation as an AI & sustainability leader to tell a compelling story: AI will do wonders to help solve the climate crisis. To fossil-fuel firms, Microsoft has a different message: AI will help them drill, baby, drill. 1/ theatlantic.com/technology/arc…
For more than a year, I’ve been poring over hundreds of pages of internal Microsoft documents, many of which were shared with the SEC, and interviewing current and former employees and execs on the giant's engagements with the oil & gas (O&G) industry. 2/
Microsoft doesn’t just passively provide its services to these companies. It develops bespoke AI-enhanced tools for them, which it also markets to them as for the explicit purpose of optimizing and automating drilling, and maximizing fossil-fuel production. 3/
For years I’ve been interviewing data annotation workers who are the lifeblood of the AI industry. For years I’ve heard the same story: the platforms they work for wield total power, leaving them precarious & vulnerable to exploitation. A horrible example of this just happened 1/
Several workers in Kenya told me - confirmed by local media reports - that Remotasks, one of the largest AI data work platforms in the country that Scale AI owns and OpenAI uses, has suddenly suspended their operations in Kenya. When workers try to log on, they see this. 2/
No reason given, little advanced notice. This was the sole source of income for some families. One of the workers I interviewed & kept in touch with is a mother of four kids now desperately struggling to feed them and on the verge of being evicted. This is not humane. 3/
A big question looms over generative AI: what really is its impact on the environment? I spent months investigating a single campus of Microsoft data centers in the Arizona desert - designated in part for OpenAI - in an attempt to find out. Thread. theatlantic.com/technology/arc…
Some broader context: Concrete numbers on the energy and water consumption, and carbon emissions, of generative AI remain frustratingly elusive. Many, many people have said this - most recently @SenMarkey, @SenatorHeinrich, @RepAnnaEshoo & @RepDonBeyer. markey.senate.gov/news/press-rel…
We have some wonderful scholarship - shout out to people like @SashaMTL and @strubell - that has measured the training costs of GPT-like large language models () and the inference costs of open-source generative AI applications ().arxiv.org/pdf/2206.05229… arxiv.org/pdf/2311.16863…