Jonathan Zittrain Profile picture
May 30 43 tweets 18 min read Twitter logo Read on Twitter
Today, a crisp one-sentence open letter warning about existential AI threat: “Mitigating the risk of extinction from AI should be a global priority alongside other societal-scale risks such as pandemics and nuclear war.”

I did not sign the letter.

A lot of smart, thoughtful, genuinely brilliant colleagues have signed the letter, and it follows on earlier alarms about AI, including one from February worried that “Advanced AI could represent a profound change in the history of life on Earth ... .” nytimes.com/2023/03/29/tec…
At the risk of being the blockhead in the first ten minutes of the horror movie who confidently heads to the basement to show that the strange noises are entirely normal, I thought I’d explain why I didn't sign.
The February letter garnered over 1,800 signatories, including Elon Musk, though in a sign of our times an indeterminate number of signatures (including those of Xi Jinping and Meta chief AI scientist @ylecun) turned out to be fake. theguardian.com/technology/202…
That letter dramatically called for a six-month moratorium -- voluntary or if necessary by law (!) -- on further AI large language model development. To be sure, several weeks after calling for the pause, Musk appears to have gone all in on AI development. businessinsider.com/elon-musk-twit…
Existential risk from AI is not a new concept; imagining machines (or non-human humans more generally) that turn on us has a rich history.

(The quote below is from the December 1948 issue of TIME about the Mark III.) Cover of Mary Shelley's Fra...The Terminator(s)TIME cover on the Mark III ...From the December 1948 issu...
Traditionally the lurid tools of our extinction have been only within the reach of a powerful few – say, those with the know-how, enriched uranium, and delivery vehicles for nuclear warheads; or, as some tell it, those who built and operate supercolliders. (!) But Walter L. Wagner and Lu...
(So far CERN hasn’t created a black hole that’s devoured the Earth, or emitted a strangelet to turn it into a “shrunken dense dead lump of … strange matter”; the quickly-dismissed lawsuit against it called for an environmental impact statement that would have been for the ages.)
Signatories no doubt have different accounts of AI’s “existential risk.” Many would blanch at a Terminator-style story of killer robots. Some see AI achieving “superintelligence” – becoming smarter than humans are, individually and even collectively. en.wikipedia.org/wiki/Superinte…
Here's OpenAI's own description of superintelligence from last week. It's perhaps necessarily vague on what that is and how it might come about. It's clearer that the authors don't think today's AI systems count, so can develop more freely. openai.com/blog/governanc… Given the picture as we see...What’s not in scope We thin...
If you’d like the full Terminator argument, Holden Karnofsky’s essay from June 2022 is a good place to start. But it also starts by more or less assuming the premise of hostile superintelligence(s): cold-takes.com/ai-could-defea… So, for what follows, let's...
And @mattyglesias, writing a year ago -- ages in AI time -- is very much down with the Terminator analogy, because it communicates a justified fear of AGI to the public, even if its details don't happen to track what the experts worry about. slowboring.com/p/the-case-for… A lot of smart people like ...At any rate, the point — wh...
But I've been brought up short by some of some arm-waving around how a superintelligence emerges: if AIs are getting better (“smarter”) over time, deep questions of what “smart” means notwithstanding, the argument is that it’s just a matter of time before they surpass us.
Before LLMs like GPT came about, the arguments about "better" were sometimes couched in analogies between processor power and brainpower, and how the former could overwhelm the latter, even as raw processors alone don't make minds, any more than a pile of brains do.
Particularly if AIs start coding their own successors – in theory, they very quickly level up. But AI implementations run the gamut. The first time we’ve really seen something that acts like a colloquial artificial intelligence has been w/“large language models” like GPT.
It’s amazing what’s come of pouring billions of fragments of humanity’s words into a big pot, performing several hundred millions of dollars’ worth of computational stirring, and then making some refinements through Q&A with the resulting model (“RLHF”). washingtonpost.com/technology/int…
Like many, the first time I tried out GPT-3 I was floored. Wait, this thing passes the Turing Test! Just like that! (Sure, the Turing Test is flawed, but still…) amazon.com/Turing-Test-Be…
Some of its most impressive characteristics might look modest. If I start it off with questions in regular case and ANSWER IN CAPS it will “know” to keep going in that style, despite no explicit training or code around upper and lower cases.
And it even seems capable of cognition at times – quite a leap when it’s trained as an unsupervised “auto-regressive” model, i.e. simply predicting what tokens and words come next in a sentence.
(@MelMitchell1 has written an informative and accessible overview of evaluating reasoning in certain AI models.) aiguide.substack.com/p/why-the-abst…
If people can’t explain how GPT approximates some form of cognition in some cases – and for most meanings of testable explanation, they can’t – then it’s awfully hard to know how much better GPT can get in version 5 or 6 or 7.
Yet then GPT turns out to be lousy at something easy (for now), in ways that suggest that the way these models work is not, in fact, much like human cognition. (We're don't achieve thought thanks to reading every word of Reddit and … everything else.)
(@ylecun has written about the difference between existing model architectures and one that might recreate the sort of common sense that we take for granted in humans. openreview.net/pdf?id=BZ5a1r-…)
It’s a strange moment to have chatbots that are so unbelievably good, head and shoulders above what came before, and also so clearly innately limited – geared for coherence rather than truth. What a time to be alive – and what a time to be not alive!
When Microsoft released a version of GPT within Bing, veteran tech columnist @kevinroose was blown away – well, more precisely, deeply creeped out. Bing, a.k.a. “Sydney,” gave 2001’s HAL-9000 a run for its money. nytimes.com/2023/02/16/tec…
I still can’t get over the fact that one of the few ways the bot-maker has of getting these bots to be nice is to … tell them to be nice, before turning the same mic over to a user, who can try to belay those orders.
Which brings us back to how these bots could appear to do harm: by a person simply asking them how best to do it, and by their trying to oblige by predictive-texting their way to a retread of a Terminator or similar script knocking around inside. vice.com/en/article/93k…
But it gets worse! In a development that an earlier me would have cheered, within the past month OpenAI added plug-ins to GPT – ways for it to not only answer text with text (or text with code), but to … run that code. Or operate OpenTable or Instacart. openai.com/blog/chatgpt-p…
GPT plug-ins are hugely generative, and it seems there’s no end of both cool and uncool things that people will do with them. Like the Internet itself! For the Internet, I’ve long argued that the good has outweighed the bad. en.wikipedia.org/wiki/Generativ…
OpenAI hopes to shape plug-ins to limit bad real-world impacts across the blood-brain barrier from the online playground. But imagine GPT being prompted to pull off a bomb scare – using Craigslist to do it. (Real story; AI piece only a thought experiment!) masslive.com/news/2023/05/f…
This sort of scenario, with later generations of LLMs both less monitored and more powerful and connected, worries me a lot. An AI doesn’t have to think for itself, contemplating escaping humans’ control, to do (or to be instructed to do) bad things. en.wikipedia.org/wiki/Instrumen…
Especially in a possible world where AI models are open sourced and run on laptops (old me would cheer), some wise and careful practices are needed to try to ensure they can’t so readily connect to make things just happen in the real world. semianalysis.com/p/google-we-ha…
But strangely, this seems just the kind of scenario that the OpenAI team does *not* so much worry about -- it's not in scope for existential risk from superintelligence.

And the superintelligence that *is* in scope has upsides that militate towards (carefully) building it, too. What’s not in scope We thin...Given the risks and difficu...
As @verityharding points out on @Samfr's blog, though, it's the prominence of ChatGPT that lends talk of AI risk oxygen rn -- even though the risk, when described by those concerned, appears to be separate from GPT, and not concretely articulated beyond "human+ intelligences." AI is, however, reaching a ...
This open reply to the February AI risk letter by @timnitGebru, @emilymbender, @mcmillan_majora, and @mmitchell_ai also gets at this issue. dair-institute.org/blog/letter-st… It is dangerous to distract...
And this essay by @random_walker and @sayashk, also responding to the February open letter on AI risks, offers similar reasons to hold back. aisnakeoil.substack.com/p/a-misleading… Table of risks from https:/...
And even if we can walk and chew gum at the same time – worrying about the speculative risks as well as the ones right in front of us like bias and misuse – placing AI tools into the ranks of nuclear and biological weapons of mass destruction jumps the gun.
(@afedercooper points out that climate change is conspicuously missing; perhaps an awkward artifact of intense computing's contribution to it; a desire to avoid an ancillary topic under controversy; or an implicit claim that its massive displacement and suffering ≠ existential.)
It jumps the gun because the problem is far more ill-defined -- except in projection of awful consequences -- than the workings of nuclear or biological proliferation, and the remedies to prevent "too much computing" at that register are, well, extreme. time.com/6266923/ai-eli… Here’s what would actually ...
From a legal-policy perspective, the regulatory perimeter is unbounded if enough computers (including virtual ones like those on computational blockchains like Ethereum) amount to a bunch of highly enriched uranium. If everything must be regulated, worldwide, nothing will be.
That's a lot harder than regulating uranium, and is why analogies to orgs like the @iaeaorg don't quite work for me yet. The IAEA has a distinct membership -- sovereign states -- and remit, including keeping an eye on nuclear facilities with the assent of those who operate them.
So, as François Chollet succinctly put it after the February letter:
Of course, in transformative tech it seems like there are only two phases in evaluating risk: too early to tell, and too late to do anything about it.

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Jonathan Zittrain

Jonathan Zittrain Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @zittrain

Feb 17, 2022
Failing to deal with this company’s behavior when it started is the biggest public policy failure in the digital space in a generation. Reining it in is that much tougher now that it’s well-funded and so frequently used by government clients. washingtonpost.com/technology/202… @drewharwell
This is … not how principles are supposed to work.
It’s just such an obvious failure, by governments who are supposed to set basic ground rules, and by the big platforms that couldn’t rouse themselves to do more than send a vaguely discomfited letter when they come down like a ton of bricks on tiny infractions elsewhere.
Read 12 tweets
Jun 19, 2021
Some extraordinary quotes in this piece. It’s hard not to conclude from it that many Republican officials believe that only Republicans can be trusted to administer elections, and only Republicans can legitimately win them. (Any Rs who do certify a D victory are also purged.)
This really drives home @AdamSerwer’s observation that “fraud” can encompass the idea that one’s opponents are illegitimate participants in politics and governing, even if they’re fellow Americans. amp.theatlantic.com/amp/article/61…
Writing in 2010, @jonathanchait observed the same phenomenon: Any D electoral wins are seen as uniquely illegitimate; any subsequent D policy as unacceptable without R support, which is naturally not forthcoming. newrepublic.com/article/74316/…
Read 7 tweets
May 21, 2021
With the help of the ace @nytimes digital team, we compiled a list of ~2.2 million externally-facing hyperlinks that had been used in nytimes.com articles since its launch in 1996. The goal was to discern how many of them had fallen victim to linkrot or content drift.
A link "rots" when it no longer works -- which can happen if the server hosting it is taken down, if a web site is reorganized (especially if it changes ownership), or if a takedown request or demand has eliminated access to that page. This happens a lot.
Content drift is related to link rot -- drift happens when the link functions, but what was there at the time the link was shared or embedded in, say, a news story or academic article, has vanished or, more confusingly, changed. Here's @NASA's web site during a gov't shutdown.(!)
Read 16 tweets
May 20, 2021
From the Epic v. Apple battle over Apple's centrally-controlled app store: to justify only allowing Apple-approved apps on iPhones and iPads (and forbidding "side-loading" of other apps by phone owners), Apple's SVP of engineering disparages Mac security. apple.slashdot.org/story/21/05/19…
These are not new worries! From 2007 (!) -- just as the iPhone was taking off -- some thoughts on how security would justify centralized control and its anti-innovation drawbacks: yupnet.org/zittrain/2008/…
The original iPhone had no outside software at all! Steve Jobs explained: “We define everything that is on the phone ... The last thing you want is to have loaded three apps on your phone and then you go to make a call and it doesn’t work any more.” today.law.harvard.edu/zittrain-in-ft…
Read 4 tweets
Apr 7, 2021
NFTs are the latest reprise of the Internet's original democratizing promise -- and the gold rush that follows when there's money in the air. In this case, out of thin air. theatlantic.com/ideas/archive/…
So, @Williamjmarks5 and I sat down to work through the current NFT mania -- itself built on top of the cryptocurrency/blockchain mania -- and see whether there's a there there. (Below: is it soh-rawr-ey? oh: so-rare) NFT Valuation Report: What To Look For When Buying Sorare Ca
Pinning down the value of NFTs gets philosophical pretty quickly. But a first observation is that people might buy something because they think others might want to buy it for more, later. Classic speculation, doable without ever asking what makes something innately valuable. Investment vs. Speculation. Source attributed to "key d
Read 15 tweets
Jan 28, 2021
Last year, just before the pandemic hit, I gave the @clarehall_cam Tanner Lectures on Human Values, reflecting on how tech has empowered humanity -- and yet it rightly feels like we have less and less control. today.law.harvard.edu/gaining-power-…
That added power, for example, doesn't just make autonomous vehicles work without drivers. It means those cars can answer to, say, law enforcement authorities instead of the passengers -- locking the doors and driving them to the nearest police station if a warrant is issued.
Our 25 years of a mainstreamed Internet has seen three eras of (non)governance. The first is "rights" -- focusing on how to enjoy the new freedoms of communication without interference from governments who would surveill it and big interests (e.g., ©) who are harmed by it.
Read 9 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us on Twitter!

:(