@emilymbender.bsky.social Profile picture
Sep 19, 2022 25 tweets 12 min read Read on X
This article in the Atlantic by Stephen Marche is so full of #AIhype it almost reads like a self-parody. So, for your entertainment/education in spotting #AIhype, I present a brief annotated reading:

theatlantic.com/technology/arc…

/1
Straight out of the gate, he's not just comparing "AI" to "miracles" but flat out calling it one and quoting Google & Tesla (ex-)execs making comparisons to "God" and "demons".

/2 Screencap from linked article: "Miracles can be perplex
This is not the writing of someone who actually knows what #NLProc is. If you use grammar checkers, autocorrect, online translation services, web search, autocaptions, a voice assistant, etc you use NLP technology in everyday life. But guess what? NLP isn't a subfield of "AI".
/3 Screencap, same article: "Early artificial intelligence
Here's the author is claiming to have inside knowledge of some "esoteric" technology development that, unbeknownst to the average human, is going to be very disruptive. But note the utter lack of citations or other grounding for this claim.

/4 Screencap, same article: "Or rather: If you are using N
Okay, agreed on fake-it-til-you-make-it, but "direct thrust at the unfathomable" and "not even the engineers understand" are just unadulterated hype. If they don't understand how it works, how are they even measuring that it works?

/5 Screencap, same article: "Science fiction, and our own
Protip: They aren't really. The capabilities that the AI boosters claim to have built are ones that we don't have effective benchmarks for, & actually can't, in principle. See: AI and the Everything in the Whole Wide World Benchmark by @rajiinio et al /6

…ets-benchmarks-proceedings.neurips.cc/paper/2021/has…
et al = @cephaloponderer @alexhanna @amandalynneP and me.

For a quick overview, see this article by @bendee983

bdtechtalks.com/2021/12/06/ai-…

/7
@cephaloponderer @alexhanna @amandalynneP @bendee983 Okay, back to the hype. This is weirdly ominous and again provides no supporting evidence. You can't see it, but that doesn't mean it isn't there ... is not an argument that it is!

/8 Screencap, same article: "But the confusion surrounding
This is kinda fun, because I was musing a few weeks ago about how we don't usually go to "superhuman" for other tools. And it does sound ridiculous, doesn't it?



/9 Screencap, same article: "All technology is, in a sense
If you don't know how something works, but can test that it works (w/certain degree of reliability), then it is usable. It's true that deep learning is opaque on the how. But we can't let any engineers off the hook in terms of testing the functionality of their systems.

/10 Screencap, same article: "But the sorcery of artificial
"What technologists call 'parameters'" makes this sound so ominous and mysterious. Our "little animal brains" have ~86 billion neurons (source: brainfacts.org/in-the-lab/mee…). So not a different scale (and with much more complexity).

/11 Screencap, same article: "The details of how this could
More to the point: None of this is inevitable. DL systems aren't naturally occurring phenomena that we can try to understand or just stand in awe of. They are things we are building & choosing to use. We can choose not to, at least w/p sufficient testing for each use case.

/12
Also, because it feels gross to compare language model parameters to human neurons, I want to plug again this great article by @AlexBaria and @doctabarz on the computational metaphor.

arxiv.org/abs/2107.14042

/13
Back to Marche: I don't think we should necessarily believe the people who got super rich off of surveillance capitalism when they say "oh noes, can't regulate, it would stop the development of the technology".

/14 Screencap, same article: "This unfathomability poses a
Again, whether or not we try to build this (and with what regulatory guardrails) is a CHOICE. But also: it would be pretty easy with today's stochastic parrots to sometimes at least get an answer like that. (While other times getting hate speech...)

/15 Screencap, same article: "Others have headed into deepe
Uh, just because you put these things in a list does not make them all the same kind of thing ("language game").

/16 Screencap, same article: "What we are doing is teaching
Yeah, just because the people who built the thing say it does something "in a ways that's not dissimilar from the way you and I do" doesn't make it true. Do they have the expertise to evaluate that? How did they evaluate that?

/17 Screencap: "PaLM, Google’s latest foray into NLP, has
Oh, and again, while "contemporary NLP" does use neural LMs for a lot of things, I wouldn't say it "derives" from them. There is more to the field than just throwing neural nets are poorly conceived tasks.

/18
What comes next is some GPT-3 authored additional hype, stating with the prompt "And if AI harnesses the power promised by quantum computing," Marche does acknowledge it (in the following paragraph). He is also responsible for deciding to include it.

/19 Screencap (GPT-3 authored text): "And if AI harnesses tScreencap (Marche): "Our AI future will be weird and su
(Note that Marche also doesn't tell us how many tries he took to get the one he chose to include.)

/20
It's not doing any of these things, actually. Having synthetic text in the style of someone who has died is not bringing them back from the dead. I'm not sure what an "imitation" of consciousness is, nor how it would benefit us.

/21 Screencap: "Technology is moving into realms that were
And it is certainly not "piercing the heart of how language works between people".

On how LM-geneated text is nothing like human linguistic behavior, see Bender & Koller 2020 and also this episode of Factually!

aclweb.org/anthology/2020…
earwolf.com/episode/the-re…

/22
And one last screencap before I end. Where is the evidence for any of these claims? None is provided.

/23 Screencap: "A fragment of humanity is about to leap for
So, I hope that was enjoyable and/or informative. I give this one #threemarvins. Could 2022 be the year of peak #AIhype? That sure would be nice. 24/24
Postscript 1: Important additional info on the (sigh) comparison of 100B parameter networks to human brains

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with @emilymbender.bsky.social

@emilymbender.bsky.social Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @emilymbender

Nov 4
As OpenAI and Meta introduce LLM-driven searchbots, I'd like to once again remind people that neither LLMs nor chatbots are good technology for information access.

A thread, with links:

>>
@chirag_shah and I wrote about this in two academic papers:
2022: dl.acm.org/doi/10.1145/34…
2024: dl.acm.org/doi/10.1145/36…

We also have an op-ed from Dec 2022:
iai.tv/articles/all-k…

>>
Why are LLMs bad for search? Because LLMs are nothing more than statistical models of the distribution of word forms in text, set up to output plausible-sounding sequences of words.



>>
Read 15 tweets
Feb 29
It seems like there are just endless bad ideas about how to use "AI". Here are some new ones courtesy of the UK government.

... and a short thread because there is so much awfulness in this one article.
/1


ft.com/content/f2ae55…
Screencap: "UK ministers are piloting the use of generative artificial intelligence to analyse responses to government consultations and write draft answers to parliamentary questions.  Oliver Dowden, the deputy prime minister, will on Thursday unveil tools that the AI “crack squad” at the heart of Whitehall is trialling with a view to wider rollouts across central departments and public services."
Either it's a version of ChatGPT OR it's a search system where people can find the actual sources of the information. Both of those things can't be true at the same time. /2 Screencap: "The AI tools include using government-hosted versions of ChatGPT and a mix of open-source AI models securely hosted in-house to draft preliminary responses to questions to ministers submitted by MPs and to freedom of information requests.  The drafts would always be checked by a human civil servant and the AI tools are programmed to ensure they cite their sources on all claims, so they can be verified."
Also: the output of "generative AI", synthetic text, is NOT information. So, UK friends, if your government is actually using it to respond to freedom of information requests, they are presumably violating their own laws about freedom of information requests. /3
Read 10 tweets
Jan 14
It is depressing how often Bender & Koller 2020 is cited incorrectly. My best guess is that ppl writing abt whether or not LLMs 'understand' or 'are agents' have such strongly held beliefs abt what they want to be true that this impedes their ability to understand what we wrote.
Or maybe they aren't actually reading the paper --- just summarizing based on what other people (with similar beliefs) have mistakenly said about the paper.

>>
Today's case in point is a new arXiv posting, "Are Language Models More Like Libraries or Like Librarians? Bibliotechnism, the Novel Reference Problem, and the Attitudes of LLMs" by Lederman & Mahowald, posted Jan 10, 2024.



>>arxiv.org/pdf/2401.04854…
Read 11 tweets
Dec 7, 2023
A quick thread on #AIhype and other issues in yesterday's Gemini release: 1/
#1 -- What an utter lack of transparency. Researchers form multiple groups, including @mmitchell_ai and @timnitgebru when they were at Google, have been calling for clear and thorough documentation of training data & trained models since 2017. 2/
In Bender & Friedman 2018, we put it like this: /3 Screecap: "These two recommendations will need to be implemented with care. We have already noted the potential barrier to access. Secrecy concerns may also arise in some situations (e.g., some groups may be willing to share datasets but not demographic information, for fear of public relations backlash or to protect the safety of contributors to the dataset). That said, as consumers of datasets or products trained with them, NLP researchers, developers, and the general public would be well advised to use systems only if there is access to the information we propose should be included ...
Read 20 tweets
Nov 24, 2023
With the OpenAI clownshow, there's been renewed media attention on the xrisk/"AI safety" nonsense. Personally, I've had a fresh wave of reporters asking me naive questions (+ some contacts from old hands who know how to handle ultra-rich man-children with god complexes). 🧵1/
As a quick reminder: AI doomerism is also #AIhype. The idea that synthetic text extruding machines are harbingers of AGI that is on the verge of combusting into consciousness and then turning on humanity is unscientific nonsense. 2/
t the same time, it serves to suggest that the software is powerful, even magically so: if the "AI" could take over the world, it must be something amazing. 3/
Read 27 tweets
Jun 11, 2023
There's a lot I like in this op-ed, but unfortunately it ends with some gratuitous ableism (and also weird remarks about AGI as a "holy grail").

First, the good parts:

theguardian.com/commentisfree/…
"[False arrests w/face rec tech] should be at the heart of one of the most urgent contemporary debates: that of artificial intelligence and the dangers it poses. That it is not, and that so few recognise it as significant, shows how warped has become the discussion of AI,"

>>
"We have stumbled into a digital panopticon almost without realising it. Yet to suggest we live in a world shaped by AI is to misplace the problem. There is no machine without a human, and nor is there likely to be."

>>
Read 7 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(