@emilymbender.bsky.social Profile picture
Prof, Linguistics, UW // Faculty Director, CLMS // she/her // @emilymbender@dair-community.social & bsky // rep by @ianbonaparte

Sep 19, 2022, 25 tweets

This article in the Atlantic by Stephen Marche is so full of #AIhype it almost reads like a self-parody. So, for your entertainment/education in spotting #AIhype, I present a brief annotated reading:

theatlantic.com/technology/arc…

/1

Straight out of the gate, he's not just comparing "AI" to "miracles" but flat out calling it one and quoting Google & Tesla (ex-)execs making comparisons to "God" and "demons".

/2

This is not the writing of someone who actually knows what #NLProc is. If you use grammar checkers, autocorrect, online translation services, web search, autocaptions, a voice assistant, etc you use NLP technology in everyday life. But guess what? NLP isn't a subfield of "AI".
/3

Here's the author is claiming to have inside knowledge of some "esoteric" technology development that, unbeknownst to the average human, is going to be very disruptive. But note the utter lack of citations or other grounding for this claim.

/4

Okay, agreed on fake-it-til-you-make-it, but "direct thrust at the unfathomable" and "not even the engineers understand" are just unadulterated hype. If they don't understand how it works, how are they even measuring that it works?

/5

Protip: They aren't really. The capabilities that the AI boosters claim to have built are ones that we don't have effective benchmarks for, & actually can't, in principle. See: AI and the Everything in the Whole Wide World Benchmark by @rajiinio et al /6

…ets-benchmarks-proceedings.neurips.cc/paper/2021/has…

et al = @cephaloponderer @alexhanna @amandalynneP and me.

For a quick overview, see this article by @bendee983

bdtechtalks.com/2021/12/06/ai-…

/7

@cephaloponderer @alexhanna @amandalynneP @bendee983 Okay, back to the hype. This is weirdly ominous and again provides no supporting evidence. You can't see it, but that doesn't mean it isn't there ... is not an argument that it is!

/8

This is kinda fun, because I was musing a few weeks ago about how we don't usually go to "superhuman" for other tools. And it does sound ridiculous, doesn't it?



/9

If you don't know how something works, but can test that it works (w/certain degree of reliability), then it is usable. It's true that deep learning is opaque on the how. But we can't let any engineers off the hook in terms of testing the functionality of their systems.

/10

"What technologists call 'parameters'" makes this sound so ominous and mysterious. Our "little animal brains" have ~86 billion neurons (source: brainfacts.org/in-the-lab/mee…). So not a different scale (and with much more complexity).

/11

More to the point: None of this is inevitable. DL systems aren't naturally occurring phenomena that we can try to understand or just stand in awe of. They are things we are building & choosing to use. We can choose not to, at least w/p sufficient testing for each use case.

/12

Also, because it feels gross to compare language model parameters to human neurons, I want to plug again this great article by @AlexBaria and @doctabarz on the computational metaphor.

arxiv.org/abs/2107.14042

/13

Back to Marche: I don't think we should necessarily believe the people who got super rich off of surveillance capitalism when they say "oh noes, can't regulate, it would stop the development of the technology".

/14

Again, whether or not we try to build this (and with what regulatory guardrails) is a CHOICE. But also: it would be pretty easy with today's stochastic parrots to sometimes at least get an answer like that. (While other times getting hate speech...)

/15

Uh, just because you put these things in a list does not make them all the same kind of thing ("language game").

/16

Yeah, just because the people who built the thing say it does something "in a ways that's not dissimilar from the way you and I do" doesn't make it true. Do they have the expertise to evaluate that? How did they evaluate that?

/17

Oh, and again, while "contemporary NLP" does use neural LMs for a lot of things, I wouldn't say it "derives" from them. There is more to the field than just throwing neural nets are poorly conceived tasks.

/18

What comes next is some GPT-3 authored additional hype, stating with the prompt "And if AI harnesses the power promised by quantum computing," Marche does acknowledge it (in the following paragraph). He is also responsible for deciding to include it.

/19

(Note that Marche also doesn't tell us how many tries he took to get the one he chose to include.)

/20

It's not doing any of these things, actually. Having synthetic text in the style of someone who has died is not bringing them back from the dead. I'm not sure what an "imitation" of consciousness is, nor how it would benefit us.

/21

And it is certainly not "piercing the heart of how language works between people".

On how LM-geneated text is nothing like human linguistic behavior, see Bender & Koller 2020 and also this episode of Factually!

aclweb.org/anthology/2020…
earwolf.com/episode/the-re…

/22

And one last screencap before I end. Where is the evidence for any of these claims? None is provided.

/23

So, I hope that was enjoyable and/or informative. I give this one #threemarvins. Could 2022 be the year of peak #AIhype? That sure would be nice. 24/24

Postscript 1: Important additional info on the (sigh) comparison of 100B parameter networks to human brains

Share this Scrolly Tale with your friends.

A Scrolly Tale is a new way to read Twitter threads with a more visually immersive experience.
Discover more beautiful Scrolly Tales like this.

Keep scrolling