No, I will not make any such change, for two reasons:
1. Process: The camera ready is done, and approved by all of the authors. If I make any changes past this point it will be literally only fixing typos/citations. No changes to content let alone the title.
2. Content: I stand by the title and the question we are asking. The question is motivated because the field has been dominated by "bigger bigger bigger!" (yes in terms of both training data and model size), with most* of the discourse only fawning over the results. >>
*(Exceptions to this: (1) the big body of work--including yours--into whether the models absorb bias and (2) the GPT-2 staged roll-out paper (and references cited in its sec 1.)

arxiv.org/abs/1908.09203
Thus, the motivation for writing this paper. We aren't saying "LLMs are bad" but rather: these are the dangers we see, that should be accounted for in risk/benefit analyses and, if research proceeds in this direction, mitigated. >>
Furthermore, I've now had a minute to read your critique, and I disagree with your claim that our criticisms are independent of model size. Difficulty in curating and documenting datasets absolutely scales with dataset size, as we clearly lay out in the paper: >> Screen cap of part of sec 4.4 of the paper, reading: "W
Likewise, nowhere do we say that small LMs are necessarily good/risk-free. There, and in your points about smaller models possibly being less energy efficient, you seem to have bought into a world view where language modeling must necessarily exist and continue. >>
This isn't a presupposition we share. (And please don't misread: I'm not saying it should all stop this instant, but rather that research in this area should include cost/benefit analyses.) >>
As for the claim that our paper is one-sided, this is exhausting. All of ML gets to write papers that talk up the benefits of the tech without mentioning any risks (at least until 2020 w/broader impact statements), but when a paper focuses on the risks, it's "one-sided"? >>
Furthermore, the "debate" you would like us to acknowledge is based on a false premise. As we lay out in detail in Sec 4, the training data emphatically do NOT represent "the world as it is". >>
And lastly, miss me with the claim that our work is "political" and therefore has a responsibility to "present the alternative views".

See also:


>>
We draw on scholarship from a range of fields that looks at understanding how systems of power and oppression work in society. >>
The claim that this kind of scholarship is "political" and "non-scientific" is precisely the kind of gate-keeping move set up to maintain "science" as the domain of people of privilege only. /fin

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Emily M. Bender

Emily M. Bender Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @emilymbender

16 Dec 20
They sure will. Here's a quick analysis of how, from my perspective as someone who studies societal impacts of natural language technology:
First, some guesses about system components, based on current tech: it will include a very large language model (akin to GPT-3) trained on huge amounts of web text, including Reddit and the like.
It will also likely be trained on sample input/output pairs, where they asked crowdworkers to create the bulleted summaries for news articles.
Read 16 tweets
14 Dec 20
To all of those who are "both-sides-ing" this: we see you. It takes courage, guts, and fortitude to speak out in the face of oppression, knowing that no matter how gently you make the point people will think you're "too angry".
I'm glad that #UWAllen publicly disavowed Domingos's meltdown. It's disheartening to see so many folks reacting to that with: well what about @AnimaAnandkumar?
Just how angry is the right amount of angry, when faced with racism, misogyny, misogynoir, gaslighting, etc? Furious is the right amount.
Read 4 tweets
13 Dec 20
"Aside from turning the paper viral, the incident offered a shocking indication of how little Google can tolerate even mild pushback, and how easily it can shed all pretense of scientific independence." Thank you, @mathbabedotorg
@mathbabedotorg Re: "Embarrassing as this episode should be for Google — the company’s CEO has apologized — I’m hoping policy makers grasp the larger lesson."

Totally agree on the main pts (about policy makers and about it being embarrassing). It doesn't seem to me that he actually apologized.
Read 4 tweets
9 Dec 20
I’ve picked up a bunch of new followers in the past few days, and I suspect many of you are here because you’re interested in what I might have to say about Google and Dr. @TimnitGebru. So, here’s what I have to say:
Dr. @TimnitGebru is a truly inspiring scholar and leader. Working with her these past few months has been an absolute highlight for me:
I was super excited to work with her because of all the excellent work she has done. For highlights, see this thread compiled by @math_rachel:
Read 23 tweets
30 Oct 20
I was feeling a little ranty this morning, but there's actually also some interesting points about context and pragmatics here, for when we write (or cause machines to write) text that will be interpreted in contexts we aren't directly participating in:
Surely, from the platform's point of view, WeCNLP is starting at 7am. For them, WeCNLP refers to an event with "start" and "end" times they have to program into their platform, so that people who have registered can access the platform during those times. >>
But for people *attending* WeCNLP (the addressees of that email), WeCNLP refers to an event with a specific internal schedule, of talks and informal meeting times. >>
Read 7 tweets
30 Oct 20
Is anyone else on the West Coast already up and surprised to see an email from #WeCNLP2020 saying the event starts in "59 minutes" when the schedule says 8am start?
Seems like an auto-generated system from the online platform because the site is opening at 7, though the program doesn't start until 8.
Given that this one is WEST COAST NLP and for once is actually in our timezone, it would be nice to not be harassed by emails making us feel late...
Read 4 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!

Follow Us on Twitter!