1. Process: The camera ready is done, and approved by all of the authors. If I make any changes past this point it will be literally only fixing typos/citations. No changes to content let alone the title.
2. Content: I stand by the title and the question we are asking. The question is motivated because the field has been dominated by "bigger bigger bigger!" (yes in terms of both training data and model size), with most* of the discourse only fawning over the results. >>
*(Exceptions to this: (1) the big body of work--including yours--into whether the models absorb bias and (2) the GPT-2 staged roll-out paper (and references cited in its sec 1.)
Thus, the motivation for writing this paper. We aren't saying "LLMs are bad" but rather: these are the dangers we see, that should be accounted for in risk/benefit analyses and, if research proceeds in this direction, mitigated. >>
Furthermore, I've now had a minute to read your critique, and I disagree with your claim that our criticisms are independent of model size. Difficulty in curating and documenting datasets absolutely scales with dataset size, as we clearly lay out in the paper: >>
Likewise, nowhere do we say that small LMs are necessarily good/risk-free. There, and in your points about smaller models possibly being less energy efficient, you seem to have bought into a world view where language modeling must necessarily exist and continue. >>
This isn't a presupposition we share. (And please don't misread: I'm not saying it should all stop this instant, but rather that research in this area should include cost/benefit analyses.) >>
As for the claim that our paper is one-sided, this is exhausting. All of ML gets to write papers that talk up the benefits of the tech without mentioning any risks (at least until 2020 w/broader impact statements), but when a paper focuses on the risks, it's "one-sided"? >>
Furthermore, the "debate" you would like us to acknowledge is based on a false premise. As we lay out in detail in Sec 4, the training data emphatically do NOT represent "the world as it is". >>
And lastly, miss me with the claim that our work is "political" and therefore has a responsibility to "present the alternative views".
We draw on scholarship from a range of fields that looks at understanding how systems of power and oppression work in society. >>
The claim that this kind of scholarship is "political" and "non-scientific" is precisely the kind of gate-keeping move set up to maintain "science" as the domain of people of privilege only. /fin
• • •
Missing some Tweet in this thread? You can try to
force a refresh
First, some guesses about system components, based on current tech: it will include a very large language model (akin to GPT-3) trained on huge amounts of web text, including Reddit and the like.
It will also likely be trained on sample input/output pairs, where they asked crowdworkers to create the bulleted summaries for news articles.
To all of those who are "both-sides-ing" this: we see you. It takes courage, guts, and fortitude to speak out in the face of oppression, knowing that no matter how gently you make the point people will think you're "too angry".
I'm glad that #UWAllen publicly disavowed Domingos's meltdown. It's disheartening to see so many folks reacting to that with: well what about @AnimaAnandkumar?
Just how angry is the right amount of angry, when faced with racism, misogyny, misogynoir, gaslighting, etc? Furious is the right amount.
"Aside from turning the paper viral, the incident offered a shocking indication of how little Google can tolerate even mild pushback, and how easily it can shed all pretense of scientific independence." Thank you, @mathbabedotorg
@mathbabedotorg Re: "Embarrassing as this episode should be for Google — the company’s CEO has apologized — I’m hoping policy makers grasp the larger lesson."
Totally agree on the main pts (about policy makers and about it being embarrassing). It doesn't seem to me that he actually apologized.
I’ve picked up a bunch of new followers in the past few days, and I suspect many of you are here because you’re interested in what I might have to say about Google and Dr. @TimnitGebru. So, here’s what I have to say:
Dr. @TimnitGebru is a truly inspiring scholar and leader. Working with her these past few months has been an absolute highlight for me:
I was feeling a little ranty this morning, but there's actually also some interesting points about context and pragmatics here, for when we write (or cause machines to write) text that will be interpreted in contexts we aren't directly participating in:
Surely, from the platform's point of view, WeCNLP is starting at 7am. For them, WeCNLP refers to an event with "start" and "end" times they have to program into their platform, so that people who have registered can access the platform during those times. >>
But for people *attending* WeCNLP (the addressees of that email), WeCNLP refers to an event with a specific internal schedule, of talks and informal meeting times. >>
Is anyone else on the West Coast already up and surprised to see an email from #WeCNLP2020 saying the event starts in "59 minutes" when the schedule says 8am start?
Seems like an auto-generated system from the online platform because the site is opening at 7, though the program doesn't start until 8.
Given that this one is WEST COAST NLP and for once is actually in our timezone, it would be nice to not be harassed by emails making us feel late...