Kawin Ethayarajh Profile picture
Jul 12, 2020 6 tweets 6 min read Read on X
Inspired by @yoavgo 's poll, I looked at the views for papers in three tracks -- Ethics, Summarization, and Theme (69 papers in total).

The median views per paper was 104.

In these three tracks, the most-viewed papers at time of writing are ... Image
1. Climbing towards NLU: On Meaning, Form, and Understanding in the Age of Data by @emilymbender and @alkoller (961 views)

2. How Can We Accelerate Progress Towards Human-like Linguistic Generalization? by @tallinzen (410 views)
3. The Unstoppable Rise of Computational Linguistics in Deep Learning by @JamieBHenderson (356 views)

4. (Re)construing Meaning in NLP by @Sean_Trott @TorrentTiago @nancy_c_chang @complingy (291 views)
5. Language (Technology) is Power: A Critical Survey of "Bias" in NLP @sulin_blodgett @haldaume3 @s010n @hannawallach (283 views)

6. To Test Machine Comprehension, Start by Defining Comprehension @jdunietz @GregHBurnham Bharadwaj, @OwenRambow @jchucarroll Ferrucci (239 views)
7. Social Bias Frames: Reasoning about Social and Power Implications of Language by @MaartenSap Saadia Gabriel, @Lianhuiq @jurafsky @nlpnoah @YejinChoinka (227 views)

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Kawin Ethayarajh

Kawin Ethayarajh Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @ethayarajh

Dec 14
As an AI faculty now surrounded by economists at Booth, the discussion on scaling and the Lucas critique has been entertaining to watch. IMO the spirit of @arpitrage's claim feels right but double descent isn't the best analogy. Instead, we should look to language understanding: Image
For a long time, our models of language were very structured. You would create trees of sentences based on their grammatical structure (e.g., dependency parsing), catalog all the different senses of words (e.g,. WordNet), etc. We did this for three reasons:

1) Language has a lot of known structure; it seems wasteful to throw our understanding away.

2) We didn't have web-scale data to learn from.

3) These structured models were useful, and we seemed to be making progress, albeit slowly.Image
Then deep learning (and later scaling) came along. The objections were similar to what we see in the replies to @arpitrage: "Language is too complex -- you'll overfit!" "The use of language itself changes how language will be used!" "There's no way this will generalize OOD!"

The scaling counter to this: _make everything in-distribution_. And it worked! Anytime people came up to some objection to why scaling language models wouldn't work --- some unlearned grammatical construction, difficulty with understanding negation, etc --- all of that disappeared with a model release a few months later. I myself was a skeptic that scaling could work this well, but it did.

Now does this mean LLMs perfectly handle OOD data? Of course not, but so much of text is now in-distribution for them that for almost everything we care about in practice, they handle it well. Moreover, you get some emergent properties like in-context learning where you no longer even need to update the weights of the model for it to learn a task.
Read 5 tweets
Dec 7, 2023
📢The problem in model alignment no one talks about — the need for preference data, which costs $$$ and time!

Enter Kahneman-Tversky Optimization (KTO), which matches or exceeds DPO without paired preferences.

And with it, the largest-ever suite of feedback-aligned LLMs. 🧵 Image
But first, what makes alignment work? Among methods that directly optimize preferences, the majority of gains <30B come from SFT.

Even a dummy one-step PPO that uses +1/-1 rewards works very well.

DPO is uniquely good at the 30B scale, however. 2/ Image
But *why* do they work?

We find that alignment methods impute a utility function to humans.

These imputed functions have many qualities of those empirically derived by Kahneman & Tversky in their Nobel Prize-winning work on how humans make decisions about uncertain outcomes. 3/
Read 13 tweets
Feb 22, 2023
📢 Models like #ChatGPT are trained on tons of human feedback. But collecting this costs $$$!

That's why we're releasing the Stanford Human Preferences Dataset (🚢SHP), a collection of 385K *naturally occurring* *collective* human preferences over text.
huggingface.co/datasets/stanf…
Given some context and two possible responses, SHP preferences reflect the helpfulness of one response over another.

The preferences are over responses to questions/instructions in 18 domains, from cooking to legal advice, drawn from Reddit.
They were inferred from the simple observation that if comment A was written after B but has a higher score despite getting less visibility, then ostensibly A > B.

If A was written before B, then we can't conclude this -- the higher score could have come from more visibility!
Read 10 tweets
Jun 5, 2021
Is there a connection between Shapley Values and attention-based explanations in NLP?

Yes! Our #ACL2021NLP paper proves that **attention flows** can be Shapley Value explanations, but regular attention and leave-one-out cannot.



w/ @jurafsky @stanfordnlp arxiv.org/abs/2105.14652
Image
Shapley Values are a solution to the credit assignment problem in cooperative games -- if 10 people work together to win some reward, how can it be equitably distributed?

For this reason, they've become a popular kind of explanation in ML. 2/
Shapley Values have been used to explain the importance of individual features, embeddings, and neurons.

@GhorbaniAmirata and @james_y_zou have even used them to value training data points.

In NLP though, attention-based explanations and leave-one-out still predominate. 3/
Read 8 tweets
Sep 22, 2020
There's been some confusion over what Microsoft's "exclusive license" really means here.

While I can't speak for OpenAI, exclusive licenses generally grant exclusivity *within some specific context*. So no, Microsoft won't be the only one able to use GPT3. That said ...
My guess is that only MS will have access to the underlying model, while everyone will have to go through the API and be at the whims of whatever terms are set by OpenAI.
This is big -- if you build a product on top of GPT3, your ability to scale will depend on OpenAI's willingness to increase your throughput, which in turn will depend on the terms of their agreement with MS. Not a great situation to be in if you're directly competing with MS.
Read 6 tweets
Jun 23, 2020
Is your NLP classifier actually (un)biased? Or is your diagnosis based on too little data?

It might be the latter!

In my #ACL2020 paper, I discuss why we need bigger datasets for conclusively identifying classification bias in NLP.

arxiv.org/abs/2004.12332 1/
Background: Large NLP datasets don't come with annotations for protected attributes (e.g., gender). To test for classification bias, one typically annotates a small sample of data (typically < 5K). WinoBias and WinoGender are great examples of these bias-specific datasets. 2/
Intuitively, the less data we annotate, the less certain we are that our estimate is close to the true bias. But how can we quantify this uncertainty? 3/
Read 8 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(