Delip Rao e/σ Profile picture
Sep 6, 2022 25 tweets 8 min read Read on X
Language Models have taken #NLProc by storm. Even if you don’t directly work in NLP, you have likely heard and possibly, used language models. But ever wonder who came up with the term “Language Model”? Recently I went on that quest, and I want to take you along with me. 🧶
I am teaching a graduate-level course on language models and transformers at @ucsc this Winter, and out of curiosity, I wanted to find out who coined the term “Language Model”.
First, I was a bit ashamed I did not know this fact after all these years in NLP. Surely, this should be in any NLP textbook, right? Wrong! I checked every NLP textbook I could get my hands on, and all of them define what an LM is, without giving any provenance to the term.
Here's an example from the 3rd edition of @jurafsky and Martin. Other books do it similarly.
I also quickly skimmed some popular NLP courses, including ones from Johns Hopkins, Stanford, Berkeley, etc., and they don't cover it either. You might be thinking that this detail is not important and thereby justifying its omission.
Normally, I would agree with you that there's no point learning random historical facts for their sake unless the provenance shows us to understand better what's being examined. So on we go to the quest! 🐈
Now, everyone knows Shannon worked out the entropy of the English language as a means to his end -- developing a theory of optimal communication.
His 1948 paper is very readable, but there is no mention of the term "Language Model" in it or in any of his follow-up works. people.math.harvard.edu/~ctm/home/text…
The next allusion to a language model was in 1958 by @ChomskyDotInfo in his famous paper on the "three models for the description of language", where he calls them "Finite State Markov Processes".
Chomsky got too caught up with grammar and did not think much of these finite state Markov Processes, and neither did he use the term "Language Model" for them. chomsky.info/wp-content/upl…
I also knew @jhuclsp had a famous 1995 workshop on adding syntax in language models ("LM95"). Going through the proceedings, it appears like the term Language Model was very operational by then. So, the term must have originated somewhere between 1958 and 1995.
After some digging, I have finally landed on this 1976 paper: citeseerx.ist.psu.edu/viewdoc/downlo…
I feel quite confident that this is the first source, and Fred Jelinek was likely the first person to use the term "language model" in the scientific literature; it happens specifically in this paragraph where the term shows up in italics:
The paper itself is a landmark paper. It first described modern pre-deep learning ASR engines to the world, and the architecture described there is still used in countless deployments even today. Fred lays out the architecture as follows:
The “Linguistic Decoder” takes a phone sequence and returns scored sequences of words.
Later in the text, Fred clearly points out that the language model (used for the first time, in italics) is just one of the ways to do this linguistic decoding.
Footnote #17 is a dig at Chomsky, where Fred is basically saying, “don’t read too much into LMs. They just define the probability of a sequence. There’s nothing linguistic about it.”
This footnote is actually quite important. Today, with large language models sweeping NLP tasks across the board with their performance, it might be understandable to think LMs "understand" language.
@Smerity dug out this gem for me from 5 years ago (yes, I have friends like that) where I contradict that. Is he right?
If you are in the Fred Jelinek camp, Language Modeling per se is not an NLP problem. That LMs are used in solving NLP problems is a consequence of the language modeling objective. Much like the combustion engine science has nothing to do with flying, but jet engines use them.
In this world of neural LMs, I would define Language Models as intelligent representations of language. I use “intelligence” here in a @SchmidhuberAI-style compression-is-intelligence manner. arxiv.org/abs/0812.4360
Pretext tasks, like the MLM task, capture various aspects of language, and reducing the cross-entropy loss is equivalent to maximizing compression, i.e., “intelligence”.
So to land this thread, knowing where terms come from matters. It allows us to understand precisely when multiple interpretations are possible from the surface meaning of the term. Next time you want to cite something canonical for LMs, don’t forget (Jelinek, 1978)! 🧶🐈
PS: More AI folks should know about Fred Jelinek. He's a giant on whose shoulders many of us stand. web.archive.org/web/2016122819…
Correction: one of my earlier tweets refers to the paper in question as (Jelinek, 1978), but as evident from the bibtex entry, it’s (Jelinek, 1976).

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Delip Rao e/σ

Delip Rao e/σ Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @deliprao

Jul 9
Since many have asked, I am posting a detailed initial review in this 🧶
The unboxing experience is unlike anything else. The reader itself is in this pillow-like case. Dosa, the hoarder of pillows, wanted it for herself! 😆🐶

So, if you use the case it comes with, holding daylight in your arms feels like hugging a small pillow.
Image
Image
The device has a comfortable heft and thickness to it. It is not so thin like Remarkable or Kindle that it feels flimsy to hold, but also not heavy. The finish on the device feels like something from an experienced design/manufacturing house. A physical back button is very nice!
Image
Image
Read 15 tweets
Mar 27
I have long maintained LLMs make the poor performers mediocre, the average slightly above average, but do not change, and maybe hinder, the performance of top performers.

Here’s a result from a university-level physics coding task.

arxiv.org/abs/2403.16977
Image
An important point not to be missed is that mixed-use students don't necessarily have a gain over no-LLM students on a sufficiently challenging task with reasonably competitive humans.
Often, when top writers/coders complain they are not getting much out of LLMs, folks like to pile on to tell them they are not using LLMs correctly. But people speak from their realities without understanding others.
Read 11 tweets
Dec 13, 2023
I have been testing mistral-medium & GPT-4’s code generation abilities for non-trivial problems. These are problems even experience engineers will take time to work it out. I am summarizing some examples and overall impression in this thread: 🧶
My high level summary is @MistralAI 1) always does the job, 2) doesn’t waste output tokens on verbose explanatory output, and 3) the suggestions it offers are concrete.

Examples 👇
Mistral-Medium: “Write cuda-optimized code for generating a PyTorch Dataset of Fibonacci primes”
Verdict: No-nonsense, full code, ✅

Image
Image
Image
Read 10 tweets
Dec 6, 2023
NEWS: Apple just entered the AI open source arena by quietly releasing their new DL framework called MLX! It runs code natively on Apple Silicon with a single pip install and no other dependencies.

Sharing what I discovered from this initial release: github.com/ml-explore/mlx
Image
It seems to follow the PyTorch API closely and provides many useful primitives right out of the box. For example, implementing a decoder-only transformer is as simple as this! Image
Notable is many latest discoveries are offered as primitives in the lib. For instance, if you have to implement rotary position encoding in PyTorch, you will need to roll it out on your own or depend on lucidrains. Image
Read 14 tweets
Jul 21, 2023
This is another one of those ill-thought, fear-mongering scientific disinformation about LLMs, and I will explain why in this long thread. 🧶
Before you think this is some influencer^ thread sh*t, I worked explicitly on social media disinformation campaigns during the 2016 elections and have worked on countermeasures.

^ also, I'm not an influencer for anything or anyone, AFAIK
I have built products, published research, presented to the National Academy of Sciences on detection of generated content, and co-organized an ICML workshop on synthetic media.
Read 25 tweets
Mar 1, 2023
OpenAI released their ChatGPT API today. Here’s a deep dive:

1. It’s not only a new model, but a new endpoint. Notice the model name says, “gpt-3.5-turbo”.

Turbo model is something the paid ChatGPT users (“PLUS”) got a preview a week or so ago.
2. what makes it “turbo” or fast is still TBA but my bet is it’s some kind of mixture-of-experts setup, besides other systems optimizations.
3. This also reduces the cost of serving and that explains why ChatGPT API is now 1/10th the price of davinci APIs. But I also feel this could be partly a strategy decision to undercut competition while getting an opportunity to observe usage patterns and improve models.
Read 23 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(