Language models are some of the most interesting and most promising research topics in AI. After all, being able to communicate with humans naturally has long been considered *the* ultimate goal for AI (Turing test).

arxiv.org/abs/2111.09509

#NLP #AI #ML #DS 1/3
However, even though large language models in particular are very powerful at generating new text, it is still an ongoing source of debate of how much of that ability is just "rote memorization", and how much is rooted in genuinely fundamental language understanding. 2/3
The interesting paper above tries to answer some of those questions. It would seem that the language models are quite capable of coming up with genuinely novel texts, especially for larger paragraphs, but they still seem to lack the basic semantic understanding of language. 3/3

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Bojan Tunguz

Bojan Tunguz Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @tunguz

16 Oct
1/ After a year of work, our paper on mRNA Degradation is finally out!

paper: arxiv.org/abs/2110.07531
code: github.com/eternagame/Kag…
2/ A year ago I was approached with a unique and exciting opportunity: I was asked to help out with setting a Kaggle Open Vaccine competition, where the goal would be to come up with a Machine Learning model for the stability of RNA molecules.
3/ This is of a pressing importance for the development of the mRNA vaccines. The task seemed a bit daunting, since I have had no prior experience with RNA or Biophysics, but wanted to help out any way I could.
Read 8 tweets
18 Dec 20
One of the unfortunate consequences of Kaggle's inability to host tabular data competitions any more will be that the fine art of feature engineering will slowly fade away. Feature engineering is rarely, if ever, covered in ML courses and textbooks. 1/
There is very little formal research on it, especially on how to come up with domain-specific nontrivial features. These features are often far more important for all aspects of the modeling pipeline than improved algorithms. 2/
I certainly would have never realized any of this were it not for tabular Kaggle competitions. There, over many years, a community treasure trove of incredible tricks and insights had accumulated. Most of them unique. 3/
Read 5 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Thank you for your support!

Follow Us on Twitter!

:(