. @OpenAI's #Codex is to programming as Tesla's FSD 2021 is to driving.

Read the paper (esp Appendix B) carefully and you will realize there is a gap between the slick videos & reality: it is often correct on simple tasks, but frequently lost on more complex challenges.

1/3
OpenAI #Codex livestream worked because execs pre-tested & knew what Codex is good for (bite-sized tasks), steering clear of what it is weak at (understanding programming task as a whole).

Codex is amazing, but like current self-driving systems, it may never be reliable.

2/3
What do I think of @OpenAI's new 1st grade math video?

No published results, no peer review. We don't know what the training set is, what the accuracy is, nor how robust results would be to slight changes in wording.

PR ≠ Science

3/3

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Gary Marcus

Gary Marcus Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @GaryMarcus

3 Jul
No, @ylecun, high dimensionality doesn’t erase the critical distinction between interpolation & extrapolation.

Thread on this below, because only way forward for ML is to confront it directly.

To deny it would invite yet another decade of AI we can't trust. (1/9)
Let's start with a simple example drawn from my 2001 book The Algebraic Mind, that anyone can try at home: (2/9)
Train a basic multilayer perceptron on the identity function (ie mulltiplying the input times one) on a random subset of 10% of the the even numbers, from 2 to 1024, representing each number as a standard distributed representation of nodes encoding binary digits. (3/9)
Read 9 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!

Follow Us on Twitter!

:(