Why is AI bad at math? πŸ“

Machine learning models today are good at generating realistic-looking text (see GPT-3), images (VQGAN+CLIP), or even code (GitHub Co-Pilot/Codex).

However, these models only learn to imitate, so the results often contain logical errors.

Thread πŸ‘‡
Simple math problems, like the ones 10-year-old kids solve, usually require several logical steps involving simple arithmetics.

The problem is that, if the ML model makes a logical mistake anywhere along the way, it will not be able to recover the correct answer.

πŸ‘‡
@OpenAI is now working on tackling this issue.

In their latest paper, they introduce the so-called verifiers. The generative model generates 100 solutions, but the verifiers select the one that has the highest chance of being factually correct.

openai.com/blog/grade-sch…

πŸ‘‡
This strategy helps them get much better at solving simple math problems - almost on par with kids aged 9-12. However, they still achieve only 55% correct answers, so there is still some way to go.

It is an interesting research field though, so great to see progress there.

πŸ‘‡
Thanks to @lacker for running and documenting a series of interesting experimets with GPT-3. The example in the first tweet is taken from there. Check all of them in this blog post:

lacker.io/ai/2020/07/06/…
Yes, this is a good point! AI is not bad at math, language models are still bad at math.

If you converted these problems to mathematical notation, they would be trivial to solve by a computer without any AI. It is how you get there...

I mostly agree with this and I really like your scrabble example. This perfectly illustrates the point you are trying to make!

And I agree that AI is still far away from human intelligence.

That being said, I think you underestimate its abilities! πŸ‘‡

It's true that language models are trained to imitate human written text and that's why you see these stupid mistakes.

However, the fact that the same model can be used to assess if a statement is true or not shows that there is more to that than just imitation! πŸ‘‡
The English scrabble player is able to imitate French scrabble by remembering the words, but he wasn't able to assess if a particular sentence makes sense, right?

πŸ‘‡
And the human thought process for complicated tasks is somewhat similar. You play around with different possible solutions in your head and assess them if they are real solutions.

Or like brainstorming - people throw ideas around and discuss and assess them. πŸ‘‡
And maybe AI learns in a different way than humans, but we also learn in different ways. Imagine learning a scientific formula.

One person may learn it by hard, while another one may learn how it is derived and not remember the formula itself by hard.
A third person may remember it by some analogy with a formula in another field.

So, AI may find different ways to "learn" things, that are not like the human ways, but are not less effective. I agree we are not there, though...
And last tweet - if you are interesting in this topic I recommend this podcast on what intelligence means!

β€’ β€’ β€’

Missing some Tweet in this thread? You can try to force a refresh
γ€€

Keep Current with Vladimir Haltakov

Vladimir Haltakov Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @haltakov

9 Nov
How I made $3000 in 3 weeks selling AI-generated art? πŸ’°

Last week I showed you how you can use VQGAN+CLIP to generate interesting images based on text prompts.

Now, I'll tell you how I sold some of these as NFTs for more than $3000 in less than 3 weeks.

Let's go πŸ‘‡ Image
Background

I've been interested in NFTs for 2 months now and one collection I find interesting is @cryptoadzNFT. What's special about it is that the creator @supergremplin published all of the art in the public domain. This spurred the creation of many derivative projects.

πŸ‘‡ Image
The Idea πŸ’‘

My idea was to use VQGAN+CLIP to create interesting versions of the CrypToadz. So, I started experimenting with my own toad #6741.

I took the original NFT image as a start and experimented a lot with different text prompts. The results were very promising!

πŸ‘‡ Image
Read 20 tweets
3 Nov
How to create art with Machine Learning? 🎨

You've probably seen these strangely beautiful AI-generated images on Twitter. Have you wondered how they are created?

In this thread, I'll tell you about a method for generating art with ML known as VQGAN+CLIP.

Let's jump in πŸ‘‡
Short History πŸ“œ

In January @OpenAI publicly released CLIP, which is a model that allows matching text to images.

Just days after that, some people like @advadnoun, @RiversHaveWings, and @quasimondo started experimenting using CLIP to guide the output of a GAN using text.

πŸ‘‡
OpenAI published an image generation model together with CLIP, called DALL-E, but without the full code and the pre-trained models.

The results from guiding StyleGAN2 or BigGAN with CLIP aren't as accurate as DALL-E, but they are weirdly artistic.



πŸ‘‡
Read 18 tweets
2 Nov
Creators only get badges πŸ…

There is a problem with how value is distributed in online communities today. It seems we take the status quo for granted and don't discuss it much.

The people that create most of the value, get none of the money! Only badges...

Thread πŸ‘‡
Online communities

I'm talking about platforms like Twitter, Reddit, Stack Overflow etc. They're wonderful places, where you can discuss interesting topics, get help with a problem, or read the latest news.

However, the people that make them truly valuable receive nothing πŸ‘‡
It usually looks like this:

β–ͺ️ Company creates a web 2.0 platform
β–ͺ️ Users create content and increase the value
β–ͺ️ Company aggregates the demand
β–ͺ️ Company monetizes with ads and subscriptions
β–ͺ️ Company gets lots of money
β–ͺ️ Creators get badges, karma and virtual gold

πŸ‘‡ Image
Read 25 tweets
13 Oct
Machine Learning Formulas Explained! πŸ‘¨β€πŸ«

This is the formula for the Binary Cross Entropy Loss. This loss function is commonly used for binary classification problems.

It may look super confusing, but I promise you that it is actually quite simple!

Let's go step by step πŸ‘‡
The Cross-Entropy Loss function is one of the most used losses for classification problems. It tells us how well a machine learning model classifies a dataset compared to the ground truth labels.

The Binary Cross-Entropy Loss is a special case when we have only 2 classes.

πŸ‘‡
The most important part to understand is this one - this is the core of the whole formula!

Here, Y denotes the ground-truth label, while ΕΆ is the predicted probability of the classifier.

Let's look at a simple example before we talk about the logarithm... πŸ‘‡
Read 15 tweets
21 Sep
There are two problems with ROC curves

❌ They don't work for imbalanced datasets
❌ They don't work for object detection problems

So what do we do to evaluate our machine learning models properly in these cases?

We use a Precision-Recall curve.

Another one of my threads πŸ‘‡
Last week I wrote another detailed thread on ROC curves. I recommend that you read it first if you don't know what they are.



Then go on πŸ‘‡
❌ Problem 1 - Imbalanced Data

ROC curves measure the True Positive Rate (also known as Accuracy). So, if you have an imbalanced dataset, the ROC curve will not tell you if your classifier completely ignores the underrepresented class.

More details:

πŸ‘‡
Read 19 tweets
20 Sep
How to spot fake images of faces generated by a GAN? Look at the eyes! πŸ‘οΈ

This is an interesting paper that shows how fake images of faces can be easily detected by looking at the shape of the pupil.

The pupils in GAN-generated images are usually not round - see the image!

πŸ‘‡
Here is the actual paper. The authors propose a way to automatically identify fake images by analyzing the pupil's shape.

arxiv.org/abs/2109.00162
The bad thing is, GANs will probably quickly catch up and include an additional constraint for pupils to be round...
Read 5 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!

Follow Us on Twitter!

:(