In their latest paper, they introduce the so-called verifiers. The generative model generates 100 solutions, but the verifiers select the one that has the highest chance of being factually correct.
This strategy helps them get much better at solving simple math problems - almost on par with kids aged 9-12. However, they still achieve only 55% correct answers, so there is still some way to go.
It is an interesting research field though, so great to see progress there.
π
Thanks to @lacker for running and documenting a series of interesting experimets with GPT-3. The example in the first tweet is taken from there. Check all of them in this blog post:
It's true that language models are trained to imitate human written text and that's why you see these stupid mistakes.
However, the fact that the same model can be used to assess if a statement is true or not shows that there is more to that than just imitation! π
The English scrabble player is able to imitate French scrabble by remembering the words, but he wasn't able to assess if a particular sentence makes sense, right?
π
And the human thought process for complicated tasks is somewhat similar. You play around with different possible solutions in your head and assess them if they are real solutions.
Or like brainstorming - people throw ideas around and discuss and assess them. π
And maybe AI learns in a different way than humans, but we also learn in different ways. Imagine learning a scientific formula.
One person may learn it by hard, while another one may learn how it is derived and not remember the formula itself by hard.
A third person may remember it by some analogy with a formula in another field.
So, AI may find different ways to "learn" things, that are not like the human ways, but are not less effective. I agree we are not there, though...
And last tweet - if you are interesting in this topic I recommend this podcast on what intelligence means!
β’ β’ β’
Missing some Tweet in this thread? You can try to
force a refresh
How I made $3000 in 3 weeks selling AI-generated art? π°
Last week I showed you how you can use VQGAN+CLIP to generate interesting images based on text prompts.
Now, I'll tell you how I sold some of these as NFTs for more than $3000 in less than 3 weeks.
Let's go π
Background
I've been interested in NFTs for 2 months now and one collection I find interesting is @cryptoadzNFT. What's special about it is that the creator @supergremplin published all of the art in the public domain. This spurred the creation of many derivative projects.
π
The Idea π‘
My idea was to use VQGAN+CLIP to create interesting versions of the CrypToadz. So, I started experimenting with my own toad #6741.
I took the original NFT image as a start and experimented a lot with different text prompts. The results were very promising!
You've probably seen these strangely beautiful AI-generated images on Twitter. Have you wondered how they are created?
In this thread, I'll tell you about a method for generating art with ML known as VQGAN+CLIP.
Let's jump in π
Short History π
In January @OpenAI publicly released CLIP, which is a model that allows matching text to images.
Just days after that, some people like @advadnoun, @RiversHaveWings, and @quasimondo started experimenting using CLIP to guide the output of a GAN using text.
π
OpenAI published an image generation model together with CLIP, called DALL-E, but without the full code and the pre-trained models.
The results from guiding StyleGAN2 or BigGAN with CLIP aren't as accurate as DALL-E, but they are weirdly artistic.
There is a problem with how value is distributed in online communities today. It seems we take the status quo for granted and don't discuss it much.
The people that create most of the value, get none of the money! Only badges...
Thread π
Online communities
I'm talking about platforms like Twitter, Reddit, Stack Overflow etc. They're wonderful places, where you can discuss interesting topics, get help with a problem, or read the latest news.
However, the people that make them truly valuable receive nothing π
It usually looks like this:
βͺοΈ Company creates a web 2.0 platform
βͺοΈ Users create content and increase the value
βͺοΈ Company aggregates the demand
βͺοΈ Company monetizes with ads and subscriptions
βͺοΈ Company gets lots of money
βͺοΈ Creators get badges, karma and virtual gold
This is the formula for the Binary Cross Entropy Loss. This loss function is commonly used for binary classification problems.
It may look super confusing, but I promise you that it is actually quite simple!
Let's go step by step π
The Cross-Entropy Loss function is one of the most used losses for classification problems. It tells us how well a machine learning model classifies a dataset compared to the ground truth labels.
The Binary Cross-Entropy Loss is a special case when we have only 2 classes.
π
The most important part to understand is this one - this is the core of the whole formula!
Here, Y denotes the ground-truth label, while ΕΆ is the predicted probability of the classifier.
Let's look at a simple example before we talk about the logarithm... π
ROC curves measure the True Positive Rate (also known as Accuracy). So, if you have an imbalanced dataset, the ROC curve will not tell you if your classifier completely ignores the underrepresented class.