We usually talk about two main types of machine learning models:
• A Classification model
• A Regression model
They are different, and it's essential to understand why.
↓ 1/6
Whenever the result of your predictions is categorical, you have a classification model.
For example, when your prediction is a binary value (True or False,) or when you want to predict a specific animal from a picture (Lion, Zebra, Horse.)
↓ 2/6
If the result of your predictions is numerical, you have a regression model.
For example, returning a stock's future price, the value of a house, or tomorrow's temperature.
↓ 3/6
Be aware that "Logistic Regression" is a classification model, not a regression one!
I know. Dumb naming. But it is what it is.
↓ 4/6
A few examples of problems so you can determine whether they are classification or regression:
1. Predicting the age of a person 2. Predicting the nationality of a person 3. Predicting whether an email is spam or not 4. Predicting the total revenue from a product
↓ 5/6
5. Predicting the number of products sold 6. Predicting whether it will rain tomorrow 7. Predicting how many inches will snow tomorrow
Answers: RCCRRCR
6/6
Many people have commented (rightfully so) about my statement saying that "Logistic regression is dumb naming."
Logistic regression is a regression algorithm that's used in classification tasks.
So it's the name really dumb that's the case?
This poses a different question: should we classify methods and techniques by how their work internally or by the task they help solve?
I think the correct answer should be both, right?
• • •
Missing some Tweet in this thread? You can try to
force a refresh
If you are looking for more, this course from Harvard University is an excellent introduction to probability as a language and a set of tools for understanding statistics, science, risk, and randomness.
A lot in machine learning is pretty dry and boring, but understanding how autoencoders work feels different.
This is a thread about autoencoders, things they can do, and a pretty cool example.
↓ 1/10
Autoencoders are lossy data compression algorithms built using neural networks.
A network encodes (compresses) the original input into an intermediate representation, and another network reverses the process to get the same input back.
↓ 2/10
The encoding process "generalizes" the input data.
I like to think of it as the Summarizer in Chief of the network: its entire job is to represent the entire dataset as compactly as possible, so the decoder can do a decent job at reproducing the original data back.
Here are my thoughts about the "HTML is not a programming language" recurrent theme.
↓ 1/5
This question is controversial not because people care about HTML but because it is used as a proxy to classify their worth.
If HTML is not a programming language, then the people working with HTML must not be real programmers, right?
↓ 2/5
This is demeaning and completely unhelpful for those who are starting and looking to find a community.
Instead of drawing lines, we should be welcoming those who want to join us. We need more programmers, coders, developers, or whatever else you want to call them!