Precision: What is the percentage of positive predictions that are actually positive?
Recall: What is the percentage of actual positives that were predicted correctly?
🧵🧵
The fewer false positives, the higher the precision. Vice-versa.
The fewer false negatives, the higher the recall. Vice-versa.
How do you increase precision? Reduce false positives.
It can depend on the problem, but generally, that might mean fixing the labels of those negative samples(being predicted as positives) or adding more of them in the training data.
How do you increase recall? Reduce false negatives.
Fix the labels of positive samples that are being classified as negatives when they are not, or add more samples to the training data.
What happens when I increase precision? I will hurt recall.
There is a tradeoff between them. Increasing one can reduce the other.
What does it mean when the precision of your classifier is 1?
False positives are 0.
Your classifier is smart about not classifying negative samples as positives.
What's about recall being 1?
False negatives are 0.
Your classifier is smart about not classifying positive samples as negatives.
What if the precision and recall are both 1? You have a perfect classifier. This is ideal!
What is the better way to know the performance of the classifier without having to balance precision and recall?
Combine them. Find their harmonic mean. If either precision or recall is low, the resulting mean will be low too.
Such harmonic mean is called the F1 Score and it is a reliable metric to use when we are dealing with imbalanced datasets.
If your dataset is balanced(positive samples are equal to negative samples in the training set), ordinary accuracy is enough.
Thanks for reading.
If you found the thread helpful, retweet or share it with people who you think might benefit from seeing it.
Follow @Jeande_d for more machine learning content.
I posted this thread a while back, but I edited it for clarity.
• • •
Missing some Tweet in this thread? You can try to
force a refresh
◆3 things from me
◆2 from from others and
◆1 from the community
This week, I explored different object detection libraries, wrote about the hyper-parameter optimization methods, and updated the introduction to machine learning in my complete ML packaged free online book.
I also reached 6000 followers 🎉. Thank you for your support again!
Activations functions are one of the most important components of any typical neural network.
What exactly are activation functions, and why do we need to inject them into the neural network?
A thread 🧵🧵
Activations functions are basically mathematical functions that are used to introduce non linearities in the network.
Without an activation function, the neural network would behave like a linear classifier/regressor.
Or simply put, it would only be able to solve linear problems or those kinds of problems where the relationship between input and output can be mapped out easily because input and output change in a proportional manner.
◆3 things from me
◆2 things from other people and
◆2 from the community
🧵🧵
This week, I wrote about what to consider while choosing a machine learning model for a particular problem, early stopping which is one of the powerful regularization techniques, and what to know about the learning rate.
The next is their corresponding threads!
1. What to know about a model selection process...