Learn to calculate regression equations and perform hypothesis tests with The Manga Guide to Regression Analysis.
You also learn: simple, multiple, and logistic regression to predict iced tea orders and bakery revenues, and calculate confidence intervals and odds ratios.
The curse of dimensionality is a major roadblock for machine learning practitioners.
But most don't fully understand it.
Don't be left in the dark - join me in this thread as I clarify and demystify this concept 👇🏽🧵
The Curse of Dimensionality (let's just call it "The Curse") refers to problems that occur when you try to use statistical methods in high-dimensional space.
As the number of features (dimensionality) increases, the data becomes relatively more sparse, and often exponentially more samples are needed to make statistically significant predictions.
Feature selection is a crucial part of building a good machine learning model.
But most data scientists don't think before they select features.
The fact is: feature selection in machine learning is not always necessary.
Here are 5 situation when you don't need it 👇🏽🧵
1. You have a small dataset that doesn't have many features.
If the data you're using is small and doesn't have many features, you don't need to do feature selection.
2. The features are already carefully selected
If the features you're using have already been carefully chosen and are important for the task you are trying to do, you don't need to do feature selection.