Notes: Looking at the graph, the majority of Kagglers do not track their ML models. All eye on the leaderboard!
AutoML Tools - Top 5
1. Google Cloud AutoML 2. Azure Automated ML 3. Amazon SageMaker Autopilot 4. H20 Driverless AI 5. Databricks AutoML
CONCLUSIONS:
1. Notebooks are still the most appreciated way of experimenting with ML. If you never did it, try them in VSCode. 2. Scikit-Learn is ahead of the game 3. All you need is XGBoost(CC: @tunguz) 4. No need for model tracking on Kaggle. There is a leaderboard
If you would like to read the whole survey, here is the link:
The machine learning research community is very and very vibrant.
Here is what I mean...🧵🧵
In 1958, Frank Rosenblatt invented a perceptron, a very simple algorithm that would later turn out to be the core and origin of to days intelligent machines.
In essence, the perceptron is a simple binary classifier that can determine whether or not a given input belongs to a specific class.
Precision: What is the percentage of positive predictions that are actually positive?
Recall: What is the percentage of actual positives that were predicted correctly?
The fewer false positives, the higher the precision. Vice-versa.
The fewer false negatives, the higher the recall. Vice-versa.
How do you increase precision? Reduce false positives.
It can depend on the problem, but generally, that might mean fixing the labels of those negative samples(being predicted as positives) or adding more of them in the training data.
It starts with a high-level overview of the model/technique being covered and then continues with the implementation.
And wherever possible, there are visuals to support the concepts.
Here is an outline of what you will find there:
PART 1 - Intro to Programming and Working with Data
â—†Intro to Python for Machine Learning
â—†Data Computation With NumPy
â—†Data Manipulation with Pandas
â—†Data Visualization
â—†Real EDA and Data Preparation