Notes: Looking at the graph, the majority of Kagglers do not track their ML models. All eye on the leaderboard!
AutoML Tools - Top 5
1. Google Cloud AutoML 2. Azure Automated ML 3. Amazon SageMaker Autopilot 4. H20 Driverless AI 5. Databricks AutoML
CONCLUSIONS:
1. Notebooks are still the most appreciated way of experimenting with ML. If you never did it, try them in VSCode. 2. Scikit-Learn is ahead of the game 3. All you need is XGBoost(CC: @tunguz) 4. No need for model tracking on Kaggle. There is a leaderboard
If you would like to read the whole survey, here is the link:
GPT-4 is finally out. As many people alluded to lately, GPT-4 is multimodal. It can take images and texts. It can answer questions about images, converse back & forth, etc. GPT-4 is essentially better GPT-3.5 + vision.
GPT-4 exhibits human-level performance on a range of language and vision academic benchmarks and shows excellent performance in professional exams.
Below is GPT-4 performance compared to GPT-3.5 & GPT-4(no vision)>> vision improves language!?
GPT-4 demonstrates remarkable few-shot(3-shots) accuracy on MMLU(Multi‑task Language Understanding). In particular, GPT-4 results on low-resourceful languages are very impressive(some better than GPT-3.5 on English).
The world is interconnected and graphs are everywhere. Deep learning is increasingly being applied in processing and analyzing graphs and alike datasets.
Here are some of the best resources for people interested in graph machine learning ⬇️⬇️⬇️
Geometric Deep Learning - AMMI
This is one of the best graph ML courses that cover a wide range of topics. Its 2nd version was recently released. Taught by M. Bronstein, P. Veličković, T. Cohen, and J. Bruna, all of whom are seminal figures in the field.
Most vision transformer-based backbones such as Swin Transformer are hierarchical(channels & resolutions increase and decrease with depth respectively), they mimic ConvNets.
ViTDet takes a different approach: keep the isotropic structure of ViT and add a simple feature pyramid.
The 2 main highlights of ViTDet:
- Uses window self-attention but stays away from shifted-windows
- Separates pretraining & finetuning. This feat can make it possible to use self-supervised pretraining. The best results are indeed achieved by MAE(masked autoencoder) pretraining.
There have been many attempts to improve the efficiency of Transformer which resulted in zillions of xformers but, most people still use vanilla Transformer.
Here is a comprehensive walkthrough of efficient Transformers
A standard Transformer works great, but its core component(self-attention) has quadratic time complexity, something that is okay for short sequences(in tasks like NMT) but not good for large sequences.
The main point of improving Transfomer is thus reducing such time complexity.
Despite having many xformers that claim to have a linear time complexity(or close), "it's still a misery to know which fundamental efficient Transformer block one should consider using."
Indeed, most language & vision applications are still dominated by standard Transformer.
Things are changing so fast. Long ago, to classify images, you typically had to extract features using traditional image processing techniques, and then feed them to a linear/SVM classifier.
The progress and ongoing unification of visual recognition architectures 🧵🧵
To detect objects, you had to generate thousands of object regions using region proposals and then have an SMV classifier & regressor that classify & predict box coordinates respectively.
After the 2012 AlexNet moment, things started to change. All you needed to do image classification was ConvNets. No more traditional feature engineering techniques.
MLOps or Machine Leaning Operations is one of the most important things to be studying these days. Building models is one thing. Making models useful is another thing.
The reason why you need to study MLOps and the 4 learning resources that you will ever need 🧵🧵
Today, we can all agree that model building is no longer the most challenging part of a machine learning project. Over the past decades, there has been massive development in models.
In most problems, it's not even required to build your own models nor recommended.
If models are no longer the bottleneck, what's important then and what does this have to do with MLOps that we started this thread talking about?
Without bells and whistles, let's see why MLOps is important and the resources that you can use to learn it.