@omarsar0 made a great summary about the papers that he recently shared such as a survey of visual transformers, video object segmentation, neural rendering, Graph Neural Networks(GNNs), etc...
I am interested in learning about Geometric Deep Learning. In the near future, I will learn about it and will share it with you along the way...
#2
OpenAI made it easy to get access to GPT-3. In a matter of seconds, I just got this massive language model. If I get time in the coming weeks, I will explore it.
I would like to keep doing weekly highlights to share things that I found helpful over the whole week. For deep dives, I am starting a newsletter in next month.
Let me know if the highlights help you catch up with what's happened in the community.
Until the next week, stay safe!
And make sure you follow @Jeande_d for more machine learning ideas and the latest news.
• • •
Missing some Tweet in this thread? You can try to
force a refresh
Activations functions are one of the most important components of any typical neural network.
What exactly are activation functions, and why do we need to inject them into the neural network?
A thread 🧵🧵
Activations functions are basically mathematical functions that are used to introduce non linearities in the network.
Without an activation function, the neural network would behave like a linear classifier/regressor.
Or simply put, it would only be able to solve linear problems or those kinds of problems where the relationship between input and output can be mapped out easily because input and output change in a proportional manner.
◆3 things from me
◆2 things from other people and
◆2 from the community
🧵🧵
This week, I wrote about what to consider while choosing a machine learning model for a particular problem, early stopping which is one of the powerful regularization techniques, and what to know about the learning rate.
The next is their corresponding threads!
1. What to know about a model selection process...
The below illustration shows early stopping, one of the effective and simplest regularization techniques used in training neural networks.
A thread on the idea behind early stopping, why it works, and why you should always use it...🧵
Usually, during training, the training loss will decrease gradually, and if everything goes well on the validation side, validation loss will decrease too.
When the validation loss hits the local minimum point, it will start to increase again. Which is a signal of overfitting.
How can we stop the training just right before the validation loss rise again? Or before the validation accuracy starts decreasing?
That's the motivation for early stopping.
With early stopping, we can stop the training when there are no improvements in the validation metrics.