Santiago Profile picture
17 Jan, 12 tweets, 2 min read
20 machine learning questions that will make you think.

(Cool questions. Not the regular, introductory stuff that you find everywhere.)

πŸ§΅πŸ‘‡
1. Why is it important to introduce non-linearities in a neural network?

2. What are the differences between a multi-class classification problem and a multi-label classification problem?

3. Why does the use of Dropout work as a regularizer?

πŸ‘‡
4. Why you shouldn't use a softmax output activation function in a multi-label classification problem when using a one-hot-encoded target?

5. Does the use of Dropout in your model slow down or speed up the training process? Why?

πŸ‘‡
6. In a Linear or Logistic Regression problem, do all Gradient Descent algorithms lead to the same model, provided you let them run long enough?

7. Explain the difference between Batch Gradient Descent, Stochastic Gradient Descent, and Mini-batch Gradient Descent.

πŸ‘‡
8. What are the advantages of Convolution Neural Networks (CNN) over a fully connected network for image classification?

9. What are the advantages of Recurrent Neural Networks (RNN) over a fully connected network when working with text data?

πŸ‘‡
10. How do you deal with the vanishing gradient problem?

11. How do you deal with the exploding gradient problem?

12. Are feature engineering and feature extraction still needed when applying Deep Learning?

13. How does Batch Normalization help?

πŸ‘‡
14. The universal approximation theorem shows that any function can be approximated as closely as needed using a single nonlinearity. Then why do we use more?

15. What are some of the limitations of Deep Learning?

16. Is there any value in weight initialization?

πŸ‘‡
17. The training loss of your model is high and almost equal to the validation loss. What does it mean? What can you do?

18. Assuming you are using Batch Gradient Descent, what advantage would you get from shuffling your training dataset?

πŸ‘‡
19. Compare the following evaluation protocols: a hold-out validation set, K-fold cross-validation, and iterated K-fold validation. When would you use each one?

20. What are the main differences between Adam and the Gradient Descent optimization algorithms?

πŸ‘‡
I'll be posting an in-depth answer for each one of these questions over the next following days.

Stay tuned and help me spread more machine learning content to more people in the community!
Feel free to post answers for those questions that you know. Give them a try!

You’ll be forced to think about them just by trying to collect your thoughts to put them here.

(There’s nothing as effective to learn than interacting with others.)

β€’ β€’ β€’

Missing some Tweet in this thread? You can try to force a refresh
γ€€

Keep Current with Santiago

Santiago Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @svpino

18 Jan
Why is it important to introduce non-linearities in a neural network?

The short answer: So we can solve more interesting problems.

The left image shows a classification problem that can be solved using a single dividing line. The image on the right is much more complex.

πŸ§΅πŸ‘‡ ImageImage
Here is a neural network with 2 hidden layers of 4 neurons. The activation is set to "Linear."

In just a few epochs, the network finds the correct solution.

Notice how the network uses a single dividing line in the output. That's all it can do with linear activations.

πŸ‘‡ Image
If we try the same network on the more complex problem, it will struggle to classify the data correctly.

We haven't introduced non-linearities in this network, so it won't find the proper solution for this type of problem.

πŸ‘‡ Image
Read 7 tweets
18 Jan
Here are the classes I took and the money I paid to get my Master's from Georgia Tech with a specialization in machine learning:

πŸ§΅πŸ‘‡
The classes:

1. Machine Learning
2. Computer Vision
3. Reinforcement Learning
4. Intro to Graduate Algorithms
5. Machine Learning for Trading
6. Database Systems Concepts and Design
7. Software Development Process
8. Software Architecture and Design

πŸ‘‡
9. Human-Computer Interaction
10. Advanced Operating Systems
11. Software Analysis and Testing

You only need 30 credits to graduate. I completed 33.

It took me 4 years to go through all the classes (2015-2019). I was 35 when I started.

πŸ‘‡
Read 6 tweets
16 Jan
Ready to take your machine learning models into production?

@awscloud offers SageMaker, a fully managed machine learning service that acts as a one-stop-shop for everything you need.

The list of services is impressive:

πŸ§΅πŸ‘‡
1. Studio
2. Projects
3. Studio Notebooks
4. Experiments
5. Debugger
6. Model Registry
7. Model Building Pipelines
8. Model Monitor
9. Lineage Tracking
10. Feature Store
11. Data Wrangler
12. Preprocessing
13. Batch Transform
14. Ground Truth
15. Augmented AI
16. Edge Manager

πŸ‘‡
17. Autopilot
18. Neo
19. Elastic Inference
20. Reinforcement Learning
21. JumpStart
22. Clarify

I've been working for years with SageMaker, and the services are incredibly comprehensive. Whatever I need to do, I can find.

(Plus, you can combine them with the rest of AWS!)

πŸ‘‡
Read 4 tweets
14 Jan
10 machine learning YouTube videos.

On libraries, algorithms, and tools.

(If you want to start with machine learning, having a comprehensive set of hands-on tutorials you can always refer to is fundamental.)

πŸ§΅πŸ‘‡
1⃣ Notebooks are a fantastic way to code, experiment, and communicate your results.

Take a look at @CoreyMSchafer's fantastic 30-minute tutorial on Jupyter Notebooks.

2⃣ The Pandas library is the gold-standard to manipulate structured data.

Check out @joejamesusa's "Pandas Tutorial. Intro to DataFrames."

Read 12 tweets
12 Jan
Here is an interesting problem:

You trained a model to classify pictures of 100 different animal species. It does a good job at it.

But when you show it a picture with a species that wasn't part of the training set, the results are obviously wrong.

How do you work around this?
This is also known as a "negative" class, and it helps with this problem, assuming you are capable of collecting images from unknown objects.

I've also found the advantages of this negative class to diminish as more random objects are thrown in there.

It turns out that knowing what you don't know is a tough problem to solve in machine learning.

You'd expect the confidence score returned by the model to be very low for unknown objects. This is, unfortunately, not necessarily the case.

Read 7 tweets
11 Jan
10 fundamental practices that will improve your career in tech.

πŸ§΅πŸ‘‡
[1] Understand the power of "good enough."

A working, good enough solution is usually better than a non-existent perfect solution.

Learn to balance constraints. Know where and when to compromise and when to say "enough."
[2] If you get stuck, ask for help.

Don't spin your wheels indefinitely, trying to solve a problem that can be easily solved by someone else.

Know when you should keep trying and when to stop and ask.
Read 11 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!

Follow Us on Twitter!