Rohan Paul Profile picture
Apr 4 13 tweets 22 min read
2/n
Following tips may boost model performance across different network structures with up to 5% (mAP or mean Average Precision) without increasing computational costs in any way.

#computervision #pytorch #deeplearning #deeplearningai #100daysofmlcode #neuralnetworks #AI
3/n
Visually Coherent Image Mix-up for Object Detection. This has already been proven to be successful in lessening adversarial fears in network classification after testing it on COCO 2017 and PASCAL datasets with YOLOv3 models.
#computervision #pytorch
4/n
Classification Head Label Smoothening

Existing models apply Softmax technique to compute a probability distribution for classes. But there’s a risk of the model becoming too confident in its predictions which can result to over-fitting.
#computervision
5/n
One possible solution to this is to relax our confidence on the labels. For instance, we can slightly lower the loss target values from 1 to, say, 0.9.
#computervision #pytorch #deeplearning #deeplearningai #100daysofmlcode #neuralnetworks #AI #ArtificialIntelligence
6/n
And naturally, we increase the target value of 0 for the others slightly as such. This idea is called label smoothing.

#computervision #pytorch #deeplearning #deeplearningai #100daysofmlcode #neuralnetworks #machinelearning #datascience #pythonprogramming #python #AI
7/n
Data Pre-processing (Mixed Results) like,
• A. Random geometry transformation for random cropping (with constraints), random expansion, random horizontal flip and random resize (with random interpolation).
#computervision #pytorch #deeplearning #deeplearningai
10/n

In model training, step scheduler is the most widely used learning rate schedule. It involves multiplying the learning rate by a constant number below 1 after a number of model iterations.

#computervision #pytorch #deeplearning #deeplearningai #100daysofmlcode
11/n

Synchronized Batch Normalization (SBN)

What is SBN ?

Standard implementations of BN in public frameworks (such as Caffe, MXNet, Torch, TF, PyTorch) are unsynchronized, which means that the data are normalized within each GPU.

#computervision #pytorch #deeplearning
12/n

Therefore the working batch-size of the BN layer is BatchSize/nGPU (batch-size in each GPU).
Hence the solution is, Synchronized Multi GPU BatchNormalization.

#computervision #pytorch #deeplearning #deeplearningai #100daysofmlcode #neuralnetworks

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Rohan Paul

Rohan Paul Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @paulr_rohan

Apr 4
2/16

"roc_auc_score" is defined as the area under the ROC curve, which is the curve having False Positive Rate on the x-axis and True Positive Rate on the y-axis at all classification thresholds.

#DataScience #MachineLearning #DeepLearning #100DaysOfMLCode #Python
Read 17 tweets
Nov 24, 2021
2/n
Types of magic commands

Line magics - starts with % character. Rest of the line is its argument passed without parentheses or quotes.

Cell magics - %% - can operate on multiple lines below their call.
#DataScience #MachineLearning #100DaysOfMLCode #Python #DataScientist
Read 21 tweets
Nov 22, 2021
2/n

GauGAN2 combines segmentation mapping, inpainting and text-to-image generation in a single model

#Computervision #AI #ArtificialIntelligence #TensorFlow #PyTorch #DeepLearning #DataScience #MachineLearning #100DaysOfMLCode #Python #DataScientist #Statistics #Mathematics
3/n

Unlike GauGAN1 the GauGAN2 can translate natural language descriptions into landscape images. Typing a phrase like “sunset at a beach” generates the scene

#Computervision #AI #ArtificialIntelligence #TensorFlow #PyTorch #DeepLearning #DataScience #MachineLearning #Math
Read 7 tweets
Nov 19, 2021
Image interpolation occurs when you resize or distort your image from one pixel grid to another.

1/n

#computervision #IMAGE #DataScience #MachineLearning #DeepLearning #100DaysOfMLCode #Python #programming #Math #Stat #dataviz #DataAnalytics #AI #ArtificialIntelligence #data
Image interpolation works in two directions, and tries to achieve a best approximation of a pixel's intensity based on the values at surrounding pixels.

2/n

#computervision #IMAGE #DataScience #MachineLearning #DeepLearning #100DaysOfMLCode #Python #programming #Math #Stat
Image resizing is necessary when you need to increase or decrease the total number of pixels, whereas remapping can occur when you are correcting for lens distortion or rotating an image.

3/n
#computervision #DataScience #MachineLearning #DeepLearning #100DaysOfMLCode #Python
Read 8 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us on Twitter!

:(