Lightning AI ⚡️ Profile picture
May 23 3 tweets 2 min read Twitter logo Read on Twitter
Last week, we discussed techniques to speed up the training speed of large language models🔥💨

How about saving memory during inference? 🧠💾 Check out int8 & int4 quantization, which is supported in Lit-LLama 👉github.com/Lightning-AI/l…

🧵1/3

#LLMs #ML #DeepLearning Image
How does int8 quantization work? 🤔

It's a 2-part procedure with
1) using 8bit quantization
2) 16-bit matmuls for outlier feature dimensions

Check out the LLM.int8(): 8-bit Matrix Multiplication for Transformers at Scale paper for details arxiv.org/abs/2208.07339

🧵2/3
And how about int4? 🤔

It's a one-shot weight quantization method based on approximate second-order information⚙️📉

For more details, see GPTQ: Accurate Post-Training Quantization for Generative Pre-trained Transformers: 📚🔍 arxiv.org/abs/2210.17323

🧵3/3

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Lightning AI ⚡️

Lightning AI ⚡️ Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @LightningAI

Jan 19
Train a 20-billion parameter GPT model for text prediction on 3 GPU nodes with Lightning. 🤯

The entire training process is contained in a simple script that you can scan, read, and understand in just a few seconds.✅

🧵(1/4) Image
You can customize any part of this process: the dataset, model, hyperparameters, and training strategy.

Also, you can easily swap out which 🛠️hardware you’re using. With a single flag, you can choose to run on 3 nodes (as in this example), or higher to fit your model.

🧵(2/4)
The Lightning Platform also enables you to perform multi-node training from scratch without the hassle of setting up infrastructure or worrying about managing multi-node communication. 🤩

🧵(3/4)
Read 4 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us on Twitter!

:(