My first @GoogleAI residency project was accepted to @emnlpmeeting #EMNLP2021!

Prompt Tuning can condition a frozen T5 XXL model to perform new tasks while only adding 0.003% more parameters and no performance loss.

Camera Ready 📸: arxiv.org/abs/2104.08691

Quick Thread 🧵(1/7) A graph plotting model performance on SuperGLUE verse the nu
Fine-tuning all the parameters of large pre-trained models works well and is the core of many SotA NLP results right now, but has some sharp edges. The size can make them difficult to work with and serve, plus each fine-tuning run creates a fork. (2/7)
Prompt Tuning, learning a small set of parameters that are prepended to the embedded input, can eliminate these problems. Freezing pre-trained models enables mixed-task batching and efficient ensembling, without the need for multiple copies. (3/7) A diagram of model tuning vs prompt tuning inference, on the
The size of the pre-trained model is critical to Prompt Tuning performance. As we scale T5 from Small to XXL, we see it closes the gap with full fine-tuning. (4/7)
By keeping the pretrained model frozen, Prompt Tuning avoids overfitting to a specific task, improving performance on domain shift problems. See our paper for details, as well as comparison with other recent “P*-Tuning” approaches. (5/7)
An interesting quirk of Prompt Tuning is that the hyper parameters are a little strange. For example, our learning rate is 0.3, two orders of magnitude larger than the default T5 learning rate of 0.001. (6/7)
A huge shout out to my amazing mentors, @noahconst and @aboSamoor, who were a big part of making this project possible. (7/7)

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Brian Lester

Brian Lester Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!

Follow Us on Twitter!

:(