Prakhar Gupta Profile picture
Oct 31 9 tweets 4 min read
Can instruction tuning improve zero and few-shot performance on dialogue tasks? We introduce InstructDial, a framework that consists of 48 dialogue tasks created from 59 openly available dialogue datasets
#EMNLP2022🚀
Paper 👉 arxiv.org/abs/2205.12673
Work done at @LTIatCMU
🧵👇 We investigate instruction ...
Instruction tuning involves fine-tuning a model on a collection of tasks specified through natural language instructions (T0, Flan models). We systematically studied instruction tuning for dialogue tasks and show it works a lot better than you might expect!
The InstructDial framework consists of 48 diverse dialogue tasks varying from classification, grounded and controlled generation, safety, QA, pretraining, summarization, NLI, and other miscellaneous tasks. All tasks are specified through instructions in a seq-2-seq format. In our taxonomy, Classifica...
We introduce two meta-tasks to learn the association between the instructions, inputs, and outputs for a task. For example, in Instruction selection task, the model selects the correct instruction for an input-output pair. We show this improves generalization to new instructions!
Our results show that our models Dial-T0 and Dial-Bart0 outperform the baselines and GPT-3 models on unseens tasks. Surprisingly, the smaller Bart-based model outperforms T5 on some tasks. Even training on a few (100) instances of the test tasks leads to significant improvements! Our models perform better t...
On zero-shot dialogue evaluation, DIAL-T0 outperforms all baseline models on multi-domain eval sets. Our models also demonstrate great zero and few-shot performance across dialogue tasks such as intent detection, slot filling, and dialogue state tracking (more results in paper). On zero-shot dialpogue eval...
We performed ablation studies showing the impact of model size/type, number of tasks, and meta-tasks. Our experiments reveal that instruction tuning does not benefit all unseen test tasks and that improvements can be made in instruction wording invariance and task interference.
Our code, data, and models are available at
github.com/prakharguptaz/…
We provide data readers for a huge collection of datasets, and scripts for training and evaluation.
Models can be found at huggingface huggingface.co/prakharz/DIAL-…,
prakharz/DIAL-BART0,
prakharz/DIAL-FLANT5-XL😀
We hope InstructDial will facilitate progress on instruction-tuning and multi-task training for dialogue. #NLProc
Credits to co-authors: Cathy Jiao, Yi-Ting Yeh, @shikibmehri, Maxine Eskenazi, @jeffbigham 🎉

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Prakhar Gupta

Prakhar Gupta Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us on Twitter!

:(