Mikel Artetxe Profile picture
Nov 3, 2022 11 tweets 7 min read Read on X
📣 9 papers accepted at #emnlp2022 (7 main conference + 2 Findings)

🧵 with links to camera ready preprints 👇
1) “Does Corpus Quality Really Matter for Low-Resource Languages?”

We introduce a new corpus for Basque that has a higher quality according to annotators, but find that this improvement does not carry over to downstream NLU tasks.

arxiv.org/abs/2203.08111 ImageImage
2) “Efficient Large Scale Language Modeling with Mixtures of Experts”

We study how MoE LMs scale in comparison with dense LMs in a wide range of settings. MoEs are more efficient, but their advantage reduces at scale and varies greatly across tasks!

arxiv.org/abs/2112.10684 Image
3) “Don't Prompt, Search! Mining-based Zero-Shot Learning with Language Models”, led by @mozeskar

We propose a mining-based approach to zero-shot learning, outperforming prompting with comparable templates.

arxiv.org/abs/2210.14803 Image
4) “Prompting ELECTRA: Few-Shot Learning with Discriminative Pre-Trained Models”, led by @xiamengzhou

We show that discriminative models like ELECTRA outperform generative models like BERT in few-shot prompting.

arxiv.org/abs/2205.15223 Image
5) “Multilingual Machine Translation with Hyper-Adapters”, led by @cbaziotis

We address prior issues on scaling hyper-networks, and use them to generate language-specific adapters for multilingual MT, matching regular adapters with 12x less parameters!

arxiv.org/abs/2205.10835 Image
6) “Few-shot Learning with Multilingual Generative Language Models”, led by @VictoriaLinML & @xl_nlp

We introduce a family of multilingual autoregressive LMs and study them on cross-lingual few-shot learning and machine translation.

arxiv.org/abs/2112.10668 Image
7) “Rethinking the Role of Demonstrations: What Makes In-Context Learning Work?”, led by @sewon__min

We show that few-shot prompting with random labels is almost as good as with true labels, providing a new perspective on how in-context learning works.

arxiv.org/abs/2202.12837 Image
8) “On the Role of Bidirectionality in Language Model Pre-Training” (Findings)

We study the role of bidirectional attention and bidirectional context in LM pre-training through a new framework that generalizes prior approaches.

arxiv.org/abs/2205.11726 Image
9) “PoeLM: A Meter- and Rhyme-Controllable Language Model for Unsupervised Poetry Generation” (Findings), led by @aormazabalo

We propose an unsupervised approach to formal verse poetry generation based on control codes.

arxiv.org/abs/2205.12206 Image
Thanks to all my awesome co-authors and see you all in Abu Dhabi!

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Mikel Artetxe

Mikel Artetxe Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(