PhD student @UMassCS studying #NLProc, member of @umass_nlp. Former undergrad @IITBombay and intern @GoogleAI, @TTIC_Connect, @mozilla.
May 24, 2022 • 8 tweets • 5 min read
Excited to share RankGen, a 1.2B param contrastive encoder mapping prefixes & model generations to vectors.
✅ large improvements over nucleus/typical sampling
✅ score & rank generations from any LM
✅ human eval with writers
✅ HuggingFace ckpts, code👇 arxiv.org/abs/2205.09726
Despite great progress, text generation continues to underperform. Even large LMs generate text that has hallucination, poor continuity etc.
Part of the issue is LMs are trained to predict the next one token given the ground truth prefix, encouraging reliance on local context.