One of the hardest challenges in low-resource machine translation or NL generation nowadays is generating *into* low-resource languages, as it's hard to fluently generate sentences with few training examples. Check out our new #EMNLP2020 findings paper that tackles this. 1/3
We argue that target-side learning of good word representations shared across languages with different spelling is crucial. To do so, we apply soft decoupled encoding (arxiv.org/pdf/1902.03499), a multilingual word embedding designed to maximize cross-lingual sharing. 2/3
Results are encouraging, with consistent improvements over 4 languages! Also, the method plays well with subword segmentation like BPE, and has no test-time overhead, keeping inference speed fast. Check it out! 3/3
• • •
Missing some Tweet in this thread? You can try to
force a refresh
Super-excited about our new #ICASSP2020 paper on "Universal Phone Recognition with a Multilingual Allophone System" arxiv.org/abs/2002.11800
We create a multi-lingual ASR model that can do zero-shot phone recognition in up to 2,186 languages! How? A little linguistics :) 1/5
In our speech there are phonemes (sounds that can support lexical contrasts in a *particular* language) and their corresponding phones (the sounds that are actually spoken, which are language *independent*). Most multilingual ASR models conflate these two concepts. 2/5
We create a model that first recognizes to language-independent phones, and then converts these phones to language-specific phonemes. This makes our underlying representations of phones more universal and generalizable across languages. 3/5