About This Document
- sl:arxiv_author :
- sl:arxiv_firstAuthor : Aditi Chaudhary
- sl:arxiv_num : 2010.12566
- sl:arxiv_published : 2020-10-23T17:53:11Z
- sl:arxiv_summary : Pre-trained multilingual language models such as mBERT have shown immense
gains for several natural language processing (NLP) tasks, especially in the
zero-shot cross-lingual setting. Most, if not all, of these pre-trained models
rely on the masked-language modeling (MLM) objective as the key language
learning objective. The principle behind these approaches is that predicting
the masked words with the help of the surrounding text helps learn potent
contextualized representations. Despite the strong representation learning
capability enabled by MLM, we demonstrate an inherent limitation of MLM for
multilingual representation learning. In particular, by requiring the model to
predict the language-specific token, the MLM objective disincentivizes learning
a language-agnostic representation -- which is a key goal of multilingual
pre-training. Therefore to encourage better cross-lingual representation
learning we propose the DICT-MLM method. DICT-MLM works by incentivizing the
model to be able to predict not just the original masked word, but potentially
any of its cross-lingual synonyms as well. Our empirical analysis on multiple
downstream tasks spanning 30+ languages, demonstrates the efficacy of the
proposed approach and its ability to learn better multilingual representations.@en
- sl:arxiv_title : DICT-MLM: Improved Multilingual Pre-Training using Bilingual Dictionaries@en
- sl:arxiv_updated : 2020-10-23T17:53:11Z
- sl:bookmarkOf : https://arxiv.org/abs/2010.12566
- sl:creationDate : 2021-09-06
- sl:creationTime : 2021-09-06T18:27:44Z
Documents with similar tags (experimental)