MPNet combines strengths of masked and permuted language modeling for language understanding

Pretrained language models have been a hot research topic in natural language processing. These models, such as BERT, are usually pretrained on large-scale language corpora with carefully designed pretraining objectives and then fine-tuned on downstream tasks to boost the accuracy. Among these, masked language modeling (MLM), adopted in BERT, and permuted language modeling (PLM), adopted in XLNet, are two representative pretraining objectives. However, both of them enjoy their own advantages but suffer from limitations.

Therefore, researchers from Microsoft Research Asia, after proposing MASS for language generation last

 

 

To finish reading, please visit source site

Leave a Reply