Introducing MASS – A pre-training method that outperforms BERT and GPT in sequence to sequence language generation tasks

Editor’s note: Since 2018, pre-training has without a doubt become one of the hottest research topics in Natural Language Processing (NLP). By leveraging generalized language models like the BERT, GPT and XLNet, great breakthroughs have been achieved in natural language understanding. However, in sequence to sequence based language generation tasks, the popular pre-training methods have … Continue reading Introducing MASS – A pre-training method that outperforms BERT and GPT in sequence to sequence language generation tasks