Different pretraining tasks for pretrained NLP models.
In recent years, there are a huge number of pretrained NLP models created by researchers. In pretraining the model, a huge amount of data is taken and passed through a model with specific task. These tasks are called as pre-training task. In this article we gonna look at different pre-training tasks:
Depending on whether the pretrained model is encoder only or encoder-decoder framework, there are various pretraining tasks:
-
Masked Language modeling
-
Span Correction
Cross-lingual text-to-text pretraining tasks:
-
Translation Pair Span correction
-
Translation Span Correction
-
Machine Translation
5. Machine Translation
Machine Translation is used as a text-to-text pretraining task for Seq-to-Seq learning. It was used in paper ‘Cross-Lingual Natural Language Generation via Pre-training’. Let e and f denote a sentence and its corresponding translation. We directly use e and f as the input and output sequences, respectively. The loss function of MT is
Lmt(e,f) = L(e -> f)