Learn Before
Translation Language Modeling
Translation language modeling is a pre-training objective designed to align token representations across different languages. It involves concatenating sequences from two languages and replacing a certain percentage of tokens with a special mask symbol, such as [MASK]. The model's objective is to maximize the probability of correctly predicting these masked tokens based on the surrounding context. By doing so, the model learns to capture cross-lingual correspondences, as predicting a masked token in one language often requires leveraging information from the unmasked tokens in the other language. This cross-lingual alignment essentially enables the model to function as a translation model.
0
1
Tags
Foundations of Large Language Models
Ch.1 Pre-training - Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Bilingual Sentence Packing for Pre-training
Pre-training Strategy for a Multilingual Model
A researcher is pre-training a multilingual model using a masked language modeling (MLM) objective. To align the pre-training process with the specific methodology of Cross-Lingual Language Models (XLMs), what is the most crucial characteristic of the input data?
Core Training Principle of XLM
Translation Language Modeling
Input Embedding in Cross-Lingual Language Models