Concept

Cross-Lingual Language Models (XLM)

Proposed by Lample and Conneau (2019), Cross-Lingual Language Models (XLMs) are a specific approach to pre-training that leverages bilingual data. An XLM can be trained using either a causal language modeling (CLM) or a masked language modeling (MLM) objective. When using the MLM approach, the training objective is identical to BERT's, where the model, treated as an encoder, learns to predict randomly selected tokens that have been masked, replaced, or left unchanged in the input.

0

1

Updated 2026-05-02

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences