Learn Before
Concept

Input Embedding in Cross-Lingual Language Models

In the work of Lample and Conneau on cross-lingual language models, the input embedding for a specific token (denoted as ei\mathbf{e}_i) is calculated as the sum of its token embedding, positional embedding, and a language embedding. The inclusion of a language embedding requires assigning a language label to each token, which enables the model to distinguish between tokens from different languages.

0

1

Updated 2026-04-18

Contributors are:

Who are from:

Tags

Foundations of Large Language Models

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences