Activity (Process)

Bilingual Sentence Packing for Pre-training

A specific technique used in bilingual pre-training involves sampling a pair of aligned sentences from two different languages. These sentences are then concatenated to form a single, combined sequence that is used as a training input. This method directly exposes the model to parallel data within a single training instance, facilitating cross-lingual learning.

0

1

Updated 2026-04-18

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Related