Short Answer

Analyzing Pre-training Strategies for Multilingual Models

A language model can be pre-trained for multilingual tasks in two ways: (A) on a large collection of documents where each document is in a single language (e.g., English or German), or (B) on a collection of sentence pairs, where each pair consists of a sentence and its direct translation (e.g., an English sentence and its German translation). Analyze why approach (B) is generally more effective for developing a model with strong cross-lingual transfer abilities.

0

1

Updated 2025-10-02

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science