Concept

Factors Influencing Multilingual Pre-training

The effectiveness of a multilingual pre-trained model, assuming a fixed architecture, is determined by several key configuration choices. These include the size of the shared vocabulary, the proportion of training data allocated to each language, and the overall size of the model itself.

0

1

Updated 2026-04-18

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences