Learn Before
Concept

Multi-lingual Pre-training for Encoder-Decoder Models

To achieve effectiveness in multi-lingual and cross-lingual applications like machine translation, pre-trained encoder-decoder models require training on multi-lingual data. This process necessitates a vocabulary containing tokens from all involved languages, which allows the model to learn shared representations and develop capabilities for both understanding and generation across different languages.

0

1

Updated 2026-04-16

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Related