Learn Before
  • Approaches to Extending BERT for Multilingual Support

Multi-lingual BERT (mBERT)

Multi-lingual BERT (mBERT) is a version of BERT trained on text from 104 different languages. Its main distinction from monolingual BERT is the use of a significantly larger vocabulary to accommodate tokens from this diverse set of languages. This design allows mBERT to map representations from different languages into a common vector space, which enables the model to share and transfer knowledge across languages.

0

1

9 days ago

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Related
  • Multi-lingual BERT (mBERT)

  • Multilingual and Language-Specific PTMs