Learn Before
Approaches to Extending BERT for Multilingual Support
Multi-lingual BERT (mBERT)
Multi-lingual BERT (mBERT) is a version of BERT trained on text from 104 different languages. Its main distinction from monolingual BERT is the use of a significantly larger vocabulary to accommodate tokens from this diverse set of languages. This design allows mBERT to map representations from different languages into a common vector space, which enables the model to share and transfer knowledge across languages.
0
1
9 days ago
Tags
Ch.1 Pre-training - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Multi-lingual BERT (mBERT)
Multilingual and Language-Specific PTMs