Learn Before
Pre-training Strategy for a Multilingual Model
Analyze the two pre-training strategies presented in the case study. Which strategy is more likely to result in a model with superior cross-lingual capabilities? Justify your choice by explaining the fundamental difference in what the model learns from each approach.
0
1
Tags
Ch.1 Pre-training - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Bilingual Sentence Packing for Pre-training
Pre-training Strategy for a Multilingual Model
A researcher is pre-training a multilingual model using a masked language modeling (MLM) objective. To align the pre-training process with the specific methodology of Cross-Lingual Language Models (XLMs), what is the most crucial characteristic of the input data?
Core Training Principle of XLM
Translation Language Modeling
Input Embedding in Cross-Lingual Language Models