Learn Before
Concept

Rationale for Categorizing Pre-training Tasks by Objective

Categorizing pre-training tasks based purely on model architecture is suboptimal because a single training objective can be applied across varying architectures (e.g., masked language modeling works for both encoder-only and encoder-decoder setups). Therefore, a better approach is to classify these tasks based on their specific training objectives.

0

1

Updated 2026-04-17

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences