Multiple Choice

A research team trains two different models: one with an encoder-only structure and another with an encoder-decoder structure. Both models are trained using the exact same objective: predicting randomly masked words within a text. A colleague argues that because the model structures are different, they should be classified under separate pre-training categories. Why is this argument fundamentally flawed from a conceptual standpoint?

0

1

Updated 2025-09-28

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science