Learn Before
Concept

Impracticality of Achieving Alignment Solely Through Pre-training

While it is theoretically conceivable that pre-training on a sufficiently massive dataset—one that comprehensively covers all possible tasks and perfectly aligns with human preferences—could produce Large Language Models that are both accurate and safe without further alignment, this approach is practically unfeasible. In reality, it is nearly impossible to gather a dataset that encompasses every potential task or adequately represents the vast spectrum of human preferences, making pre-training alone insufficient for achieving proper model alignment.

0

1

Updated 2026-04-30

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences