Concept

Sources of LLM Misalignment

Misalignment in Large Language Models stems from several key issues. One primary source is a model's failure to follow user instructions, a capability not inherently acquired during pre-training. Another significant cause is the model's tendency to generate harmful content or replicate societal biases that are present in its training data.

0

1

Updated 2026-01-15

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences