Learn Before
Sources of LLM Misalignment
Misalignment in Large Language Models stems from several key issues. One primary source is a model's failure to follow user instructions, a capability not inherently acquired during pre-training. Another significant cause is the model's tendency to generate harmful content or replicate societal biases that are present in its training data.
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Learn After
A user asks a large language model to write a short biography of a famous female scientist. The model's response disproportionately focuses on her personal life and relationships, while similar requests for male scientists result in biographies centered on their professional accomplishments. Which source of model misalignment does this scenario best illustrate?
Analyzing Misalignment in a Customer Support LLM
Match each example of a large language model's output with the primary source of misalignment it demonstrates.