Learn Before
Problem

Inarticulacy of Human Preferences as an Alignment Challenge

LLM alignment is complicated by the fact that humans often struggle to precisely or consistently articulate their preferences beforehand. This inherent ambiguity makes it difficult to define a clear objective for the model. Often, desired behavior only becomes clear to a user after they have observed an undesirable or unexpected response from the LLM, highlighting the challenge of creating comprehensive alignment guidelines in advance.

0

1

Updated 2026-04-20

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.2 Generative Models - Foundations of Large Language Models