Learn Before
Concept

Challenges in Defining Human Preferences for LLM Alignment

A fundamental challenge in aligning Large Language Models is that humans often have difficulty precisely articulating their own preferences and values upfront. In many cases, it is hard to accurately describe what is desired until we actually observe the model's responses to user requests. This ambiguity complicates the process of creating comprehensive guidelines and training datasets.

0

1

Updated 2026-04-20

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.2 Generative Models - Foundations of Large Language Models