Concept

Limitations of Human Feedback in LLM Alignment

While learning from human preferences is an effective method for aligning Large Language Models, it has significant practical limitations. The process of annotating preference data is costly and difficult to scale. Furthermore, human feedback is inherently subjective, which can introduce biases and inconsistencies into the model's alignment.

0

1

Updated 2026-05-03

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Related