Comparison

Differing Motivations of Instruction and Human Preference Alignment

Instruction alignment and human preference alignment are driven by distinct goals. The primary motivation for instruction alignment is to make a model generate outputs that adhere closely to explicit user commands, whereas human preference alignment is motivated by the need to train a model based on broader, often implicit, human feedback.

0

1

Updated 2026-04-30

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences