Definition

LLM Alignment with Human Expectations

In the context of Large Language Models, alignment is the process of ensuring that a model's outputs conform to human expectations and intentions. This modern definition shifts the focus from simple data mapping to shaping the model's overall behavior to be helpful, harmless, and in accordance with user goals.

0

1

Updated 2025-10-08

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences