Multiple Choice

A research team is refining a language model's ability to be helpful and harmless. They use two distinct datasets for this process. Dataset 1 contains prompts, each paired with a single, meticulously crafted, ideal response. Dataset 2 contains prompts, each paired with two different model-generated responses, along with a label indicating which of the two responses a human preferred. Which statement best distinguishes the fundamental optimization objective when training on Dataset 1 versus Dataset 2?

0

1

Updated 2025-10-06

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science