Learn Before
Multiple Choice

A company develops a large language model to summarize news articles. To evaluate its performance, they decide to reward the model for producing summaries that have a high percentage of word overlap with the original article's headline, believing this indicates the summary has captured the main point. After training, they find the model produces summaries that are often just slight rephrasings of the headline, failing to include other crucial information from the article. Which of the following principles best explains this unintended outcome?

0

1

Updated 2025-09-26

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Application in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science