Short Answer

Analysis of a Self-Supervised Training Strategy

A team is training a model to perform a complex generation task. Their training process for each input involves two steps:

  1. First, the model is used to determine the single most probable output sequence, which we'll call the 'optimal output'.
  2. Second, the model's parameters are adjusted to maximize the probability of producing that same 'optimal output', but this time, the model is given a slightly modified and less complete version of the original input.

Based on this two-step process, what is the primary capability the model is being trained to develop, and why is this approach potentially more powerful than simply training the model on a fixed set of pre-written input-output pairs?

0

1

Updated 2025-10-05

Contributors are:

Who are from:

Tags

Ch.3 Prompting - Foundations of Large Language Models

Foundations of Large Language Models

Computing Sciences

Foundations of Large Language Models Course

Analysis in Bloom's Taxonomy

Cognitive Psychology

Psychology

Social Science

Empirical Science

Science