Formula

Log-Probability Loss with Model-Generated Target

In certain training paradigms, the learning target is generated by the model itself rather than being a fixed ground-truth label. First, an optimal prediction, y^\hat{\mathbf{y}}, is determined, often by maximizing a log-probability function. This prediction y^\hat{\mathbf{y}} is then used as the target for learning. The loss function is subsequently defined as the log-probability of this model-generated target, conditioned on variables such as a modified context c\mathbf{c}' and a latent variable z\mathbf{z}. The formula is: Loss=logPrθs(y^c,z)\text{Loss} = \log \text{Pr}_{\theta}^{s}(\hat{\mathbf{y}}|\mathbf{c}', \mathbf{z}) This objective is typically maximized during training, which is equivalent to minimizing its negative.

Image 0

0

1

Updated 2025-10-08

Contributors are:

Who are from:

Tags

Ch.3 Prompting - Foundations of Large Language Models

Foundations of Large Language Models

Computing Sciences

Foundations of Large Language Models Course