Formula

Formula for Soft Prompt Optimization by Minimizing Prediction Dissimilarity

The optimal soft prompt, denoted as σ^\hat{\sigma}, can be determined by finding the prompt that minimizes the dissimilarity between the model's predictions with and without the full context. This is expressed by the formula: σ^=argminσs(y^,y^σ)\hat{\sigma} = \arg \min_{\sigma} s(\hat{\mathbf{y}}, \hat{\mathbf{y}}_{\sigma}) Here, ss is a function measuring the dissimilarity (e.g., distance) between the prediction from the full context, y^\hat{\mathbf{y}}, and the prediction using the soft prompt, y^σ\hat{\mathbf{y}}_{\sigma}. This method aligns the compact prompt's behavior with that of the original, more descriptive prompt.

Image 0

0

1

Updated 2025-10-08

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences