Formula

Formula for Optimizing Soft Prompts via Context Compression

The optimal soft prompt, denoted as σ^\hat{\sigma}, is determined by minimizing a function s(,)s(\cdot, \cdot) that compares the prediction from the full context, y^\hat{y}, with the prediction from the compressed context, y^σ\hat{y}_{\sigma}. This function typically represents a loss or similarity measure. The optimization problem is formally expressed as: σ^=argminσs(y^,y^σ)\hat{\sigma} = \underset{\sigma}{\arg\min}\, s(\hat{y}, \hat{y}_{\sigma})

Image 0

0

1

Updated 2026-05-02

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.3 Prompting - Foundations of Large Language Models

Related