Formula for Soft Prompt Optimization by Minimizing Prediction Dissimilarity
The optimal soft prompt, denoted as , can be determined by finding the prompt that minimizes the dissimilarity between the model's predictions with and without the full context. This is expressed by the formula: Here, is a function measuring the dissimilarity (e.g., distance) between the prediction from the full context, , and the prediction using the soft prompt, . This method aligns the compact prompt's behavior with that of the original, more descriptive prompt.

0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Formula for Soft Prompt Optimization by Minimizing Prediction Dissimilarity
Optimizing Language Model API Costs
A team is training a set of learnable, continuous parameters to serve as a compact substitute for a long, detailed textual instruction set for a language model. The goal is for these compact parameters to guide the model to produce the same quality of output as the original long instructions when given any user input. Which of the following best describes the core objective of this training process?
Characteristics of Teacher and Student Models in Knowledge Distillation
In the framework of learning a soft prompt via knowledge distillation to compress a longer context, match each component with its corresponding role in the process.
Learn After
Troubleshooting Soft Prompt Optimization
A researcher is using the following formula to find the best soft prompt (σ) for a large language model:
hat(σ) = arg min_σ s(hat(y), hat(y)_σ)In this formula,
hat(y)is the model's prediction given a full, descriptive context, andhat(y)_σis the prediction given the soft prompt. What is the fundamental goal of this optimization process?Evaluating Soft Prompt Generalization