Alternative Methods for Soft Prompt Optimization
Soft prompt optimization can be framed in at least two ways. The first method involves maximizing the log-probability of the desired output, treating it as a maximum likelihood problem. An alternative approach is to minimize the Kullback-Leibler (KL) divergence between the model's output distribution when using the full context and its distribution when using the compressed soft prompt.
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Soft Prompt Learning as Context Compression via Knowledge Distillation
Formula for Optimizing Soft Prompts via Context Compression
Alternative Methods for Soft Prompt Optimization
A developer is tasked with creating a compact, learned 'soft prompt' that can effectively replace a very long and detailed set of instructions (the 'full context') for a language model. The objective is to ensure that for any given user query, the model's final output is nearly identical whether it's conditioned on the long instructions or the new compact prompt. Which of the following optimization strategies directly targets this specific objective?
When training a soft prompt to act as a compressed version of a longer context, the primary optimization objective is to ensure the learned soft prompt's vector representation is as close as possible to the vector representation of the original context.
Debugging Soft Prompt Optimization
Interpreting the Soft Prompt Optimization Formula
Learn After
Formula for Soft Prompt Optimization via Log-Likelihood Maximization
Formula for Soft Prompt Optimization by Minimizing KL Divergence
A team is creating a soft prompt to summarize a complex user manual for a question-answering model. Their main objective is not just to get the single correct answer, but to ensure the model's uncertainty and its ranking of other plausible-but-incorrect answers are the same with the soft prompt as they were with the full manual. Which of the following optimization strategies best aligns with this specific objective?
Choosing an Optimization Strategy for Soft Prompts
A researcher is optimizing a soft prompt. With the original, long context, the model predicts the correct answer with 60% probability and a plausible alternative with 30% probability. The researcher's goal is to create a soft prompt that causes the model to predict the correct answer with over 95% probability, even if this significantly changes the probability of the alternative answer. Which optimization approach is better suited for this specific goal?