Troubleshooting Soft Prompt Optimization
A machine learning engineer is training a soft prompt to summarize complex technical articles. The goal is to create a short, efficient prompt that produces summaries of the same quality as those generated from a much longer, more detailed prompt. The engineer uses the optimization formula:
Where:
- is the high-quality summary from the long prompt.
- is the summary from the soft prompt being trained.
- is a function measuring the dissimilarity between the two summaries.
After extensive training, the dissimilarity score remains high, and the summaries generated using the soft prompt are of poor quality. Analyze two distinct potential reasons for this failure. For each reason, explain how it relates to the components of the provided optimization formula.
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
Troubleshooting Soft Prompt Optimization
A researcher is using the following formula to find the best soft prompt (σ) for a large language model:
hat(σ) = arg min_σ s(hat(y), hat(y)_σ)In this formula,
hat(y)is the model's prediction given a full, descriptive context, andhat(y)_σis the prediction given the soft prompt. What is the fundamental goal of this optimization process?Evaluating Soft Prompt Generalization