Evaluating Soft Prompt Performance
Based on the optimization objective of minimizing the KL divergence between the distribution from the full context and the distribution from the soft prompt, which soft prompt (σ_A or σ_B) is superior? Justify your answer by explaining how the distributions relate to the optimization goal, without calculating the exact KL divergence value.
0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Analysis in Bloom's Taxonomy
Cognitive Psychology
Psychology
Social Science
Empirical Science
Science
Related
A researcher is training a soft prompt, denoted as (\sigma), to mimic the behavior of a full context, (c), for a given input, (z). They use the Kullback-Leibler (KL) divergence between the model's output probability distributions as their objective function: After extensive training, the researcher observes that the KL divergence has reached a value of 0. What is the most accurate conclusion to draw from this result?
Evaluating Soft Prompt Performance
Analyzing the Asymmetry in Soft Prompt Optimization