Learn Before
Optimization Goal for Soft Prompt Learning via Context Compression
When learning a soft prompt, denoted as , through context compression, the objective is to make the model's predictions for a given input, , nearly identical whether using the compact soft prompt or the original, longer context . This is achieved by minimizing the difference between the outputs generated under these two conditions. This optimization goal can be formalized with a mathematical expression that seeks the prompt that minimizes the dissimilarity between the two predictions.

0
1
Tags
Ch.4 Alignment - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Ch.3 Prompting - Foundations of Large Language Models
Related
Optimization Goal for Soft Prompt Learning via Context Compression
Challenge of Context Compression for Long Sequences
Prompt as a Form of Context
A research team is developing a system where a very long, detailed set of instructions is 'compressed' into a compact, learnable set of numerical values. This compact representation is then used to guide a language model in performing a specific task, aiming to replicate the performance that would be achieved if the model had processed the full set of instructions. What is the most significant practical challenge the team will face when implementing this 'compression' process?
Applying Context Compression for a Specialized Task
Sequential Context Compression with an RNN-like Mechanism
The Goal of Context Compression for Soft Prompts
Learn After
Soft Prompt Learning as Context Compression via Knowledge Distillation
Formula for Optimizing Soft Prompts via Context Compression
Alternative Methods for Soft Prompt Optimization
A developer is tasked with creating a compact, learned 'soft prompt' that can effectively replace a very long and detailed set of instructions (the 'full context') for a language model. The objective is to ensure that for any given user query, the model's final output is nearly identical whether it's conditioned on the long instructions or the new compact prompt. Which of the following optimization strategies directly targets this specific objective?
When training a soft prompt to act as a compressed version of a longer context, the primary optimization objective is to ensure the learned soft prompt's vector representation is as close as possible to the vector representation of the original context.
Debugging Soft Prompt Optimization
Interpreting the Soft Prompt Optimization Formula