Concept

Soft Prompt Learning as Context Compression via Knowledge Distillation

This technique frames soft prompt learning as a context compression problem solved using knowledge distillation. In this process, the knowledge from a lengthy, standard prompt is distilled from a teacher model into a compact set of 'pseudo tokens'. The embeddings for these pseudo tokens, which are appended to the user's input sequence, are then optimized to replicate the predictions of the teacher model, effectively capturing the essence of the original, more complex prompt.

Image 0

0

1

Updated 2026-05-02

Contributors are:

Who are from:

Tags

Ch.4 Alignment - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.3 Prompting - Foundations of Large Language Models

Related