Input Composition Formula for Prompt Tuning
The actual input sequence provided to a Large Language Model during prompt tuning is formed by prepending trainable pseudo embeddings to the standard token embeddings. This composition is formally expressed as: Here, are the trainable soft prompt embeddings optimized for a specific task, and are the fixed token embeddings representing the user's input.

0
1
Tags
Ch.3 Prompting - Foundations of Large Language Models
Foundations of Large Language Models
Foundations of Large Language Models Course
Computing Sciences
Related
Input Composition Formula for Prompt Tuning
An engineer is adapting a large, pre-trained language model for a new task. To do this efficiently, they keep all the original model's parameters frozen. Their adaptation strategy involves modifying the input sequence before it is processed by the model. For any given text, they first convert the text into its standard sequence of numerical token representations. Then, they prepend a separate, short sequence of newly initialized, trainable numerical vectors to the beginning of that sequence. Only these new vectors are updated during training on the new task. Which statement best distinguishes the nature of these prepended, trainable vectors from the standard token representations?
You are examining the input layer of a large language model adapted using a parameter-efficient technique. The input is formed by combining two distinct types of numerical vectors. Match each vector type with its correct description.
Prefix Tuning Architecture
Parameter-Efficient Model Adaptation
Learn After
Visual Representation of Input Composition in Prompt Tuning
A language model is being adapted for a new task. The input sequence for this model is constructed by prepending a series of 10 newly introduced, trainable vectors to the 50 standard, frozen word vectors that represent an input sentence. Which statement accurately analyzes the composition of the final 60-vector input sequence fed into the model?
A large language model is being adapted for a specific task by modifying its input. The final input sequence is created by combining a set of newly introduced, learnable vectors with the standard vectors representing the input text. Arrange these two components in the correct order to form the final sequence that is fed into the model.
Debugging an Input Composition Method