Activity (Process)

Inter-Layer Data Flow in Prefix-Tuning

In a prefix-tuned Transformer, the data flow between layers follows a recursive pattern. The output from a given layer, which consists of the hidden states corresponding only to the original input sequence (after the prefix-related states are discarded), becomes the input for the subsequent layer. This output is then concatenated with the next layer's unique set of trainable prefix vectors, forming the complete input for that layer's computation.

Image 0

0

1

Updated 2026-05-02

Contributors are:

Who are from:

Tags

Ch.3 Prompting - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences