Comparison

Comparison of Prefix Language Modeling and Causal Language Modeling

The primary distinction between Prefix Language Modeling (PrefixLM) and Causal Language Modeling (CLM) lies in how context is processed and generation is initiated. In standard CLM, the entire text sequence is generated autoregressively, with each token prediction conditioned on all preceding tokens starting from the very beginning. In contrast, PrefixLM splits the task: an encoder first processes an initial prefix non-causally (all at once) to create a contextual representation. Then, a decoder uses this context to autoregressively generate only the subsequent part of the sequence.

Image 0

0

1

Updated 2026-04-16

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences