Learn Before
Comparison

Comparison of Prefilling and BERT Encoding

The prefilling phase is analogous to BERT's encoding process in that both convert an input sequence into contextualized token representations. However, they differ fundamentally in their operational mode: prefilling is unidirectional, consistent with standard autoregressive language modeling, whereas BERT employs a bidirectional approach to generate its representations.

0

1

Updated 2026-05-03

Contributors are:

Who are from:

Tags

Ch.5 Inference - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences