Learn Before
Activity (Process)

Standard Transformer Encoding Procedure

The standard procedure for processing an input sequence with a Transformer encoder begins by representing each input token, xix_i, as its corresponding embedding, ei\mathbf{e}_i. This sequence of embeddings, e0,...,em\mathbf{e}_{0},...,\mathbf{e}_m, is then fed into the encoder. The encoder processes this input to produce a sequence of contextualized output vectors, or hidden states, h0,...,hm\mathbf{h}_{0},...,\mathbf{h}_m.

0

1

Updated 2026-04-17

Contributors are:

Who are from:

Tags

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences