Learn Before
Activity (Process)

Single-Text Classification with BERT Models

As one of the most widely-used applications of BERT, single-text classification processes an input sequence to determine its overall category. The input text is typically formatted as a sequence of tokens, such as [CLS] x1x2...xmx_1 x_2 ... x_m. The BERT model receives this sequence and encodes it into a corresponding sequence of vectors. The initial output vector, denoted as hcls\mathbf{h}_{\mathrm{cls}} (or h0\mathbf{h}_{0}), is typically extracted as the comprehensive representation of the entire input text. A prediction network then takes this single hcls\mathbf{h}_{\mathrm{cls}} vector as its input to produce a probability distribution over the possible labels.

Image 0

0

1

Updated 2026-04-18

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences

Ch.1 Pre-training - Foundations of Large Language Models

Learn After