Learn Before
Concept

Negative Likelihood Loss in Sequence Labeling

A sequence labeling model, such as one used for Named Entity Recognition (NER), outputs a probability distribution over the possible tag set at each position, generating a sequence of distributions p1,,pm{\mathbf{p}_1, \dots, \mathbf{p}_m}. The training or fine-tuning of the model can be performed over these distributions. For example, if pi(tagi)p_i(\mathrm{tag}_i) represents the model's assigned probability for the correct tag at position ii, the training loss can be defined as the negative likelihood.

0

1

Updated 2026-04-18

Contributors are:

Who are from:

Tags

Foundations of Large Language Models

Ch.1 Pre-training - Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences