Concept

N-gram Language Modeling

For quite a long period, particularly before 2010, the dominant approach to language modeling was the nn-gram approach. In nn-gram language modeling, we estimate the probability of a word given its preceding n1n-1 words, and thus the probability of a sequence can be approximated by the product of a series of nn-gram probabilities. These probabilities are typically estimated by collecting smoothed relative counts of nn-grams in text. This straightforward approach has been extensively used in NLP, and the success of modern statistical speech recognition and machine translation systems has largely depended on the utilization of nn-gram language models.

0

1

Updated 2026-04-18

Tags

Data Science

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences