Theory

Chinchilla Scaling Law

The Chinchilla scaling law provides a framework for predicting language model performance. According to this principle, the test loss per token is derived by adding a constant irreducible error term to two separate inverse proportional functions: one based on the model size (NN) and the other based on the training dataset size (DD).

Image 0

0

1

Updated 2026-04-22

Contributors are:

Who are from:

Tags

Ch.2 Generative Models - Foundations of Large Language Models

Foundations of Large Language Models

Foundations of Large Language Models Course

Computing Sciences